# PyTorch

Bytedance Flux
Bytedance Flux
Flux is a high-performance communication overlap library developed by ByteDance, designed for tensor and expert parallelism on GPUs. Through efficient kernels and compatibility with PyTorch, it supports various parallelization strategies and is suitable for large-scale model training and inference. Flux's main advantages include high performance, ease of integration, and support for multiple NVIDIA GPU architectures. It excels in large-scale distributed training, particularly with Mixture-of-Experts (MoE) models, significantly improving computational efficiency.
Model Training and Deployment
66.0K
Profiling Data in DeepSeek Infra
Profiling Data In DeepSeek Infra
DeepSeek Profile Data is a project focused on performance analysis of deep learning frameworks. It captures performance data of training and inference frameworks through PyTorch Profiler, helping researchers and developers better understand computation and communication overlap strategies and underlying implementation details. This data is crucial for optimizing large-scale distributed training and inference tasks, significantly improving system efficiency and performance. This project is a significant contribution from the DeepSeek team in the field of deep learning infrastructure, aiming to promote the community's exploration of efficient computing strategies.
Model Training and Deployment
47.2K
InspireMusic
Inspiremusic
InspireMusic is an AIGC toolkit and model framework focused on music, songs, and audio generation, developed using PyTorch. It achieves high-quality music generation through audio tokenization and decoding processes, combining autoregressive transformers and conditional flow matching models. This toolkit supports multiple conditional controls such as text prompts, music styles, and structures, enabling the generation of high-quality audio at both 24kHz and 48kHz, as well as supporting long audio generation. Additionally, it offers convenient fine-tuning and inference scripts for users to adjust the model according to their needs. The open-source nature of InspireMusic aims to empower everyday users to enhance sound effects in their research through music creation.
Music Generation
60.7K
OLMo-2-1124-7B-DPO
Olmo 2 1124 7B DPO
OLMo-2-1124-7B-DPO is a large language model developed by the Allen Institute for Artificial Intelligence, which has been fine-tuned through supervised training on specific datasets, followed by DPO training. The model is designed to deliver high performance across a variety of tasks, including chat, solving mathematical problems, and text generation. It is built on the Transformers library, supports PyTorch, and is licensed under the Apache 2.0 license.
Chatbot
46.1K
RMBG-2.0
RMBG 2.0
RMBG-2.0 is a background removal model developed by BRIA AI, aimed at effectively separating the foreground and background in images. The model is trained on a curated dataset including general stock images, e-commerce, gaming, and advertising content, making it suitable for commercial use and capable of driving large-scale content creation for enterprises. Its accuracy, efficiency, and versatility are comparable to leading open-source models. RMBG-2.0 is available as source code for non-commercial use.
Background Removal
73.4K
LLaMA-O1
Llama O1
LLaMA-O1 is a large inference model framework that integrates Monte Carlo Tree Search (MCTS), self-reinforcement learning, Proximal Policy Optimization (PPO), and draws from the dual strategy paradigm of AlphaGo Zero alongside large language models. This model primarily targets Olympic-level mathematical reasoning problems, providing an open platform for training, inference, and evaluation. According to product background information, this is an individual experimental project and is not affiliated with any third-party organizations or institutions.
Research Instruments
48.3K
Sparsh
Sparsh
Sparsh is a series of general tactile representations trained through self-supervised algorithms such as MAE, DINO, and JEPA. It can generate useful representations for DIGIT, Gelsight'17, and Gelsight Mini, significantly outperforming end-to-end models on downstream tasks proposed by TacBench while supporting data-efficient training for new downstream tasks. The Sparsh project includes PyTorch implementations, pre-trained models, and datasets released alongside Sparsh.
Research Instruments
44.4K
Meta Lingua
Meta Lingua
Meta Lingua is a lightweight and efficient library for training and inference of large language models (LLMs) designed specifically for research purposes. It utilizes easy-to-modify PyTorch components, enabling researchers to experiment with new architectures, loss functions, and datasets. The library aims to facilitate end-to-end training, inference, and evaluation, providing tools for better understanding the speed and stability of the models. Although Meta Lingua is still under development, it already offers several sample applications demonstrating how to use this repository.
Model Training and Deployment
48.3K
Fresh Picks
torchao
Torchao
Torchao is a library for PyTorch focused on custom data types and optimization, supporting the quantization and sparsification of weights, gradients, optimizers, and activation functions for both inference and training. It is compatible with torch.compile() and FSDP2, enabling acceleration for most PyTorch models. Torchao aims to enhance model inference speed and memory efficiency while minimizing accuracy loss through techniques such as Quantization Aware Training (QAT) and Post Training Quantization (PTQ).
AI Development Assistant
57.7K
FluxMusic
Fluxmusic
FluxMusic is a text-to-music generation model implemented in PyTorch that explores a straightforward method for generating music from text prompts using a diffusion-based flow transformer. This model can create music segments based on textual cues, showcasing both innovation and technical complexity. It represents cutting-edge technology in the field of music generation, offering new possibilities for musical creativity.
AI music generation
61.0K
zero_to_gpt
Zero To Gpt
zero_to_gpt is a tutorial aimed at helping users learn deep learning from the ground up, ultimately enabling them to train their own GPT models. As AI technologies emerge from labs and find wide applications across various industries, the demand for professionals who can understand and apply AI is increasing. This tutorial integrates theory and practice by addressing real-world problems (such as weather prediction and language translation) to explore the theoretical foundations of deep learning, including gradient descent and backpropagation. The course content starts with basic neural network architectures and training methods, gradually advancing to complex topics such as transformers, GPU programming, and distributed training.
AI course
48.9K
Data-Juicer
Data Juicer
Data-Juicer is a comprehensive multimodal data processing system aimed at delivering higher quality, richer, and more digestible data for large language models (LLMs). It offers a systematic and reusable data processing library, supports collaborative development between data and models, allows rapid iteration through a sandbox lab, and provides features like data and model feedback loops, visualization, and multidimensional automated evaluation, helping users better understand and improve their data and models. Data-Juicer is actively maintained and regularly enhanced with more features, data recipes, and datasets.
AI Data Mining
62.4K
ml-mdm
Ml Mdm
ml-mdm is a Python package designed for the efficient training of high-quality text-to-image diffusion models. Utilizing Matryoshka diffusion model technology, it can train a single pixel-space model at a resolution of 1024x1024 pixels, demonstrating impressive zero-shot generalization capabilities.
AI image generation
54.9K
AuraSR-v2
Aurasr V2
AuraSR-v2 is a GAN-based image super-resolution model designed for enlarging generated images, and it is a variant of the GigaGAN paper. The PyTorch implementation of this model is based on the unofficial lucidrains/gigagan-pytorch repository. It significantly enhances the resolution of images while preserving their quality, which is particularly important for applications requiring high-definition image outputs.
AI image enhancement
63.2K
DiT-MoE
Dit MoE
DiT-MoE is a diffusion transformer model implemented in PyTorch that can scale up to 16 billion parameters while competing with dense networks and demonstrating highly optimized inference capabilities. It represents cutting-edge technology in deep learning for handling large-scale datasets, carrying significant research and application value.
AI Model
50.8K
ComfyUI-Fast-Style-Transfer
Comfyui Fast Style Transfer
ComfyUI-Fast-Style-Transfer is a rapid neural style transfer plugin developed based on the PyTorch framework. It allows users to achieve image style conversion through simple operations. This plugin is based on the fast-neural-style-pytorch project and currently only ports the basic inference functionality. Users can customize styles and achieve unique style transfer effects by training their own models.
AI image generation
54.9K
Fresh Picks
ToucanTTS
Toucantts
Developed by the Natural Language Processing Institute at Stuttgart University in Germany, ToucanTTS is a multilingual and controllable text-to-speech synthesis toolkit. Built using pure Python and PyTorch, it strives to maintain simplicity and ease of use while being as powerful as possible. The toolkit supports teaching, training, and using cutting-edge speech synthesis models, offering high flexibility and customizability, making it suitable for both education and research.
AI text translation and audio
83.1K
AudioLCM
Audiolcm
AudioLCM is a text-to-audio generation model implemented in PyTorch. It generates high-quality and efficient audio using a latent consistency model. Developed by Huadai Liu and others, it provides an open-source implementation and pre-trained models. It can convert text descriptions into near-realistic audio, holding significant application value, particularly in areas like speech synthesis and audio production.
AI text translation and voice
83.9K
kan-gpt
Kan Gpt
kan-gpt is a PyTorch-based implementation of Generative Pre-trained Transformers (GPTs) that employs Kolmogorov-Arnold Networks (KANs) for language modeling. The model demonstrates potential in text generation tasks, particularly in handling long-range dependencies. Its significance lies in providing a new model architecture for the field of natural language processing, which can enhance the performance of language models.
AI Model
53.3K
LeRobot
Lerobot
LeRobot is an open-source project aimed at lowering the barriers to entry into the field of robotics, enabling everyone to contribute and benefit from shared datasets and pre-trained models. It includes the most advanced methods verified in the real world, with a special focus on imitation learning and reinforcement learning. LeRobot provides a set of pre-trained models, datasets with human-collected demonstration videos, and simulation environments, allowing users to begin without assembling a robot. In the coming weeks, we plan to add support for the most affordable and capable real-world robots.
AI Development Assistant
102.4K
contrastors
Contrastors
contrastors is a contrastive learning toolkit that enables researchers and engineers to efficiently train and evaluate contrastive models. Built upon Flash Attention, it supports multi-GPU training, GradCache support for large batch training in memory-constrained environments. It also supports Huggingface, allowing for seamless loading of common models. Additionally, it supports masked language modeling pretraining and Matryoshka representation learning.
AI Model
63.8K
stable-audio-tools
Stable Audio Tools
stable-audio-tools is an open-source PyTorch library that provides training and inference code for generative models used in conditional audio generation. It includes automatic encoders, implicit diffusion models, MusicGen, and more. Support for multi-GPU training allows generation of high-quality audio.
AI Music Generation
75.6K
honeybee
Honeybee
Honeybee is a local-enhancement predictor for multimodal language models. It enhances the performance of multimodal language models on various downstream tasks, such as natural language inference and visual question answering. The advantage of Honeybee lies in the introduction of a local perception mechanism, which can better model the dependencies between input samples, thereby strengthening the inference and question-answering abilities of the multimodal language model.
AI Model
57.7K
SIFU
SIFU
SIFU is a method for reconstructing high-quality 3D virtual human models from lateral images. Its core innovation lies in proposing a new implicit function based on lateral images, which enhances feature extraction and improves geometric accuracy. Additionally, SIFU introduces a 3D consistent texture optimization process that significantly enhances texture quality and enables texture editing through a text-to-image diffusion model. SIFU excels in handling complex poses and loose clothing, making it an ideal solution for practical applications.
AI image generation
69.8K
MLX
MLX
MLX is a NumPy-like array framework designed for efficient and flexible machine learning on Apple silicon, provided by the Apple Machine Learning Research team. Its Python API closely resembles NumPy, though some exceptions exist. MLX also boasts a complete C++ API that closely adheres to the Python API. Key differences between MLX and NumPy include composable function transforms, lazy computation, and multi-device support. MLX draws inspiration from frameworks like PyTorch, Jax, and ArrayFire. Unlike these frameworks, MLX utilizes a unified memory model. Arrays in MLX reside in shared memory, enabling operations to be performed on any supported device type (CPU, GPU, etc.) without data copying.
AI Development Assistant
76.2K
YOLO-NAS Pose
YOLO NAS Pose
YOLO-NAS Pose is a free and open-source library designed for training PyTorch-based computer vision models. It offers training scripts and examples for easily replicating model results. Featuring state-of-the-art (SOTA) models, it allows you to effortlessly load and fine-tune production-ready pre-trained models with optimized best practices and validated hyperparameters for achieving superior accuracy. This streamlines the training process and eliminates guesswork. It provides models for various tasks, including classification, detection, and segmentation, making it seamlessly integrable into your codebase.
Model Training and Deployment
95.8K
English Picks
Lightning AI
Lightning AI
Lightning AI is a platform based on PyTorch that allows users to easily train and deploy AI models between local machines and cloud environments. It supports the construction of various popular AI models such as large language models, Transformers, and Stable Diffusion. Key features include support for distributed multi-GPU training, built-in MLOps capabilities, and serverless cloud deployment. It is suitable for AI research teams, companies aiming to rapidly develop AI products, and institutions that have GPU resources.
Development & Tools
81.4K
RunPod
Runpod
RunPod is a scalable cloud GPU infrastructure for training and inference. Rent cloud GPUs starting at $0.2 per hour, with support for TensorFlow, PyTorch, and other AI frameworks. We provide reliable cloud services, free bandwidth, multiple GPU options, server endpoints, and AI endpoints to suit various needs.
Development & Tools
65.4K
Featured AI Tools
Flow AI
Flow AI
Flow is an AI-driven movie-making tool designed for creators, utilizing Google DeepMind's advanced models to allow users to easily create excellent movie clips, scenes, and stories. The tool provides a seamless creative experience, supporting user-defined assets or generating content within Flow. In terms of pricing, the Google AI Pro and Google AI Ultra plans offer different functionalities suitable for various user needs.
Video Production
43.1K
NoCode
Nocode
NoCode is a platform that requires no programming experience, allowing users to quickly generate applications by describing their ideas in natural language, aiming to lower development barriers so more people can realize their ideas. The platform provides real-time previews and one-click deployment features, making it very suitable for non-technical users to turn their ideas into reality.
Development Platform
44.7K
ListenHub
Listenhub
ListenHub is a lightweight AI podcast generation tool that supports both Chinese and English. Based on cutting-edge AI technology, it can quickly generate podcast content of interest to users. Its main advantages include natural dialogue and ultra-realistic voice effects, allowing users to enjoy high-quality auditory experiences anytime and anywhere. ListenHub not only improves the speed of content generation but also offers compatibility with mobile devices, making it convenient for users to use in different settings. The product is positioned as an efficient information acquisition tool, suitable for the needs of a wide range of listeners.
AI
42.5K
MiniMax Agent
Minimax Agent
MiniMax Agent is an intelligent AI companion that adopts the latest multimodal technology. The MCP multi-agent collaboration enables AI teams to efficiently solve complex problems. It provides features such as instant answers, visual analysis, and voice interaction, which can increase productivity by 10 times.
Multimodal technology
43.3K
Chinese Picks
Tencent Hunyuan Image 2.0
Tencent Hunyuan Image 2.0
Tencent Hunyuan Image 2.0 is Tencent's latest released AI image generation model, significantly improving generation speed and image quality. With a super-high compression ratio codec and new diffusion architecture, image generation speed can reach milliseconds, avoiding the waiting time of traditional generation. At the same time, the model improves the realism and detail representation of images through the combination of reinforcement learning algorithms and human aesthetic knowledge, suitable for professional users such as designers and creators.
Image Generation
42.2K
OpenMemory MCP
Openmemory MCP
OpenMemory is an open-source personal memory layer that provides private, portable memory management for large language models (LLMs). It ensures users have full control over their data, maintaining its security when building AI applications. This project supports Docker, Python, and Node.js, making it suitable for developers seeking personalized AI experiences. OpenMemory is particularly suited for users who wish to use AI without revealing personal information.
open source
42.8K
FastVLM
Fastvlm
FastVLM is an efficient visual encoding model designed specifically for visual language models. It uses the innovative FastViTHD hybrid visual encoder to reduce the time required for encoding high-resolution images and the number of output tokens, resulting in excellent performance in both speed and accuracy. FastVLM is primarily positioned to provide developers with powerful visual language processing capabilities, applicable to various scenarios, particularly performing excellently on mobile devices that require rapid response.
Image Processing
41.7K
Chinese Picks
LiblibAI
Liblibai
LiblibAI is a leading Chinese AI creative platform offering powerful AI creative tools to help creators bring their imagination to life. The platform provides a vast library of free AI creative models, allowing users to search and utilize these models for image, text, and audio creations. Users can also train their own AI models on the platform. Focused on the diverse needs of creators, LiblibAI is committed to creating inclusive conditions and serving the creative industry, ensuring that everyone can enjoy the joy of creation.
AI Model
6.9M
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase