# Diffusion models

Text To Pose
text-to-pose is a research project aimed at generating character poses from text descriptions and using these poses to create images. This technology combines natural language processing and computer vision, achieving text-to-image generation by enhancing the control and quality of diffusion models. The project is based on a paper published at the NeurIPS 2024 Workshop, showcasing innovation and cutting-edge advancements. The key advantages of this technology include improved accuracy and controllability in image generation, as well as potential applications in artistic creation and virtual reality.
Image Generation
52.4K

Gamengen
GameNGen is a fully neural model-driven game engine capable of real-time interaction with complex environments while maintaining high quality over extended trajectories. It can interactively simulate the classic game 'DOOM' at over 20 frames per second, with its next-frame prediction achieving a PSNR of 29.4, comparable to lossy JPEG compression. Human evaluators only slightly outperform random chance in distinguishing between game clips and simulated clips. GameNGen is trained through two phases: (1) an RL-agent learns to play the game and records the actions and observations from the training sessions, which become the training data for the generative model; (2) a diffusion model is trained to predict the next frame, conditioned on the past actions and observation sequences. Conditional enhancement allows for stable autoregressive generation over long trajectories.
AI game creation
65.1K

Bootstrap3d
Bootstrap3D is a framework aimed at improving 3D content creation by leveraging synthetic data generation techniques to address the scarcity of high-quality 3D assets. It utilizes 2D and video diffusion models to generate multi-view images based on text prompts and employs the 3D-perceptive MV-LLaVA model to filter high-quality data and rewrite inaccurate titles. The framework has generated 1 million high-quality synthetic multi-view images with dense descriptive captions to alleviate the shortage of high-quality 3D data. It also introduces a Training Timestep Reschedule (TTR) strategy that leverages the denoising process to learn multi-view consistency while preserving the original 2D diffusion prior.
AI image generation
60.2K

Make An Audio 2
Make-An-Audio 2 is a text-to-audio generation technology based on diffusion models, co-developed by researchers from Zhejiang University, ByteDance, and the Chinese University of Hong Kong. This technology utilizes pre-trained large language models (LLMs) to parse text, optimizing for semantic alignment and temporal consistency, thereby improving the quality of generated audio. It also incorporates a feed-forward Transformer-based diffusion denoiser to enhance performance in generating variable-length audio and bolster the extraction of temporal information. Furthermore, by leveraging LLMs to convert abundant audio label data into audio-text datasets, the issue of time data scarcity is addressed.
AI Music Generation
52.7K
Chinese Picks

Instructvideo
InstructVideo is a method for training text-to-video diffusion models using reward fine-tuning guided by human feedback. It employs an editing-based reward fine-tuning approach, which reduces fine-tuning cost while enhancing efficiency. Leveraging pre-established image reward models, it provides reward signals through segment-wise sparse sampling and temporal decay rewards, significantly improving the visual quality of generated videos. InstructVideo not only enhances the visual quality of generated videos but also maintains strong generalization capabilities. For more information, please visit the official website.
AI video generation
131.4K

Freeu
FreeU is a method that can significantly improve the sampling quality of diffusion models without any additional cost: no training, no extra parameters, no increased memory or sampling time. It achieves this by reweighting the contributions of U-Net's skip connections and main branch feature maps, leveraging the advantages of both components of the U-Net architecture to enhance generation quality. Experiments on image and video generation tasks demonstrate that FreeU can be easily integrated into existing diffusion models such as Stable Diffusion, DreamBooth, ModelScope, Rerender, and ReVersion with just a few lines of code to improve generation quality.
Image Generation
62.1K
Featured AI Tools

Flow AI
Flow is an AI-driven movie-making tool designed for creators, utilizing Google DeepMind's advanced models to allow users to easily create excellent movie clips, scenes, and stories. The tool provides a seamless creative experience, supporting user-defined assets or generating content within Flow. In terms of pricing, the Google AI Pro and Google AI Ultra plans offer different functionalities suitable for various user needs.
Video Production
43.1K

Nocode
NoCode is a platform that requires no programming experience, allowing users to quickly generate applications by describing their ideas in natural language, aiming to lower development barriers so more people can realize their ideas. The platform provides real-time previews and one-click deployment features, making it very suitable for non-technical users to turn their ideas into reality.
Development Platform
45.5K

Listenhub
ListenHub is a lightweight AI podcast generation tool that supports both Chinese and English. Based on cutting-edge AI technology, it can quickly generate podcast content of interest to users. Its main advantages include natural dialogue and ultra-realistic voice effects, allowing users to enjoy high-quality auditory experiences anytime and anywhere. ListenHub not only improves the speed of content generation but also offers compatibility with mobile devices, making it convenient for users to use in different settings. The product is positioned as an efficient information acquisition tool, suitable for the needs of a wide range of listeners.
AI
43.3K

Minimax Agent
MiniMax Agent is an intelligent AI companion that adopts the latest multimodal technology. The MCP multi-agent collaboration enables AI teams to efficiently solve complex problems. It provides features such as instant answers, visual analysis, and voice interaction, which can increase productivity by 10 times.
Multimodal technology
44.2K
Chinese Picks

Tencent Hunyuan Image 2.0
Tencent Hunyuan Image 2.0 is Tencent's latest released AI image generation model, significantly improving generation speed and image quality. With a super-high compression ratio codec and new diffusion architecture, image generation speed can reach milliseconds, avoiding the waiting time of traditional generation. At the same time, the model improves the realism and detail representation of images through the combination of reinforcement learning algorithms and human aesthetic knowledge, suitable for professional users such as designers and creators.
Image Generation
43.6K

Openmemory MCP
OpenMemory is an open-source personal memory layer that provides private, portable memory management for large language models (LLMs). It ensures users have full control over their data, maintaining its security when building AI applications. This project supports Docker, Python, and Node.js, making it suitable for developers seeking personalized AI experiences. OpenMemory is particularly suited for users who wish to use AI without revealing personal information.
open source
43.6K

Fastvlm
FastVLM is an efficient visual encoding model designed specifically for visual language models. It uses the innovative FastViTHD hybrid visual encoder to reduce the time required for encoding high-resolution images and the number of output tokens, resulting in excellent performance in both speed and accuracy. FastVLM is primarily positioned to provide developers with powerful visual language processing capabilities, applicable to various scenarios, particularly performing excellently on mobile devices that require rapid response.
Image Processing
41.7K
Chinese Picks

Liblibai
LiblibAI is a leading Chinese AI creative platform offering powerful AI creative tools to help creators bring their imagination to life. The platform provides a vast library of free AI creative models, allowing users to search and utilize these models for image, text, and audio creations. Users can also train their own AI models on the platform. Focused on the diverse needs of creators, LiblibAI is committed to creating inclusive conditions and serving the creative industry, ensuring that everyone can enjoy the joy of creation.
AI Model
6.9M