

MCVD
Overview :
MCVD is a general-purpose model for video generation, prediction, and interpolation. It utilizes a score-based diffusion loss function to generate novel frames by injecting Gaussian noise into the current frame and conditioning on past and/or future frames for denoising. Training involves randomly masking past and/or future frames to achieve four capabilities: unconditional generation, future prediction, past reconstruction, and interpolation. The model employs a 2D convolutional U-Net architecture that conditions on past and future frames using concatenated or spatiotemporal adaptive normalization, resulting in high-quality and diverse video samples. Trained on 1-4 GPUs, it can be scaled to more channels. MCVD, a simple non-recursive 2D convolutional architecture, generates videos of arbitrary lengths and achieves SOTA results.
Target Users :
Video Generation, Prediction, and Interpolation
Use Cases
Movie Effect Generation
Video Game Development
Animation Production
Features
Video Generation
Video Prediction
Video Interpolation
Featured AI Tools

Sora
AI video generation
17.0M

Animate Anyone
Animate Anyone aims to generate character videos from static images driven by signals. Leveraging the power of diffusion models, we propose a novel framework tailored for character animation. To maintain consistency of complex appearance features present in the reference image, we design ReferenceNet to merge detailed features via spatial attention. To ensure controllability and continuity, we introduce an efficient pose guidance module to direct character movements and adopt an effective temporal modeling approach to ensure smooth cross-frame transitions between video frames. By extending the training data, our method can animate any character, achieving superior results in character animation compared to other image-to-video approaches. Moreover, we evaluate our method on benchmarks for fashion video and human dance synthesis, achieving state-of-the-art results.
AI video generation
11.4M