Generate videos
These models can generate and edit videos from text prompts and images. They use advanced AI techniques like diffusion models and latent space interpolation to create high-quality, controllable video content.
Key capabilities:
- Text-to-video generation - Convert text prompts into video clips and animations. Useful for quickly prototyping video concepts.
- Image-to-video generation - Animate still images into video.
- Inpainting for infinite zoom - Use image inpainting to extrapolate video frames and create infinite zoom effects.
- Stylization - Apply artistic filters like cartoonification to give videos a unique look and feel.
State of the art: minimax/video-01
For most people looking to generate custom videos from text or image prompts, we recommend minimax/video-01, also known as Hailuo . This model supports the generation of high-definition videos at 720p resolution and 25fps, featuring cinematic camera movement effects. It can quickly create visually striking content based on text descriptions.
Open source: tencent/hunyuan-video
The Hunyuan Video model by Tencent is an excellent open-source option, competitive with the best proprietary video models. Try adjusting the number of steps used for each frame to trade off between generation speed and detail.
Other rankings
Generative video is a rapidly advancing field. Check out the arena and leaderboard at Artificial Analysis to see what’s popular today.
Featured models
lightricks / ltx-video
LTX-Video is the first DiT-based video generation model capable of generating high-quality videos in real-time. It produces 24 FPS videos at a 768x512 resolution faster than they can be watched.
minimax / video-01-live
An image-to-video (I2V) model specifically trained for Live2D and general animation use cases
luma / ray
Fast, high quality text-to-video and image-to-video (Also known as Dream Machine)
tencent / hunyuan-video
A state-of-the-art text-to-video generation model capable of creating high-quality videos with realistic motion from text descriptions
haiper-ai / haiper-video-2
Generate 4s and 6s videos from a prompt or image
genmoai / mochi-1
Mochi 1 preview is an open video generation model with high-fidelity motion and strong prompt adherence in preliminary evaluation
minimax / video-01
Generate 6s videos with prompts or images. (Also known as Hailuo)
Recommended models
zsxkib / hunyuan-video2video
A state-of-the-art text-to-video generation model capable of creating high-quality videos with realistic motion from text descriptions
zsxkib / samurai
SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memory
zsxkib / pyramid-flow
Text-to-Video + Image-to-Video: Pyramid Flow Autoregressive Video Generation method based on Flow Matching
cuuupid / cogvideox-5b
Generate high quality videos from a prompt
meta / sam-2-video
SAM 2: Segment Anything v2 (for videos)
fofr / tooncrafter
Create videos from illustrated input images
fofr / video-morpher
Generate a video that morphs between subjects, with an optional style
camenduru / animatediff-lightning-4-step
AnimateDiff-Lightning: Cross-Model Diffusion Distillation
cjwbw / videocrafter
VideoCrafter2: Text-to-Video and Image-to-Video Generation and Editing
ali-vilab / i2vgen-xl
RESEARCH/NON-COMMERCIAL USE ONLY: High-Quality Image-to-Video Synthesis via Cascaded Diffusion Models
open-mmlab / pia
Personalized Image Animator
zsxkib / animatediff-illusions
Monster Labs' Controlnet QR Code Monster v2 For SD-1.5 on top of AnimateDiff Prompt Travel (Motion Module SD 1.5 v2)
lucataco / hotshot-xl
😊 Hotshot-XL is an AI text-to-GIF model trained to work alongside Stable Diffusion XL
zsxkib / animatediff-prompt-travel
🎨AnimateDiff Prompt Travel🧭 Seamlessly Navigate and Animate Between Text-to-Image Prompts for Dynamic Visual Narratives
zsxkib / animate-diff
🎨 AnimateDiff (w/ MotionLoRAs for Panning, Zooming, etc): Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning
lucataco / animate-diff
Animate Your Personalized Text-to-Image Diffusion Models
anotherjesse / zeroscope-v2-xl
Zeroscope V2 XL & 576w
cjwbw / controlvideo
Training-free Controllable Text-to-Video Generation
cjwbw / text2video-zero
Text-to-Image Diffusion Models are Zero-Shot Video Generators
cjwbw / damo-text-to-video
Multi-stage text-to-video generation
andreasjansson / tile-morph
Create tileable animations with seamless transitions
pollinations / stable-diffusion-dance
Audio Reactive Stable Diffusion
arielreplicate / deoldify_video
Add colours to old video footage.
pollinations / real-basicvsr-video-superresolution
RealBasicVSR: Investigating Tradeoffs in Real-World Video Super-Resolution
pollinations / tune-a-video
About Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation
arielreplicate / robust_video_matting
extract foreground of a video
arielreplicate / stable_diffusion_infinite_zoom
Use Runway's Stable-diffusion inpainting model to create an infinite loop video
andreasjansson / stable-diffusion-animation
Animate Stable Diffusion by interpolating between two prompts
nateraw / stable-diffusion-videos
Generate videos by interpolating the latent space of Stable Diffusion
deforum / deforum_stable_diffusion
Animating prompts with stable diffusion