Collections

Generate videos

These models can generate and edit videos from text prompts and images. They use advanced AI techniques like diffusion models and latent space interpolation to create high-quality, controllable video content.

Key capabilities:

  • Text-to-video generation - Convert text prompts into video clips and animations. Useful for quickly prototyping video concepts.
  • Image-to-video generation - Animate still images into video.
  • Inpainting for infinite zoom - Use image inpainting to extrapolate video frames and create infinite zoom effects.
  • Stylization - Apply artistic filters like cartoonification to give videos a unique look and feel.

State of the art: minimax/video-01

For most people looking to generate custom videos from text or image prompts, we recommend minimax/video-01, also known as Hailuo . This model supports the generation of high-definition videos at 720p resolution and 25fps, featuring cinematic camera movement effects. It can quickly create visually striking content based on text descriptions.

Open source: tencent/hunyuan-video

The Hunyuan Video model by Tencent is an excellent open-source option, competitive with the best proprietary video models. Try adjusting the number of steps used for each frame to trade off between generation speed and detail.

Other rankings

Generative video is a rapidly advancing field. Check out the arena and leaderboard at Artificial Analysis to see what’s popular today.

Featured models

pixverse / pixverse-v4

Quickly generate smooth 5s or 8s videos at 540p, 720p or 1080p

3.6K runs

wavespeedai / wan-2.1-t2v-480p

Accelerated inference for Wan 2.1 14B text to video, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

77.8K runs

wavespeedai / wan-2.1-t2v-720p

Accelerated inference for Wan 2.1 14B text to video with high resolution, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

29.6K runs

wavespeedai / wan-2.1-i2v-480p

Accelerated inference for Wan 2.1 14B image to video, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

185K runs

wavespeedai / wan-2.1-i2v-720p

Accelerated inference for Wan 2.1 14B image to video with high resolution, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation.

38.6K runs

google / veo-2

State of the art video generation model. Veo 2 can faithfully follow simple and complex instructions, and convincingly simulates real-world physics as well as a wide range of visual styles.

40.3K runs

minimax / video-01-director

Generate videos with specific camera movements

23.8K runs

minimax / video-01

Generate 6s videos with prompts or images. (Also known as Hailuo). Use a subject reference to make a video with a character and the S2V-01 model.

462.9K runs

minimax / video-01-live

An image-to-video (I2V) model specifically trained for Live2D and general animation use cases

103K runs

Recommended models

luma / ray-flash-2-720p

Generate 5s and 9s 720p videos, faster and cheaper than Ray 2

11.2K runs

luma / ray-flash-2-540p

Generate 5s and 9s 540p videos, faster and cheaper than Ray 2

9.3K runs

luma / ray-2-720p

Generate 5s and 9s 720p videos

10.6K runs

luma / ray-2-540p

Generate 5s and 9s 540p videos

2.9K runs

kwaivgi / kling-v2.0

Generate 5s and 10s videos in 720p resolution

9K runs

kwaivgi / kling-v1.6-pro

Generate 5s and 10s videos in 1080p resolution

253.7K runs

wan-video / wan-2.1-1.3b

Generate 5s 480p videos. Wan is an advanced and powerful visual generation model developed by Tongyi Lab of Alibaba Group

14.6K runs

tencent / hunyuan-video

A state-of-the-art text-to-video generation model capable of creating high-quality videos with realistic motion from text descriptions

91.2K runs

kwaivgi / kling-v1.6-standard

Generate 5s and 10s videos in 720p resolution

500.4K runs

lightricks / ltx-video

LTX-Video is the first DiT-based video generation model capable of generating high-quality videos in real-time. It produces 24 FPS videos at a 768x512 resolution faster than they can be watched.

85.3K runs

luma / ray

Fast, high quality text-to-video and image-to-video (Also known as Dream Machine)

30.8K runs

zsxkib / hunyuan-video2video

A state-of-the-art text-to-video generation model capable of creating high-quality videos with realistic motion from text descriptions

2.4K runs

haiper-ai / haiper-video-2

Generate 4s and 6s videos from a prompt or image

10.1K runs

genmoai / mochi-1

Mochi 1 preview is an open video generation model with high-fidelity motion and strong prompt adherence in preliminary evaluation

2.6K runs

zsxkib / samurai

SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memory

109 runs

zsxkib / pyramid-flow

Text-to-Video + Image-to-Video: Pyramid Flow Autoregressive Video Generation method based on Flow Matching

8.4K runs

cuuupid / cogvideox-5b

Generate high quality videos from a prompt

2.1K runs

meta / sam-2-video

SAM 2: Segment Anything v2 (for videos)

5.3K runs

fofr / tooncrafter

Create videos from illustrated input images

50.2K runs

fofr / video-morpher

Generate a video that morphs between subjects, with an optional style

14.5K runs

cjwbw / videocrafter

VideoCrafter2: Text-to-Video and Image-to-Video Generation and Editing

74.8K runs

ali-vilab / i2vgen-xl

RESEARCH/NON-COMMERCIAL USE ONLY: High-Quality Image-to-Video Synthesis via Cascaded Diffusion Models

127K runs

open-mmlab / pia

Personalized Image Animator

103.3K runs

zsxkib / animatediff-illusions

Monster Labs' Controlnet QR Code Monster v2 For SD-1.5 on top of AnimateDiff Prompt Travel (Motion Module SD 1.5 v2)

10.2K runs

lucataco / hotshot-xl

😊 Hotshot-XL is an AI text-to-GIF model trained to work alongside Stable Diffusion XL

642.7K runs

zsxkib / animatediff-prompt-travel

🎨AnimateDiff Prompt Travel🧭 Seamlessly Navigate and Animate Between Text-to-Image Prompts for Dynamic Visual Narratives

5.7K runs

zsxkib / animate-diff

🎨 AnimateDiff (w/ MotionLoRAs for Panning, Zooming, etc): Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning

57.3K runs

lucataco / animate-diff

Animate Your Personalized Text-to-Image Diffusion Models

297.8K runs

anotherjesse / zeroscope-v2-xl

Zeroscope V2 XL & 576w

288.4K runs

cjwbw / controlvideo

Training-free Controllable Text-to-Video Generation

2.4K runs

cjwbw / text2video-zero

Text-to-Image Diffusion Models are Zero-Shot Video Generators

41.7K runs

cjwbw / damo-text-to-video

Multi-stage text-to-video generation

146.6K runs

andreasjansson / tile-morph

Create tileable animations with seamless transitions

529.3K runs

pollinations / stable-diffusion-dance

Audio Reactive Stable Diffusion

5.7K runs

arielreplicate / deoldify_​video

Add colours to old video footage.

5K runs

pollinations / real-basicvsr-video-superresolution

RealBasicVSR: Investigating Tradeoffs in Real-World Video Super-Resolution

9.1K runs

pollinations / tune-a-video

About Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation

2.9K runs

arielreplicate / robust_​video_​matting

extract foreground of a video

55.2K runs

arielreplicate / stable_​diffusion_​infinite_​zoom

Use Runway's Stable-diffusion inpainting model to create an infinite loop video

38.2K runs

andreasjansson / stable-diffusion-animation

Animate Stable Diffusion by interpolating between two prompts

118.9K runs

nateraw / stable-diffusion-videos

Generate videos by interpolating the latent space of Stable Diffusion

58.4K runs

deforum / deforum_​stable_​diffusion

Animating prompts with stable diffusion

263.8K runs