zurk / hunyuan-video-8bit

Hunyuan Video 8bit model API for video generation

  • Public
  • 238 runs
  • L40S
  • GitHub
Iterate in playground

Input

string
Shift + Return to add a new line

Input prompt for video generation

Default: "A cat walks on the grass, realistic style."

string
Shift + Return to add a new line

Negative prompt

Default: "Ugly"

integer
(minimum: 128, maximum: 1920)

Video width

Default: 960

integer
(minimum: 128, maximum: 1080)

Video height

Default: 544

integer
(minimum: 16, maximum: 256)

Number of frames

Default: 65

integer
(minimum: 1, maximum: 100)

Number of inference steps

Default: 40

number
(minimum: 1, maximum: 20)

Embedded guidance scale

Default: 6

number
(minimum: 0, maximum: 20)

Flow shift parameter

Default: 7

integer

Random seed

Output

Generated in

Run time and cost

This model costs approximately $0.27 to run on Replicate, or 3 runs per $1, but this varies depending on your inputs. It is also open source and you can run it on your own computer with Docker.

This model runs on Nvidia L40S GPU hardware. Predictions typically complete within 5 minutes. The predict time for this model varies significantly based on the inputs.

Readme

Here’s the improved version of your README:


HunyuanVideo API (8-bit Version)

HunyuanVideo is a cutting-edge text-to-video generation model capable of creating high-quality videos from text descriptions. It surpasses many closed-source alternatives in text alignment, motion quality, and overall visual fidelity.

This API provides access to the 8-bit version of the model, which is optimized for performance on less expensive GPUs and offers faster inference compared to the full HunyuanVideo model.

Examples

import replicate

output = replicate.run(
    "zurk/hunyuan-video-8bit:main",
    input={
        "prompt": "A cat walks on the grass, realistic style.",
        "negative_prompt": "Ugly",
        "width": 960,
        "height": 544,
        "video_length": 65,
        "embedded_guidance_scale": 6.0,
        "num_inference_steps": 40,
        "seed": 43,
    }
)

Parameters

  • prompt (string, required): Text description of the video you want to generate.
  • negative_prompt (string, optional): Text describing elements you want to exclude from the video.
  • width (integer, default: 960): Video width in pixels.
  • height (integer, default: 544): Video height in pixels.
  • video_length (integer, default: 65): Number of frames (maximum 129).
  • seed (integer, optional): Random seed for reproducibility. If not specified, you can find its value in the logs.
  • embedded_guidance_scale (float, default: 6.0): Scale for embedded guidance during generation.
  • num_inference_steps (integer, default: 40): Number of denoising steps.
  • flow_shift (float, default: 7.0): Parameter for motion control (flow shift).

Limitations

  • The maximum video length is 129 frames (approximately 5.3 seconds).
  • The video_length parameter must follow the formula 4*n+1 (e.g., 17, 21, 25, etc.).

Feedback

If you encounter any issues while using this API, please report them by creating an issue at GitHub Issues. I will address them as soon as possible.

For further details, visit the HunyuanVideo GitHub repository or explore the ComfyUI wrapper nodes for HunyuanVideo.