harmonai / dance-diffusion

Tools to train a generative model on arbitrary audio samples

  • Public
  • 5.2K runs
  • T4
  • GitHub
  • Paper
  • License

Input

string

Model

Default: "maestro-150k"

number

Number of seconds to generate

Default: 8

integer

How many samples to generate

Default: 1

integer

Number of steps, higher numbers will give more refined output but will take longer. The maximum is 150.

Default: 100

Output

Video Player is loading.
Current Time 00:00:000
Duration 00:00:000
Loaded: 0%
Stream Type LIVE
Remaining Time 00:00:000
 
1x
Generated in

Run time and cost

This model costs approximately $0.12 to run on Replicate, or 8 runs per $1, but this varies depending on your inputs. It is also open source and you can run it on your own computer with Docker.

This model runs on Nvidia T4 GPU hardware. Predictions typically complete within 10 minutes. The predict time for this model varies significantly based on the inputs.

Readme

Dance Diffusion v0.10

Welcome to the Dance Diffusion beta!

Dance Diffusion is the first in a suite of generative audio tools for producers and musicians to be released by Harmonai. For more info or to get involved in the development of these tools, please visit https://harmonai.org and fill out the form on the front page.

Click here to ensure you are using the latest version

Audio diffusion tools in this notebook:

  • Unconditional random audio sample generation
  • Audio sample regeneration/style transfer using a single audio file
  • Audio interpolation between two audio files

Model ported to cog by Pollinations