You're looking at a specific version of this model. Jump to the model overview.

vaibhavs10 /incredibly-fast-whisper:c6433aab

Input

*file

Audio file

string

Task to perform: transcribe or translate to another language. (default: transcribe).

Default: "transcribe"

string

Language spoken in the audio, specify 'None' to perform language detection.

Default: "None"

integer

Number of parallel batches you want to compute. Reduce if you face OOMs. (default: 24).

Default: 24

string

Whisper supports both chunked as well as word level timestamps. (default: chunk).

Default: "chunk"

boolean

Use Pyannote.audio to diarise the audio clips. You will need to provide hf_token below too.

Default: false

string
Shift + Return to add a new line

Provide a hf.co/settings/token for Pyannote.audio to diarise the audio clips. You need to agree to the terms in 'https://huggingface.co/pyannote/speaker-diarization-3.1' and 'https://huggingface.co/pyannote/segmentation-3.0' first.

Output

No output yet! Press "Submit" to start a prediction.