ttsds
/
fishspeech_1_2
The Fish Speech V1.2 model.
Prediction
ttsds/fishspeech_1_2:0cfe0d652ead3df835da5a020063427419b28c66f60b322471fb5a456079659fIDj1jtgshehnrme0cmnqpvwcd8zrStatusSucceededSourceWebHardwareL40STotal durationCreatedInput
- text
- With tenure, Suzie'd have all the more leisure for yachting, but her publications are no good.
- text_reference
- and keeping eternity before the eyes, though much
- speaker_reference
- Video Player is loading.Current Time 00:00:000/Duration 00:00:000Loaded: 0%00:00:000Stream Type LIVERemaining Time -00:00:0001x
- Chapters
- descriptions off, selected
- captions settings, opens captions settings dialog
- captions off, selected
This is a modal window.
Beginning of dialog window. Escape will cancel and close the window.
End of dialog window.
{ "text": "With tenure, Suzie'd have all the more leisure for yachting, but her publications are no good.", "text_reference": "and keeping eternity before the eyes, though much", "speaker_reference": "https://replicate.delivery/pbxt/MNFXdPaUPOwYCZjZM4azsymbzE2TCV2WJXfGpeV2DrFWaSq8/example_en.wav" }
Install Replicate’s Node.js client library:npm install replicate
Import and set up the client:import Replicate from "replicate"; const replicate = new Replicate({ auth: process.env.REPLICATE_API_TOKEN, });
Run ttsds/fishspeech_1_2 using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
const output = await replicate.run( "ttsds/fishspeech_1_2:0cfe0d652ead3df835da5a020063427419b28c66f60b322471fb5a456079659f", { input: { text: "With tenure, Suzie'd have all the more leisure for yachting, but her publications are no good.", text_reference: "and keeping eternity before the eyes, though much", speaker_reference: "https://replicate.delivery/pbxt/MNFXdPaUPOwYCZjZM4azsymbzE2TCV2WJXfGpeV2DrFWaSq8/example_en.wav" } } ); // To access the file URL: console.log(output.url()); //=> "http://example.com" // To write the file to disk: fs.writeFile("my-image.png", output);
To learn more, take a look at the guide on getting started with Node.js.
Install Replicate’s Python client library:pip install replicate
Import the client:import replicate
Run ttsds/fishspeech_1_2 using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
output = replicate.run( "ttsds/fishspeech_1_2:0cfe0d652ead3df835da5a020063427419b28c66f60b322471fb5a456079659f", input={ "text": "With tenure, Suzie'd have all the more leisure for yachting, but her publications are no good.", "text_reference": "and keeping eternity before the eyes, though much", "speaker_reference": "https://replicate.delivery/pbxt/MNFXdPaUPOwYCZjZM4azsymbzE2TCV2WJXfGpeV2DrFWaSq8/example_en.wav" } ) print(output)
To learn more, take a look at the guide on getting started with Python.
Run ttsds/fishspeech_1_2 using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
curl -s -X POST \ -H "Authorization: Bearer $REPLICATE_API_TOKEN" \ -H "Content-Type: application/json" \ -H "Prefer: wait" \ -d $'{ "version": "0cfe0d652ead3df835da5a020063427419b28c66f60b322471fb5a456079659f", "input": { "text": "With tenure, Suzie\'d have all the more leisure for yachting, but her publications are no good.", "text_reference": "and keeping eternity before the eyes, though much", "speaker_reference": "https://replicate.delivery/pbxt/MNFXdPaUPOwYCZjZM4azsymbzE2TCV2WJXfGpeV2DrFWaSq8/example_en.wav" } }' \ https://api.replicate.com/v1/predictions
To learn more, take a look at Replicate’s HTTP API reference docs.
Output
Video Player is loading.Current Time 00:00:000/Duration 00:00:000Loaded: 0%00:00:000Stream Type LIVERemaining Time -00:00:0001x- Chapters
- descriptions off, selected
- captions settings, opens captions settings dialog
- captions off, selected
This is a modal window.
Beginning of dialog window. Escape will cancel and close the window.
End of dialog window.
{ "completed_at": "2025-01-28T17:21:06.102323Z", "created_at": "2025-01-28T17:19:49.901000Z", "data_removed": false, "error": null, "id": "j1jtgshehnrme0cmnqpvwcd8zr", "input": { "text": "With tenure, Suzie'd have all the more leisure for yachting, but her publications are no good.", "text_reference": "and keeping eternity before the eyes, though much", "speaker_reference": "https://replicate.delivery/pbxt/MNFXdPaUPOwYCZjZM4azsymbzE2TCV2WJXfGpeV2DrFWaSq8/example_en.wav" }, "logs": "2025-01-28 17:21:00.785 | INFO | tools.llama.generate:generate_long:432 - Encoded text: With tenure, Suzie'd have all the more leisure for yachting, but her publications are no good.\n2025-01-28 17:21:00.785 | INFO | tools.llama.generate:generate_long:450 - Generating sentence 1/1 of sample 1/1\n 0%| | 0/3892 [00:00<?, ?it/s]/root/.pyenv/versions/3.11.10/lib/python3.11/site-packages/torch/backends/cuda/__init__.py:342: FutureWarning: torch.backends.cuda.sdp_kernel() is deprecated. In the future, this context manager will be removed. Please see, torch.nn.attention.sdpa_kernel() for the new context manager, with updated signature.\nwarnings.warn(\n 0%| | 6/3892 [00:00<01:12, 53.45it/s]\n 0%| | 12/3892 [00:00<01:10, 54.65it/s]\n 0%| | 18/3892 [00:00<01:10, 55.23it/s]\n 1%| | 24/3892 [00:00<01:09, 55.57it/s]\n 1%| | 30/3892 [00:00<01:09, 55.69it/s]\n 1%| | 36/3892 [00:00<01:09, 55.78it/s]\n 1%| | 42/3892 [00:00<01:10, 54.58it/s]\n 1%| | 48/3892 [00:00<01:11, 53.49it/s]\n 1%|▏ | 54/3892 [00:01<01:15, 50.94it/s]\n 2%|▏ | 60/3892 [00:01<01:18, 49.01it/s]\n 2%|▏ | 66/3892 [00:01<01:15, 50.55it/s]\n 2%|▏ | 72/3892 [00:01<01:13, 51.68it/s]\n 2%|▏ | 78/3892 [00:01<01:13, 51.75it/s]\n 2%|▏ | 84/3892 [00:01<01:13, 51.73it/s]\n 2%|▏ | 90/3892 [00:01<01:14, 50.99it/s]\n 2%|▏ | 96/3892 [00:01<01:14, 50.77it/s]\n 3%|▎ | 102/3892 [00:01<01:14, 50.97it/s]\n 3%|▎ | 108/3892 [00:02<01:13, 51.26it/s]\n 3%|▎ | 114/3892 [00:02<01:13, 51.37it/s]\n 3%|▎ | 120/3892 [00:02<01:13, 51.29it/s]\n 3%|▎ | 126/3892 [00:02<01:11, 52.36it/s]\n 3%|▎ | 132/3892 [00:02<01:10, 53.35it/s]\n 4%|▎ | 138/3892 [00:02<01:09, 54.03it/s]\n 4%|▎ | 144/3892 [00:02<01:09, 54.20it/s]\n 4%|▍ | 150/3892 [00:02<01:08, 54.64it/s]\n 4%|▍ | 156/3892 [00:02<01:08, 54.90it/s]\n 4%|▍ | 162/3892 [00:03<01:07, 55.14it/s]\n 4%|▍ | 168/3892 [00:03<01:07, 55.32it/s]\n 4%|▍ | 174/3892 [00:03<01:07, 55.43it/s]\n 5%|▍ | 180/3892 [00:03<01:06, 55.55it/s]\n 5%|▍ | 186/3892 [00:03<01:06, 55.57it/s]\n 5%|▍ | 192/3892 [00:03<01:06, 55.63it/s]\n 5%|▌ | 198/3892 [00:03<01:07, 55.06it/s]\n 5%|▌ | 204/3892 [00:03<01:06, 55.13it/s]\n 5%|▌ | 210/3892 [00:03<01:06, 54.96it/s]\n 6%|▌ | 216/3892 [00:04<01:06, 55.16it/s]\n 6%|▌ | 222/3892 [00:04<01:06, 55.38it/s]\n 6%|▌ | 228/3892 [00:04<01:06, 55.31it/s]\n 6%|▌ | 234/3892 [00:04<01:05, 55.47it/s]\n 6%|▌ | 240/3892 [00:04<01:05, 55.52it/s]\n 6%|▋ | 246/3892 [00:04<01:05, 55.37it/s]\n 6%|▋ | 252/3892 [00:04<01:05, 55.25it/s]\n7%|▋ | 255/3892 [00:04<01:07, 53.56it/s]\n2025-01-28 17:21:05.732 | INFO | tools.llama.generate:generate_long:505 - Generated 257 tokens in 4.95 seconds, 51.95 tokens/sec\n2025-01-28 17:21:05.732 | INFO | tools.llama.generate:generate_long:508 - Bandwidth achieved: 25.47 GB/s\n2025-01-28 17:21:05.732 | INFO | tools.llama.generate:generate_long:513 - GPU Memory used: 1.56 GB\n/root/.pyenv/versions/3.11.10/lib/python3.11/site-packages/torch/nn/modules/conv.py:306: UserWarning: Plan failed with a cudnnException: CUDNN_BACKEND_EXECUTION_PLAN_DESCRIPTOR: cudnnFinalize Descriptor Failed cudnn_status: CUDNN_STATUS_NOT_SUPPORTED (Triggered internally at ../aten/src/ATen/native/cudnn/Conv_v8.cpp:919.)\nreturn F.conv1d(input, weight, bias, self.stride,\nNext sample", "metrics": { "predict_time": 5.640252315, "total_time": 76.201323 }, "output": "https://replicate.delivery/xezq/I6kOPqNuL0afek6we5chuQZEuJmVZ085jqrGiqTWWWdEgYToA/generated.wav", "started_at": "2025-01-28T17:21:00.462070Z", "status": "succeeded", "urls": { "stream": "https://stream.replicate.com/v1/files/bsvm-lxu4pk5mr37uzc56fncm4ixffl2ufkith6wcmxh7ddjomvjwnpwa", "get": "https://api.replicate.com/v1/predictions/j1jtgshehnrme0cmnqpvwcd8zr", "cancel": "https://api.replicate.com/v1/predictions/j1jtgshehnrme0cmnqpvwcd8zr/cancel" }, "version": "0cfe0d652ead3df835da5a020063427419b28c66f60b322471fb5a456079659f" }
Generated in2025-01-28 17:21:00.785 | INFO | tools.llama.generate:generate_long:432 - Encoded text: With tenure, Suzie'd have all the more leisure for yachting, but her publications are no good. 2025-01-28 17:21:00.785 | INFO | tools.llama.generate:generate_long:450 - Generating sentence 1/1 of sample 1/1 0%| | 0/3892 [00:00<?, ?it/s]/root/.pyenv/versions/3.11.10/lib/python3.11/site-packages/torch/backends/cuda/__init__.py:342: FutureWarning: torch.backends.cuda.sdp_kernel() is deprecated. In the future, this context manager will be removed. Please see, torch.nn.attention.sdpa_kernel() for the new context manager, with updated signature. warnings.warn( 0%| | 6/3892 [00:00<01:12, 53.45it/s] 0%| | 12/3892 [00:00<01:10, 54.65it/s] 0%| | 18/3892 [00:00<01:10, 55.23it/s] 1%| | 24/3892 [00:00<01:09, 55.57it/s] 1%| | 30/3892 [00:00<01:09, 55.69it/s] 1%| | 36/3892 [00:00<01:09, 55.78it/s] 1%| | 42/3892 [00:00<01:10, 54.58it/s] 1%| | 48/3892 [00:00<01:11, 53.49it/s] 1%|▏ | 54/3892 [00:01<01:15, 50.94it/s] 2%|▏ | 60/3892 [00:01<01:18, 49.01it/s] 2%|▏ | 66/3892 [00:01<01:15, 50.55it/s] 2%|▏ | 72/3892 [00:01<01:13, 51.68it/s] 2%|▏ | 78/3892 [00:01<01:13, 51.75it/s] 2%|▏ | 84/3892 [00:01<01:13, 51.73it/s] 2%|▏ | 90/3892 [00:01<01:14, 50.99it/s] 2%|▏ | 96/3892 [00:01<01:14, 50.77it/s] 3%|▎ | 102/3892 [00:01<01:14, 50.97it/s] 3%|▎ | 108/3892 [00:02<01:13, 51.26it/s] 3%|▎ | 114/3892 [00:02<01:13, 51.37it/s] 3%|▎ | 120/3892 [00:02<01:13, 51.29it/s] 3%|▎ | 126/3892 [00:02<01:11, 52.36it/s] 3%|▎ | 132/3892 [00:02<01:10, 53.35it/s] 4%|▎ | 138/3892 [00:02<01:09, 54.03it/s] 4%|▎ | 144/3892 [00:02<01:09, 54.20it/s] 4%|▍ | 150/3892 [00:02<01:08, 54.64it/s] 4%|▍ | 156/3892 [00:02<01:08, 54.90it/s] 4%|▍ | 162/3892 [00:03<01:07, 55.14it/s] 4%|▍ | 168/3892 [00:03<01:07, 55.32it/s] 4%|▍ | 174/3892 [00:03<01:07, 55.43it/s] 5%|▍ | 180/3892 [00:03<01:06, 55.55it/s] 5%|▍ | 186/3892 [00:03<01:06, 55.57it/s] 5%|▍ | 192/3892 [00:03<01:06, 55.63it/s] 5%|▌ | 198/3892 [00:03<01:07, 55.06it/s] 5%|▌ | 204/3892 [00:03<01:06, 55.13it/s] 5%|▌ | 210/3892 [00:03<01:06, 54.96it/s] 6%|▌ | 216/3892 [00:04<01:06, 55.16it/s] 6%|▌ | 222/3892 [00:04<01:06, 55.38it/s] 6%|▌ | 228/3892 [00:04<01:06, 55.31it/s] 6%|▌ | 234/3892 [00:04<01:05, 55.47it/s] 6%|▌ | 240/3892 [00:04<01:05, 55.52it/s] 6%|▋ | 246/3892 [00:04<01:05, 55.37it/s] 6%|▋ | 252/3892 [00:04<01:05, 55.25it/s] 7%|▋ | 255/3892 [00:04<01:07, 53.56it/s] 2025-01-28 17:21:05.732 | INFO | tools.llama.generate:generate_long:505 - Generated 257 tokens in 4.95 seconds, 51.95 tokens/sec 2025-01-28 17:21:05.732 | INFO | tools.llama.generate:generate_long:508 - Bandwidth achieved: 25.47 GB/s 2025-01-28 17:21:05.732 | INFO | tools.llama.generate:generate_long:513 - GPU Memory used: 1.56 GB /root/.pyenv/versions/3.11.10/lib/python3.11/site-packages/torch/nn/modules/conv.py:306: UserWarning: Plan failed with a cudnnException: CUDNN_BACKEND_EXECUTION_PLAN_DESCRIPTOR: cudnnFinalize Descriptor Failed cudnn_status: CUDNN_STATUS_NOT_SUPPORTED (Triggered internally at ../aten/src/ATen/native/cudnn/Conv_v8.cpp:919.) return F.conv1d(input, weight, bias, self.stride, Next sample
Want to make some of these yourself?
Run this model