cbh123
/
sdxl-spongebob
SDXL fine-tuned on Spongebob
- Public
- 110 runs
-
L40S
- SDXL fine-tune
Prediction
cbh123/sdxl-spongebob:32b470e7IDge6b7sdbtw65xs3cq5hby4d7niStatusSucceededSourceWebHardwareA40 (Large)Total durationCreatedInput
- width
- 1024
- height
- 1024
- prompt
- a rock in the style of TOK
- refine
- no_refiner
- scheduler
- K_EULER
- lora_scale
- 0.6
- num_outputs
- 1
- guidance_scale
- 7.5
- apply_watermark
- high_noise_frac
- 0.8
- negative_prompt
- prompt_strength
- 0.8
- num_inference_steps
- 50
{ "width": 1024, "height": 1024, "prompt": "a rock in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 }
Install Replicate’s Node.js client library:npm install replicate
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Import and set up the client:import Replicate from "replicate"; const replicate = new Replicate({ auth: process.env.REPLICATE_API_TOKEN, });
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
const output = await replicate.run( "cbh123/sdxl-spongebob:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", { input: { width: 1024, height: 1024, prompt: "a rock in the style of TOK", refine: "no_refiner", scheduler: "K_EULER", lora_scale: 0.6, num_outputs: 1, guidance_scale: 7.5, apply_watermark: true, high_noise_frac: 0.8, negative_prompt: "", prompt_strength: 0.8, num_inference_steps: 50 } } ); console.log(output);
To learn more, take a look at the guide on getting started with Node.js.
Install Replicate’s Python client library:pip install replicate
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Import the client:import replicate
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
output = replicate.run( "cbh123/sdxl-spongebob:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", input={ "width": 1024, "height": 1024, "prompt": "a rock in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": True, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } ) print(output)
To learn more, take a look at the guide on getting started with Python.
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
curl -s -X POST \ -H "Authorization: Bearer $REPLICATE_API_TOKEN" \ -H "Content-Type: application/json" \ -H "Prefer: wait" \ -d $'{ "version": "32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", "input": { "width": 1024, "height": 1024, "prompt": "a rock in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } }' \ https://api.replicate.com/v1/predictions
To learn more, take a look at Replicate’s HTTP API reference docs.
You can run this model locally using Cog. First, install Cog:brew install cog
If you don’t have Homebrew, there are other installation options available.
Run this to download the model and run it in your local environment:
cog predict r8.im/cbh123/sdxl-spongebob@sha256:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2 \ -i 'width=1024' \ -i 'height=1024' \ -i 'prompt="a rock in the style of TOK"' \ -i 'refine="no_refiner"' \ -i 'scheduler="K_EULER"' \ -i 'lora_scale=0.6' \ -i 'num_outputs=1' \ -i 'guidance_scale=7.5' \ -i 'apply_watermark=true' \ -i 'high_noise_frac=0.8' \ -i 'negative_prompt=""' \ -i 'prompt_strength=0.8' \ -i 'num_inference_steps=50'
To learn more, take a look at the Cog documentation.
Run this to download the model and run it in your local environment:
docker run -d -p 5000:5000 --gpus=all r8.im/cbh123/sdxl-spongebob@sha256:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2
curl -s -X POST \ -H "Content-Type: application/json" \ -d $'{ "input": { "width": 1024, "height": 1024, "prompt": "a rock in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } }' \ http://localhost:5000/predictions
To learn more, take a look at the Cog documentation.
Output
{ "completed_at": "2023-11-01T13:49:03.779441Z", "created_at": "2023-11-01T13:48:43.791315Z", "data_removed": false, "error": null, "id": "ge6b7sdbtw65xs3cq5hby4d7ni", "input": { "width": 1024, "height": 1024, "prompt": "a rock in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 }, "logs": "Using seed: 5330\nEnsuring enough disk space...\nFree disk space: 1709969444864\nDownloading weights: https://pbxt.replicate.delivery/TaVwE3krc0pMLxaP7wipUtHVZv329Jc4vRwXbGkR0INf145IA/trained_model.tar\nb'Downloaded 186 MB bytes in 0.252s (739 MB/s)\\nExtracted 186 MB in 0.052s (3.6 GB/s)\\n'\nDownloaded weights in 0.42419958114624023 seconds\nLoading fine-tuned model\nDoes not have Unet. assume we are using LoRA\nLoading Unet LoRA\nPrompt: a rock in the style of <s0><s1>\ntxt2img mode\n 0%| | 0/50 [00:00<?, ?it/s]\n 2%|▏ | 1/50 [00:00<00:13, 3.50it/s]\n 4%|▍ | 2/50 [00:00<00:13, 3.59it/s]\n 6%|▌ | 3/50 [00:00<00:12, 3.62it/s]\n 8%|▊ | 4/50 [00:01<00:12, 3.64it/s]\n 10%|█ | 5/50 [00:01<00:12, 3.65it/s]\n 12%|█▏ | 6/50 [00:01<00:12, 3.65it/s]\n 14%|█▍ | 7/50 [00:01<00:11, 3.66it/s]\n 16%|█▌ | 8/50 [00:02<00:11, 3.65it/s]\n 18%|█▊ | 9/50 [00:02<00:11, 3.66it/s]\n 20%|██ | 10/50 [00:02<00:10, 3.66it/s]\n 22%|██▏ | 11/50 [00:03<00:10, 3.66it/s]\n 24%|██▍ | 12/50 [00:03<00:10, 3.66it/s]\n 26%|██▌ | 13/50 [00:03<00:10, 3.66it/s]\n 28%|██▊ | 14/50 [00:03<00:09, 3.66it/s]\n 30%|███ | 15/50 [00:04<00:09, 3.66it/s]\n 32%|███▏ | 16/50 [00:04<00:09, 3.66it/s]\n 34%|███▍ | 17/50 [00:04<00:09, 3.65it/s]\n 36%|███▌ | 18/50 [00:04<00:08, 3.65it/s]\n 38%|███▊ | 19/50 [00:05<00:08, 3.65it/s]\n 40%|████ | 20/50 [00:05<00:08, 3.65it/s]\n 42%|████▏ | 21/50 [00:05<00:07, 3.65it/s]\n 44%|████▍ | 22/50 [00:06<00:07, 3.65it/s]\n 46%|████▌ | 23/50 [00:06<00:07, 3.65it/s]\n 48%|████▊ | 24/50 [00:06<00:07, 3.65it/s]\n 50%|█████ | 25/50 [00:06<00:06, 3.65it/s]\n 52%|█████▏ | 26/50 [00:07<00:06, 3.65it/s]\n 54%|█████▍ | 27/50 [00:07<00:06, 3.65it/s]\n 56%|█████▌ | 28/50 [00:07<00:06, 3.65it/s]\n 58%|█████▊ | 29/50 [00:07<00:05, 3.65it/s]\n 60%|██████ | 30/50 [00:08<00:05, 3.64it/s]\n 62%|██████▏ | 31/50 [00:08<00:05, 3.64it/s]\n 64%|██████▍ | 32/50 [00:08<00:04, 3.64it/s]\n 66%|██████▌ | 33/50 [00:09<00:04, 3.64it/s]\n 68%|██████▊ | 34/50 [00:09<00:04, 3.64it/s]\n 70%|███████ | 35/50 [00:09<00:04, 3.64it/s]\n 72%|███████▏ | 36/50 [00:09<00:03, 3.64it/s]\n 74%|███████▍ | 37/50 [00:10<00:03, 3.65it/s]\n 76%|███████▌ | 38/50 [00:10<00:03, 3.64it/s]\n 78%|███████▊ | 39/50 [00:10<00:03, 3.64it/s]\n 80%|████████ | 40/50 [00:10<00:02, 3.64it/s]\n 82%|████████▏ | 41/50 [00:11<00:02, 3.64it/s]\n 84%|████████▍ | 42/50 [00:11<00:02, 3.64it/s]\n 86%|████████▌ | 43/50 [00:11<00:01, 3.63it/s]\n 88%|████████▊ | 44/50 [00:12<00:01, 3.64it/s]\n 90%|█████████ | 45/50 [00:12<00:01, 3.64it/s]\n 92%|█████████▏| 46/50 [00:12<00:01, 3.63it/s]\n 94%|█████████▍| 47/50 [00:12<00:00, 3.63it/s]\n 96%|█████████▌| 48/50 [00:13<00:00, 3.64it/s]\n 98%|█████████▊| 49/50 [00:13<00:00, 3.63it/s]\n100%|██████████| 50/50 [00:13<00:00, 3.63it/s]\n100%|██████████| 50/50 [00:13<00:00, 3.64it/s]", "metrics": { "predict_time": 16.655912, "total_time": 19.988126 }, "output": [ "https://replicate.delivery/pbxt/dbqRAWpmf4Ten0839TWkV0H6aC3qBD4wCi4azzyFv4iOlA0RA/out-0.png" ], "started_at": "2023-11-01T13:48:47.123529Z", "status": "succeeded", "urls": { "get": "https://api.replicate.com/v1/predictions/ge6b7sdbtw65xs3cq5hby4d7ni", "cancel": "https://api.replicate.com/v1/predictions/ge6b7sdbtw65xs3cq5hby4d7ni/cancel" }, "version": "32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2" }
Generated inUsing seed: 5330 Ensuring enough disk space... Free disk space: 1709969444864 Downloading weights: https://pbxt.replicate.delivery/TaVwE3krc0pMLxaP7wipUtHVZv329Jc4vRwXbGkR0INf145IA/trained_model.tar b'Downloaded 186 MB bytes in 0.252s (739 MB/s)\nExtracted 186 MB in 0.052s (3.6 GB/s)\n' Downloaded weights in 0.42419958114624023 seconds Loading fine-tuned model Does not have Unet. assume we are using LoRA Loading Unet LoRA Prompt: a rock in the style of <s0><s1> txt2img mode 0%| | 0/50 [00:00<?, ?it/s] 2%|▏ | 1/50 [00:00<00:13, 3.50it/s] 4%|▍ | 2/50 [00:00<00:13, 3.59it/s] 6%|▌ | 3/50 [00:00<00:12, 3.62it/s] 8%|▊ | 4/50 [00:01<00:12, 3.64it/s] 10%|█ | 5/50 [00:01<00:12, 3.65it/s] 12%|█▏ | 6/50 [00:01<00:12, 3.65it/s] 14%|█▍ | 7/50 [00:01<00:11, 3.66it/s] 16%|█▌ | 8/50 [00:02<00:11, 3.65it/s] 18%|█▊ | 9/50 [00:02<00:11, 3.66it/s] 20%|██ | 10/50 [00:02<00:10, 3.66it/s] 22%|██▏ | 11/50 [00:03<00:10, 3.66it/s] 24%|██▍ | 12/50 [00:03<00:10, 3.66it/s] 26%|██▌ | 13/50 [00:03<00:10, 3.66it/s] 28%|██▊ | 14/50 [00:03<00:09, 3.66it/s] 30%|███ | 15/50 [00:04<00:09, 3.66it/s] 32%|███▏ | 16/50 [00:04<00:09, 3.66it/s] 34%|███▍ | 17/50 [00:04<00:09, 3.65it/s] 36%|███▌ | 18/50 [00:04<00:08, 3.65it/s] 38%|███▊ | 19/50 [00:05<00:08, 3.65it/s] 40%|████ | 20/50 [00:05<00:08, 3.65it/s] 42%|████▏ | 21/50 [00:05<00:07, 3.65it/s] 44%|████▍ | 22/50 [00:06<00:07, 3.65it/s] 46%|████▌ | 23/50 [00:06<00:07, 3.65it/s] 48%|████▊ | 24/50 [00:06<00:07, 3.65it/s] 50%|█████ | 25/50 [00:06<00:06, 3.65it/s] 52%|█████▏ | 26/50 [00:07<00:06, 3.65it/s] 54%|█████▍ | 27/50 [00:07<00:06, 3.65it/s] 56%|█████▌ | 28/50 [00:07<00:06, 3.65it/s] 58%|█████▊ | 29/50 [00:07<00:05, 3.65it/s] 60%|██████ | 30/50 [00:08<00:05, 3.64it/s] 62%|██████▏ | 31/50 [00:08<00:05, 3.64it/s] 64%|██████▍ | 32/50 [00:08<00:04, 3.64it/s] 66%|██████▌ | 33/50 [00:09<00:04, 3.64it/s] 68%|██████▊ | 34/50 [00:09<00:04, 3.64it/s] 70%|███████ | 35/50 [00:09<00:04, 3.64it/s] 72%|███████▏ | 36/50 [00:09<00:03, 3.64it/s] 74%|███████▍ | 37/50 [00:10<00:03, 3.65it/s] 76%|███████▌ | 38/50 [00:10<00:03, 3.64it/s] 78%|███████▊ | 39/50 [00:10<00:03, 3.64it/s] 80%|████████ | 40/50 [00:10<00:02, 3.64it/s] 82%|████████▏ | 41/50 [00:11<00:02, 3.64it/s] 84%|████████▍ | 42/50 [00:11<00:02, 3.64it/s] 86%|████████▌ | 43/50 [00:11<00:01, 3.63it/s] 88%|████████▊ | 44/50 [00:12<00:01, 3.64it/s] 90%|█████████ | 45/50 [00:12<00:01, 3.64it/s] 92%|█████████▏| 46/50 [00:12<00:01, 3.63it/s] 94%|█████████▍| 47/50 [00:12<00:00, 3.63it/s] 96%|█████████▌| 48/50 [00:13<00:00, 3.64it/s] 98%|█████████▊| 49/50 [00:13<00:00, 3.63it/s] 100%|██████████| 50/50 [00:13<00:00, 3.63it/s] 100%|██████████| 50/50 [00:13<00:00, 3.64it/s]
Prediction
cbh123/sdxl-spongebob:32b470e7IDlw4ziulbbl4ihcrri5mx5j3ykyStatusSucceededSourceWebHardwareA40 (Large)Total durationCreatedby @cbh123Input
- width
- 1024
- height
- 1024
- prompt
- a cartoon starfish in the style of TOK
- refine
- no_refiner
- scheduler
- K_EULER
- lora_scale
- 0.6
- num_outputs
- 1
- guidance_scale
- 7.5
- apply_watermark
- high_noise_frac
- 0.8
- negative_prompt
- prompt_strength
- 0.8
- num_inference_steps
- 50
{ "width": 1024, "height": 1024, "prompt": "a cartoon starfish in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 }
Install Replicate’s Node.js client library:npm install replicate
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Import and set up the client:import Replicate from "replicate"; const replicate = new Replicate({ auth: process.env.REPLICATE_API_TOKEN, });
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
const output = await replicate.run( "cbh123/sdxl-spongebob:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", { input: { width: 1024, height: 1024, prompt: "a cartoon starfish in the style of TOK", refine: "no_refiner", scheduler: "K_EULER", lora_scale: 0.6, num_outputs: 1, guidance_scale: 7.5, apply_watermark: true, high_noise_frac: 0.8, negative_prompt: "", prompt_strength: 0.8, num_inference_steps: 50 } } ); console.log(output);
To learn more, take a look at the guide on getting started with Node.js.
Install Replicate’s Python client library:pip install replicate
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Import the client:import replicate
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
output = replicate.run( "cbh123/sdxl-spongebob:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", input={ "width": 1024, "height": 1024, "prompt": "a cartoon starfish in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": True, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } ) print(output)
To learn more, take a look at the guide on getting started with Python.
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
curl -s -X POST \ -H "Authorization: Bearer $REPLICATE_API_TOKEN" \ -H "Content-Type: application/json" \ -H "Prefer: wait" \ -d $'{ "version": "32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", "input": { "width": 1024, "height": 1024, "prompt": "a cartoon starfish in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } }' \ https://api.replicate.com/v1/predictions
To learn more, take a look at Replicate’s HTTP API reference docs.
You can run this model locally using Cog. First, install Cog:brew install cog
If you don’t have Homebrew, there are other installation options available.
Run this to download the model and run it in your local environment:
cog predict r8.im/cbh123/sdxl-spongebob@sha256:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2 \ -i 'width=1024' \ -i 'height=1024' \ -i 'prompt="a cartoon starfish in the style of TOK"' \ -i 'refine="no_refiner"' \ -i 'scheduler="K_EULER"' \ -i 'lora_scale=0.6' \ -i 'num_outputs=1' \ -i 'guidance_scale=7.5' \ -i 'apply_watermark=true' \ -i 'high_noise_frac=0.8' \ -i 'negative_prompt=""' \ -i 'prompt_strength=0.8' \ -i 'num_inference_steps=50'
To learn more, take a look at the Cog documentation.
Run this to download the model and run it in your local environment:
docker run -d -p 5000:5000 --gpus=all r8.im/cbh123/sdxl-spongebob@sha256:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2
curl -s -X POST \ -H "Content-Type: application/json" \ -d $'{ "input": { "width": 1024, "height": 1024, "prompt": "a cartoon starfish in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } }' \ http://localhost:5000/predictions
To learn more, take a look at the Cog documentation.
Output
{ "completed_at": "2023-11-01T13:51:06.678896Z", "created_at": "2023-11-01T13:50:48.115591Z", "data_removed": false, "error": null, "id": "lw4ziulbbl4ihcrri5mx5j3yky", "input": { "width": 1024, "height": 1024, "prompt": "a cartoon starfish in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 }, "logs": "Using seed: 41163\nEnsuring enough disk space...\nFree disk space: 1439866560512\nDownloading weights: https://pbxt.replicate.delivery/TaVwE3krc0pMLxaP7wipUtHVZv329Jc4vRwXbGkR0INf145IA/trained_model.tar\nb'Downloaded 186 MB bytes in 1.081s (172 MB/s)\\nExtracted 186 MB in 0.049s (3.8 GB/s)\\n'\nDownloaded weights in 1.215386152267456 seconds\nLoading fine-tuned model\nDoes not have Unet. assume we are using LoRA\nLoading Unet LoRA\nPrompt: a cartoon starfish in the style of <s0><s1>\ntxt2img mode\n 0%| | 0/50 [00:00<?, ?it/s]\n 2%|▏ | 1/50 [00:00<00:13, 3.63it/s]\n 4%|▍ | 2/50 [00:00<00:13, 3.63it/s]\n 6%|▌ | 3/50 [00:00<00:12, 3.62it/s]\n 8%|▊ | 4/50 [00:01<00:12, 3.62it/s]\n 10%|█ | 5/50 [00:01<00:12, 3.63it/s]\n 12%|█▏ | 6/50 [00:01<00:12, 3.62it/s]\n 14%|█▍ | 7/50 [00:01<00:11, 3.63it/s]\n 16%|█▌ | 8/50 [00:02<00:11, 3.62it/s]\n 18%|█▊ | 9/50 [00:02<00:11, 3.62it/s]\n 20%|██ | 10/50 [00:02<00:11, 3.62it/s]\n 22%|██▏ | 11/50 [00:03<00:10, 3.62it/s]\n 24%|██▍ | 12/50 [00:03<00:10, 3.62it/s]\n 26%|██▌ | 13/50 [00:03<00:10, 3.61it/s]\n 28%|██▊ | 14/50 [00:03<00:09, 3.62it/s]\n 30%|███ | 15/50 [00:04<00:09, 3.62it/s]\n 32%|███▏ | 16/50 [00:04<00:09, 3.62it/s]\n 34%|███▍ | 17/50 [00:04<00:09, 3.62it/s]\n 36%|███▌ | 18/50 [00:04<00:08, 3.62it/s]\n 38%|███▊ | 19/50 [00:05<00:08, 3.62it/s]\n 40%|████ | 20/50 [00:05<00:08, 3.62it/s]\n 42%|████▏ | 21/50 [00:05<00:08, 3.61it/s]\n 44%|████▍ | 22/50 [00:06<00:07, 3.61it/s]\n 46%|████▌ | 23/50 [00:06<00:07, 3.61it/s]\n 48%|████▊ | 24/50 [00:06<00:07, 3.61it/s]\n 50%|█████ | 25/50 [00:06<00:06, 3.61it/s]\n 52%|█████▏ | 26/50 [00:07<00:06, 3.61it/s]\n 54%|█████▍ | 27/50 [00:07<00:06, 3.61it/s]\n 56%|█████▌ | 28/50 [00:07<00:06, 3.61it/s]\n 58%|█████▊ | 29/50 [00:08<00:05, 3.61it/s]\n 60%|██████ | 30/50 [00:08<00:05, 3.61it/s]\n 62%|██████▏ | 31/50 [00:08<00:05, 3.61it/s]\n 64%|██████▍ | 32/50 [00:08<00:04, 3.61it/s]\n 66%|██████▌ | 33/50 [00:09<00:04, 3.61it/s]\n 68%|██████▊ | 34/50 [00:09<00:04, 3.61it/s]\n 70%|███████ | 35/50 [00:09<00:04, 3.61it/s]\n 72%|███████▏ | 36/50 [00:09<00:03, 3.61it/s]\n 74%|███████▍ | 37/50 [00:10<00:03, 3.61it/s]\n 76%|███████▌ | 38/50 [00:10<00:03, 3.61it/s]\n 78%|███████▊ | 39/50 [00:10<00:03, 3.61it/s]\n 80%|████████ | 40/50 [00:11<00:02, 3.61it/s]\n 82%|████████▏ | 41/50 [00:11<00:02, 3.61it/s]\n 84%|████████▍ | 42/50 [00:11<00:02, 3.61it/s]\n 86%|████████▌ | 43/50 [00:11<00:01, 3.61it/s]\n 88%|████████▊ | 44/50 [00:12<00:01, 3.61it/s]\n 90%|█████████ | 45/50 [00:12<00:01, 3.61it/s]\n 92%|█████████▏| 46/50 [00:12<00:01, 3.60it/s]\n 94%|█████████▍| 47/50 [00:13<00:00, 3.60it/s]\n 96%|█████████▌| 48/50 [00:13<00:00, 3.61it/s]\n 98%|█████████▊| 49/50 [00:13<00:00, 3.61it/s]\n100%|██████████| 50/50 [00:13<00:00, 3.60it/s]\n100%|██████████| 50/50 [00:13<00:00, 3.61it/s]", "metrics": { "predict_time": 18.106444, "total_time": 18.563305 }, "output": [ "https://replicate.delivery/pbxt/xU99cC6f4FUZMKffHnvfMYzcszvazf1HIBBYqto0V4nP5EgOC/out-0.png" ], "started_at": "2023-11-01T13:50:48.572452Z", "status": "succeeded", "urls": { "get": "https://api.replicate.com/v1/predictions/lw4ziulbbl4ihcrri5mx5j3yky", "cancel": "https://api.replicate.com/v1/predictions/lw4ziulbbl4ihcrri5mx5j3yky/cancel" }, "version": "32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2" }
Generated inUsing seed: 41163 Ensuring enough disk space... Free disk space: 1439866560512 Downloading weights: https://pbxt.replicate.delivery/TaVwE3krc0pMLxaP7wipUtHVZv329Jc4vRwXbGkR0INf145IA/trained_model.tar b'Downloaded 186 MB bytes in 1.081s (172 MB/s)\nExtracted 186 MB in 0.049s (3.8 GB/s)\n' Downloaded weights in 1.215386152267456 seconds Loading fine-tuned model Does not have Unet. assume we are using LoRA Loading Unet LoRA Prompt: a cartoon starfish in the style of <s0><s1> txt2img mode 0%| | 0/50 [00:00<?, ?it/s] 2%|▏ | 1/50 [00:00<00:13, 3.63it/s] 4%|▍ | 2/50 [00:00<00:13, 3.63it/s] 6%|▌ | 3/50 [00:00<00:12, 3.62it/s] 8%|▊ | 4/50 [00:01<00:12, 3.62it/s] 10%|█ | 5/50 [00:01<00:12, 3.63it/s] 12%|█▏ | 6/50 [00:01<00:12, 3.62it/s] 14%|█▍ | 7/50 [00:01<00:11, 3.63it/s] 16%|█▌ | 8/50 [00:02<00:11, 3.62it/s] 18%|█▊ | 9/50 [00:02<00:11, 3.62it/s] 20%|██ | 10/50 [00:02<00:11, 3.62it/s] 22%|██▏ | 11/50 [00:03<00:10, 3.62it/s] 24%|██▍ | 12/50 [00:03<00:10, 3.62it/s] 26%|██▌ | 13/50 [00:03<00:10, 3.61it/s] 28%|██▊ | 14/50 [00:03<00:09, 3.62it/s] 30%|███ | 15/50 [00:04<00:09, 3.62it/s] 32%|███▏ | 16/50 [00:04<00:09, 3.62it/s] 34%|███▍ | 17/50 [00:04<00:09, 3.62it/s] 36%|███▌ | 18/50 [00:04<00:08, 3.62it/s] 38%|███▊ | 19/50 [00:05<00:08, 3.62it/s] 40%|████ | 20/50 [00:05<00:08, 3.62it/s] 42%|████▏ | 21/50 [00:05<00:08, 3.61it/s] 44%|████▍ | 22/50 [00:06<00:07, 3.61it/s] 46%|████▌ | 23/50 [00:06<00:07, 3.61it/s] 48%|████▊ | 24/50 [00:06<00:07, 3.61it/s] 50%|█████ | 25/50 [00:06<00:06, 3.61it/s] 52%|█████▏ | 26/50 [00:07<00:06, 3.61it/s] 54%|█████▍ | 27/50 [00:07<00:06, 3.61it/s] 56%|█████▌ | 28/50 [00:07<00:06, 3.61it/s] 58%|█████▊ | 29/50 [00:08<00:05, 3.61it/s] 60%|██████ | 30/50 [00:08<00:05, 3.61it/s] 62%|██████▏ | 31/50 [00:08<00:05, 3.61it/s] 64%|██████▍ | 32/50 [00:08<00:04, 3.61it/s] 66%|██████▌ | 33/50 [00:09<00:04, 3.61it/s] 68%|██████▊ | 34/50 [00:09<00:04, 3.61it/s] 70%|███████ | 35/50 [00:09<00:04, 3.61it/s] 72%|███████▏ | 36/50 [00:09<00:03, 3.61it/s] 74%|███████▍ | 37/50 [00:10<00:03, 3.61it/s] 76%|███████▌ | 38/50 [00:10<00:03, 3.61it/s] 78%|███████▊ | 39/50 [00:10<00:03, 3.61it/s] 80%|████████ | 40/50 [00:11<00:02, 3.61it/s] 82%|████████▏ | 41/50 [00:11<00:02, 3.61it/s] 84%|████████▍ | 42/50 [00:11<00:02, 3.61it/s] 86%|████████▌ | 43/50 [00:11<00:01, 3.61it/s] 88%|████████▊ | 44/50 [00:12<00:01, 3.61it/s] 90%|█████████ | 45/50 [00:12<00:01, 3.61it/s] 92%|█████████▏| 46/50 [00:12<00:01, 3.60it/s] 94%|█████████▍| 47/50 [00:13<00:00, 3.60it/s] 96%|█████████▌| 48/50 [00:13<00:00, 3.61it/s] 98%|█████████▊| 49/50 [00:13<00:00, 3.61it/s] 100%|██████████| 50/50 [00:13<00:00, 3.60it/s] 100%|██████████| 50/50 [00:13<00:00, 3.61it/s]
Prediction
cbh123/sdxl-spongebob:32b470e7IDm7bzc7tbdcsbngruenvhevv2aeStatusSucceededSourceWebHardwareA40 (Large)Total durationCreatedby @cbh123Input
- width
- 1024
- height
- 1024
- prompt
- a hamburger in the style of TOK
- refine
- no_refiner
- scheduler
- K_EULER
- lora_scale
- 0.6
- num_outputs
- 1
- guidance_scale
- 7.5
- apply_watermark
- high_noise_frac
- 0.8
- negative_prompt
- prompt_strength
- 0.8
- num_inference_steps
- 50
{ "width": 1024, "height": 1024, "prompt": "a hamburger in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 }
Install Replicate’s Node.js client library:npm install replicate
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Import and set up the client:import Replicate from "replicate"; const replicate = new Replicate({ auth: process.env.REPLICATE_API_TOKEN, });
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
const output = await replicate.run( "cbh123/sdxl-spongebob:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", { input: { width: 1024, height: 1024, prompt: "a hamburger in the style of TOK", refine: "no_refiner", scheduler: "K_EULER", lora_scale: 0.6, num_outputs: 1, guidance_scale: 7.5, apply_watermark: true, high_noise_frac: 0.8, negative_prompt: "", prompt_strength: 0.8, num_inference_steps: 50 } } ); console.log(output);
To learn more, take a look at the guide on getting started with Node.js.
Install Replicate’s Python client library:pip install replicate
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Import the client:import replicate
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
output = replicate.run( "cbh123/sdxl-spongebob:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", input={ "width": 1024, "height": 1024, "prompt": "a hamburger in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": True, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } ) print(output)
To learn more, take a look at the guide on getting started with Python.
Set theREPLICATE_API_TOKEN
environment variable:export REPLICATE_API_TOKEN=<paste-your-token-here>
Find your API token in your account settings.
Run cbh123/sdxl-spongebob using Replicate’s API. Check out the model's schema for an overview of inputs and outputs.
curl -s -X POST \ -H "Authorization: Bearer $REPLICATE_API_TOKEN" \ -H "Content-Type: application/json" \ -H "Prefer: wait" \ -d $'{ "version": "32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2", "input": { "width": 1024, "height": 1024, "prompt": "a hamburger in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } }' \ https://api.replicate.com/v1/predictions
To learn more, take a look at Replicate’s HTTP API reference docs.
You can run this model locally using Cog. First, install Cog:brew install cog
If you don’t have Homebrew, there are other installation options available.
Run this to download the model and run it in your local environment:
cog predict r8.im/cbh123/sdxl-spongebob@sha256:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2 \ -i 'width=1024' \ -i 'height=1024' \ -i 'prompt="a hamburger in the style of TOK"' \ -i 'refine="no_refiner"' \ -i 'scheduler="K_EULER"' \ -i 'lora_scale=0.6' \ -i 'num_outputs=1' \ -i 'guidance_scale=7.5' \ -i 'apply_watermark=true' \ -i 'high_noise_frac=0.8' \ -i 'negative_prompt=""' \ -i 'prompt_strength=0.8' \ -i 'num_inference_steps=50'
To learn more, take a look at the Cog documentation.
Run this to download the model and run it in your local environment:
docker run -d -p 5000:5000 --gpus=all r8.im/cbh123/sdxl-spongebob@sha256:32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2
curl -s -X POST \ -H "Content-Type: application/json" \ -d $'{ "input": { "width": 1024, "height": 1024, "prompt": "a hamburger in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 } }' \ http://localhost:5000/predictions
To learn more, take a look at the Cog documentation.
Output
{ "completed_at": "2023-11-01T13:58:15.350679Z", "created_at": "2023-11-01T13:57:55.161281Z", "data_removed": false, "error": null, "id": "m7bzc7tbdcsbngruenvhevv2ae", "input": { "width": 1024, "height": 1024, "prompt": "a hamburger in the style of TOK", "refine": "no_refiner", "scheduler": "K_EULER", "lora_scale": 0.6, "num_outputs": 1, "guidance_scale": 7.5, "apply_watermark": true, "high_noise_frac": 0.8, "negative_prompt": "", "prompt_strength": 0.8, "num_inference_steps": 50 }, "logs": "Using seed: 11632\nEnsuring enough disk space...\nFree disk space: 1389147619328\nDownloading weights: https://pbxt.replicate.delivery/TaVwE3krc0pMLxaP7wipUtHVZv329Jc4vRwXbGkR0INf145IA/trained_model.tar\nb'Downloaded 186 MB bytes in 0.592s (314 MB/s)\\nExtracted 186 MB in 0.054s (3.4 GB/s)\\n'\nDownloaded weights in 0.7257587909698486 seconds\nLoading fine-tuned model\nDoes not have Unet. assume we are using LoRA\nLoading Unet LoRA\nPrompt: a hamburger in the style of <s0><s1>\ntxt2img mode\n 0%| | 0/50 [00:00<?, ?it/s]\n 2%|▏ | 1/50 [00:00<00:13, 3.65it/s]\n 4%|▍ | 2/50 [00:00<00:13, 3.64it/s]\n 6%|▌ | 3/50 [00:00<00:12, 3.65it/s]\n 8%|▊ | 4/50 [00:01<00:12, 3.64it/s]\n 10%|█ | 5/50 [00:01<00:12, 3.64it/s]\n 12%|█▏ | 6/50 [00:01<00:12, 3.64it/s]\n 14%|█▍ | 7/50 [00:01<00:11, 3.64it/s]\n 16%|█▌ | 8/50 [00:02<00:11, 3.64it/s]\n 18%|█▊ | 9/50 [00:02<00:11, 3.64it/s]\n 20%|██ | 10/50 [00:02<00:10, 3.64it/s]\n 22%|██▏ | 11/50 [00:03<00:10, 3.64it/s]\n 24%|██▍ | 12/50 [00:03<00:10, 3.64it/s]\n 26%|██▌ | 13/50 [00:03<00:10, 3.64it/s]\n 28%|██▊ | 14/50 [00:03<00:09, 3.64it/s]\n 30%|███ | 15/50 [00:04<00:09, 3.64it/s]\n 32%|███▏ | 16/50 [00:04<00:09, 3.64it/s]\n 34%|███▍ | 17/50 [00:04<00:09, 3.64it/s]\n 36%|███▌ | 18/50 [00:04<00:08, 3.64it/s]\n 38%|███▊ | 19/50 [00:05<00:08, 3.64it/s]\n 40%|████ | 20/50 [00:05<00:08, 3.63it/s]\n 42%|████▏ | 21/50 [00:05<00:07, 3.63it/s]\n 44%|████▍ | 22/50 [00:06<00:07, 3.63it/s]\n 46%|████▌ | 23/50 [00:06<00:07, 3.63it/s]\n 48%|████▊ | 24/50 [00:06<00:07, 3.63it/s]\n 50%|█████ | 25/50 [00:06<00:06, 3.62it/s]\n 52%|█████▏ | 26/50 [00:07<00:06, 3.63it/s]\n 54%|█████▍ | 27/50 [00:07<00:06, 3.63it/s]\n 56%|█████▌ | 28/50 [00:07<00:06, 3.62it/s]\n 58%|█████▊ | 29/50 [00:07<00:05, 3.62it/s]\n 60%|██████ | 30/50 [00:08<00:05, 3.62it/s]\n 62%|██████▏ | 31/50 [00:08<00:05, 3.62it/s]\n 64%|██████▍ | 32/50 [00:08<00:04, 3.62it/s]\n 66%|██████▌ | 33/50 [00:09<00:04, 3.62it/s]\n 68%|██████▊ | 34/50 [00:09<00:04, 3.62it/s]\n 70%|███████ | 35/50 [00:09<00:04, 3.62it/s]\n 72%|███████▏ | 36/50 [00:09<00:03, 3.62it/s]\n 74%|███████▍ | 37/50 [00:10<00:03, 3.62it/s]\n 76%|███████▌ | 38/50 [00:10<00:03, 3.62it/s]\n 78%|███████▊ | 39/50 [00:10<00:03, 3.62it/s]\n 80%|████████ | 40/50 [00:11<00:02, 3.62it/s]\n 82%|████████▏ | 41/50 [00:11<00:02, 3.62it/s]\n 84%|████████▍ | 42/50 [00:11<00:02, 3.62it/s]\n 86%|████████▌ | 43/50 [00:11<00:01, 3.62it/s]\n 88%|████████▊ | 44/50 [00:12<00:01, 3.62it/s]\n 90%|█████████ | 45/50 [00:12<00:01, 3.62it/s]\n 92%|█████████▏| 46/50 [00:12<00:01, 3.62it/s]\n 94%|█████████▍| 47/50 [00:12<00:00, 3.62it/s]\n 96%|█████████▌| 48/50 [00:13<00:00, 3.62it/s]\n 98%|█████████▊| 49/50 [00:13<00:00, 3.61it/s]\n100%|██████████| 50/50 [00:13<00:00, 3.62it/s]\n100%|██████████| 50/50 [00:13<00:00, 3.63it/s]", "metrics": { "predict_time": 17.360879, "total_time": 20.189398 }, "output": [ "https://replicate.delivery/pbxt/mAwkvxbnW9bmCl5qKeQg7mkjM4QBsyi1BnSilwXk9fd2tA0RA/out-0.png" ], "started_at": "2023-11-01T13:57:57.989800Z", "status": "succeeded", "urls": { "get": "https://api.replicate.com/v1/predictions/m7bzc7tbdcsbngruenvhevv2ae", "cancel": "https://api.replicate.com/v1/predictions/m7bzc7tbdcsbngruenvhevv2ae/cancel" }, "version": "32b470e78297e4e79cd59599a16854b6d243e2e07d0d51cb21c7cac5838034f2" }
Generated inUsing seed: 11632 Ensuring enough disk space... Free disk space: 1389147619328 Downloading weights: https://pbxt.replicate.delivery/TaVwE3krc0pMLxaP7wipUtHVZv329Jc4vRwXbGkR0INf145IA/trained_model.tar b'Downloaded 186 MB bytes in 0.592s (314 MB/s)\nExtracted 186 MB in 0.054s (3.4 GB/s)\n' Downloaded weights in 0.7257587909698486 seconds Loading fine-tuned model Does not have Unet. assume we are using LoRA Loading Unet LoRA Prompt: a hamburger in the style of <s0><s1> txt2img mode 0%| | 0/50 [00:00<?, ?it/s] 2%|▏ | 1/50 [00:00<00:13, 3.65it/s] 4%|▍ | 2/50 [00:00<00:13, 3.64it/s] 6%|▌ | 3/50 [00:00<00:12, 3.65it/s] 8%|▊ | 4/50 [00:01<00:12, 3.64it/s] 10%|█ | 5/50 [00:01<00:12, 3.64it/s] 12%|█▏ | 6/50 [00:01<00:12, 3.64it/s] 14%|█▍ | 7/50 [00:01<00:11, 3.64it/s] 16%|█▌ | 8/50 [00:02<00:11, 3.64it/s] 18%|█▊ | 9/50 [00:02<00:11, 3.64it/s] 20%|██ | 10/50 [00:02<00:10, 3.64it/s] 22%|██▏ | 11/50 [00:03<00:10, 3.64it/s] 24%|██▍ | 12/50 [00:03<00:10, 3.64it/s] 26%|██▌ | 13/50 [00:03<00:10, 3.64it/s] 28%|██▊ | 14/50 [00:03<00:09, 3.64it/s] 30%|███ | 15/50 [00:04<00:09, 3.64it/s] 32%|███▏ | 16/50 [00:04<00:09, 3.64it/s] 34%|███▍ | 17/50 [00:04<00:09, 3.64it/s] 36%|███▌ | 18/50 [00:04<00:08, 3.64it/s] 38%|███▊ | 19/50 [00:05<00:08, 3.64it/s] 40%|████ | 20/50 [00:05<00:08, 3.63it/s] 42%|████▏ | 21/50 [00:05<00:07, 3.63it/s] 44%|████▍ | 22/50 [00:06<00:07, 3.63it/s] 46%|████▌ | 23/50 [00:06<00:07, 3.63it/s] 48%|████▊ | 24/50 [00:06<00:07, 3.63it/s] 50%|█████ | 25/50 [00:06<00:06, 3.62it/s] 52%|█████▏ | 26/50 [00:07<00:06, 3.63it/s] 54%|█████▍ | 27/50 [00:07<00:06, 3.63it/s] 56%|█████▌ | 28/50 [00:07<00:06, 3.62it/s] 58%|█████▊ | 29/50 [00:07<00:05, 3.62it/s] 60%|██████ | 30/50 [00:08<00:05, 3.62it/s] 62%|██████▏ | 31/50 [00:08<00:05, 3.62it/s] 64%|██████▍ | 32/50 [00:08<00:04, 3.62it/s] 66%|██████▌ | 33/50 [00:09<00:04, 3.62it/s] 68%|██████▊ | 34/50 [00:09<00:04, 3.62it/s] 70%|███████ | 35/50 [00:09<00:04, 3.62it/s] 72%|███████▏ | 36/50 [00:09<00:03, 3.62it/s] 74%|███████▍ | 37/50 [00:10<00:03, 3.62it/s] 76%|███████▌ | 38/50 [00:10<00:03, 3.62it/s] 78%|███████▊ | 39/50 [00:10<00:03, 3.62it/s] 80%|████████ | 40/50 [00:11<00:02, 3.62it/s] 82%|████████▏ | 41/50 [00:11<00:02, 3.62it/s] 84%|████████▍ | 42/50 [00:11<00:02, 3.62it/s] 86%|████████▌ | 43/50 [00:11<00:01, 3.62it/s] 88%|████████▊ | 44/50 [00:12<00:01, 3.62it/s] 90%|█████████ | 45/50 [00:12<00:01, 3.62it/s] 92%|█████████▏| 46/50 [00:12<00:01, 3.62it/s] 94%|█████████▍| 47/50 [00:12<00:00, 3.62it/s] 96%|█████████▌| 48/50 [00:13<00:00, 3.62it/s] 98%|█████████▊| 49/50 [00:13<00:00, 3.61it/s] 100%|██████████| 50/50 [00:13<00:00, 3.62it/s] 100%|██████████| 50/50 [00:13<00:00, 3.63it/s]
Want to make some of these yourself?
Run this model