We've made running fine-tunes on Replicate much faster, and the optimizations are open-source.
A new set of image generation capabilities for FLUX models, including inpainting, outpainting, canny edge detection, and depth maps.
NVIDIA L40S GPUs are here, with better performance and lower cost.
We've decided to stop returning data URLs in sync API outputs based on feedback from users. We're going to take a few steps back and figure out the best way to get your model output as fast as possible.
We've partnered with Ideogram to bring their inpainting model to Replicate's API.
Stability AI's latest text-to-image model is now available on Replicate and you can run it with an API.
FLUX is now much faster on Replicate, and we’ve made our optimizations open-source so you can see exactly how they work and build upon them.
Black Forest Labs continue to push boundaries with their latest release of FLUX.1 image generation model.
It's easy to fine-tune Flux, but sometimes you need to do a little more work to get the best results. This post covers techniques you can use to improve your fine-tuned Flux models.
Create and run your own fine-tuned Flux models programmatically using Replicate's HTTP API.
Create your own fine-tuned Flux model to generate new images of yourself.
Flux LoRAs, Hot Zuck, and Replicate on Lex Fridman
Fine tune FLUX.1, generative video games, a vision for the metaverse
We've added fine-tuning (LoRA) support to FLUX.1 image generation models. You can train FLUX.1 on your own images with one line of code using Replicate's API.
Flux developments, Minecraft bot, Streamlit cookbook with Zeke
We explore FLUX.1's unique strengths and aesthetics to see what we can generate.
Open source frontier image model, cut objects from videos, new Python web framework from Jeremy Howard
FLUX.1 is a new text-to-image model from Black Forest Labs, the creators of Stable Diffusion, that exceeds the capabilities of previous open-source models.
A top-tier open-ish language model, new safety classifiers, model search API
Llama 3.1 405B: is the most powerful open-source language model from Meta. Learn how to run it in the cloud with one line of code.
Data curation, data generation, data data data
Google's Gemma2 models, language model leaderboard, tips for Stable Diffusion 3
Really good coding model, AI search breakthroughs, Discord support bot
We show you how to use Stable Diffusion 3 to get the best images, including new techniques for prompting.
A step-by-step guide to generating images with Stable Diffusion 3 on your M-series Mac using MPS acceleration.
Create your own custom version of Stability's latest image generation model and run it on Replicate via the web or API.
Find concepts in GPT models, real-time speech to text in the browser, H100s are coming
Copy and paste a few commands into terminal to play with Stable Diffusion 3 on your own GPU-powered machine.
We'll soon support NVIDIA's H100 GPUs for predictions and training. Let us know if you want early access.
Stable Diffusion 3 is the latest text-to-image model from Stability, with improved image quality, typography, prompt understanding, and resource efficiency. Learn how to run it in the cloud with one line of code.
Garden State Llama, applied LLMs guide, real-time image generation
Faster image generation, AI-powered world simulator, insights on AI dataset complexity
DIY Llama 3 implementation, open-source smart glasses, steering language models with dictionary learning
Arctic is a new open-source language model from Snowflake. Learn how to run it in the cloud with one line of code.
Llama 3 is the latest language model from Meta. Learn how to run it in the cloud with one line of code.
Code Llama 70B is one of the powerful open-source code generation models. Learn how to run it in the cloud with one line of code.
Or, how I met a virtual David Attenborough.
We’ve added fine-tuning for realistic voice cloning (RVC). You can train RVC on your own dataset from a YouTube video with a few lines of code using Replicate's API.
We've raised a $40 million Series B led by a16z.
The Yi series models are large language models trained from scratch by developers at 01.AI. Learn how to run them in the cloud with one line of code.
We've added a CLI command that makes it easy to get started with Replicate.
An interactive example showing how to embed text using a state-of-the-art embedding model that beats OpenAI's embeddings API on price and performance.
We’ve added chord conditioning to Meta’s MusicGen model, so you can create automatic backing tracks in any style using text prompts and chord progressions.
How to run a latent consistency model on your M1 or M2 Mac
In this post we'll explore the basics of retrieval augmented generation by creating an example app that uses bge-large-en for embeddings, ChromaDB for vector store, and mistral-7b-instruct for language model generation.
We’ve added fine-tuning support to MusicGen. You can train the small, medium and melody models on your own audio files using Replicate.
How to use Llama 2 models with grammars for information extraction tasks.
Mistral 7B is an open-source large language model. Learn what it's good at and how to run it in the cloud with one line of code.
Combine AnimateDiff and the ST-MFNet frame interpolator to create smooth and realistic videos from a text prompt
We've made some dramatic improvements to cold boots for fine-tuned models.
With the recent release of Stable Diffusion XL fine-tuning on Replicate, and today being the 1-year anniversary of Stable Diffusion, now feels like the perfect opportunity to take a step back and reflect on how text-to-image AI has improved over the last few years.
The price of public models is being cut in half, and soon we'll start charging new users for setup and idle time on private models.
Learn the art of the Llama prompt.
Our API now supports server-sent event streams for language models. Learn how to use them to make your apps more responsive.
We’ve added fine-tuning (Dreambooth, Textual Inversion and LoRA) support to SDXL 1.0. You can train SDXL on your own images with one line of code using the Replicate API.
Llama 2 is the first open source language model of the same caliber as OpenAI’s models. Learn how to run it in the cloud with one line of code.
How to run Stable Diffusion XL 1.0 using the Replicate API
How to run Llama 2 on Mac, Linux, Windows, and your phone.
So you want to train a llama...
A roundup of recent developments from the llamaverse following the second major release of Meta's open-source large language model.
Prompt engineering and training are often the first solutions we reach for to improve language model behavior, but they're not the only way.
We've added a status page to provide real-time updates on the health of Replicate.
A roundup of recent developments from the world of open-source language models.
Give it a machine learning directory and AutoCog will create predict.py and cog.yaml until it successfully runs a prediction
A roundup of recent developments from the llamaverse.
With a small amount of data and an hour of training you can make LLaMA output text in the voice of the dataset.
We'll show you how to train Alpaca, a fine-tuned version of LLaMA that can respond to instructions like ChatGPT.
Lots of people want to build things with machine learning, but they don't have the expertise to use it.
It's like DreamBooth, but much faster. And you can run it in the cloud on Replicate.
With just a handful of images and a single API call, you can train a model, publish it to Replicate, and run predictions on it in the cloud.
How to run Stable Diffusion locally so you can hack on it
How to use Replicate to integrate Stable Diffusion into hacks, apps, and projects
A tutorial for building a chat bot that replies to prompts with the output of a text-to-image model.
We're bringing people together to explore what's being created with machine learning.
Using CLIP and LAION5B to collect thousands of captioned images.
The basics of using the API to create your own images from text.
Inspired by model cards, we've created templates for documenting models on Replicate.
An introduction to differentiable programming and the process of refining generative art models.
We're a small team of engineers and machine learning enthusiasts working to make machine learning more accessible.