Collections

Edit any image

What you can do

Remove or replace objects in your images.

Change the style of your images (e.g. "make this Studio Ghibli style")

Add text that looks natural. Generate images with text that matches specific fonts and styles.

Change specific parts of an image while keeping its structure. Use depth maps or edge detection to control what changes.

Create variations of your images. Keep what works while exploring new possibilities.

Models we recommend

For precise edits

Nano-banana could truly be the end of Photoshop. Great at following instructions and supports multi-image input.

FLUX.1 Kontext Pro allows you to edit images using text. Here are some sample prompts:

  • "Add a plant next to the woman"
  • "Change the text in the image from 'Hi' to 'Hello'"
  • "Turn me into a Simpsons character"

FLUX.1 Kontext Max is the advanced version of FLUX.1 Kontext Pro. It is much more powerful.

gpt-image-1 from OpenAI is also a solid choice for image editing and supports multiple image inputs (i.e. "combine all of these images into one scene"). (This model requires an OpenAI API key.)

Learn more about FLUX Tools →

For text and typography

Ideogram v3 excels at:

  • Adding text that looks natural
  • Matching specific fonts and styles
  • High-quality general inpainting

Need faster results? Try Ideogram v3 Turbo. Or learn more about running Ideogram models with an API.

For style changes

These models help you make targeted changes while preserving image structure:

Try it out

Test different editing approaches in the playground. Compare models side by side to find what works best for your project.

Open the playground →

Want to learn more about inpainting? Check out our guide →

Questions? Join us on Discord.

Frequently asked questions

Which models are the fastest?

If you want quick edits to an image, google/nano-banana is a strong choice—it handles editing using simple instructions in text, and it supports multi-image input.

Another fast option is bytedance/seedream-4, which supports editing at higher resolution without huge compute time.

Which models offer the best balance of cost and quality?

For reliable edits with good prompt following and identity preservation, black-forest-labs/flux-kontext-pro offers a solid middle ground.

If you’re working on a more premium workflow (typography, precise edits, full stylization), black-forest-labs/flux-kontext-max scales up quality and control.

What works best for removing or replacing objects in an image?

If you need to remove or swap items (for example a sign, person, or piece of furniture), models like bria/eraser and bria/genfill are designed for clean object removal and visual continuity.

For broader edits—such as changing a background or adding new elements while retaining structure—black-forest-labs/flux-kontext-pro works well with directed prompts.

What works best for changing style or adding text in images?

If your edit centers on changing style (for instance “make this image look like a Studio Ghibli painting”) or adding text that looks like it belongs, ideogram-ai/ideogram-v3 is excellent for natural typography and stylized inpainting.

For depth-aware or edge-preserving edits (e.g., changing pose or structure while keeping main subjects intact), black-forest-labs/flux-depth-pro or black-forest-labs/flux-canny-pro provide more control.

What’s the difference between key subtypes or approaches in this collection?

There are two main approaches:

What kinds of outputs can I expect from these models?

You’ll typically get an edited image file (same resolution or slightly changed depending on settings) where the requested modifications have been applied.

Some models also output additional metadata or allow multi-image input (for example combining two photos or layering edits) depending on the version.

How can I self-host or push a model to Replicate?

If the model is open source, you can clone the repo and run it locally using Cog or Docker.

To publish your own model, prepare a replicate.yaml file that defines inputs (image, mask, prompt) and outputs, then push it to your Replicate account for use on managed hardware.

Can I use these models for commercial work?

Yes—many image-editing models support commercial use. Always check the License section on each model’s page to confirm.

Also ensure you have rights to the image you are editing, especially if you plan to publish or monetize the output.

How do I use or run these models?

Open a model’s page on Replicate, upload your image (and optional mask or reference image), and enter a prompt describing your edit (e.g., “change the car colour to red and add a banner”).

The model will return a modified image you can download. Some models support further options like preserving identity, controlling style strength, or combining multiple inputs.

What should I know before running a job in this collection?

  • Use a good quality input image (well lit, clear subject) for best results—editing messy inputs is harder.
  • If you want to keep a person’s identity, mention that in the prompt (e.g., “keep the same face, just change the outfit”).
  • For structural edits (pose, background, mask), models that support control maps (depth/edges) will yield better fidelity.
  • Complex edits are often better performed in stages (first replace object, then refine style) rather than one large prompt.

Any other collection-specific tips or considerations?

  • For combining images (e.g., merge two photos into one scene), pick models that accept multi-image input, like black-forest-labs/flux-kontext-max or qwen/qwen-image-edit.
  • For text overlays (signs, posters), prefer models optimized for typography (like ideogram-ai/ideogram-v3).
  • When doing style transfer (photo → painting, sketch, anime), be explicit in your prompt about the style and what to keep.
  • Always check that your output respects original subjects (especially people), and if you’re using commercial content verify rights accordingly.