openai/gpt-image-1.5

OpenAI's latest image generation model with better instruction following and adherence to prompts

409 runs

GPT Image 1.5

Generate and edit images with precise control. GPT Image 1.5 is OpenAI’s latest image generation model, built for production-quality visuals and controllable creative workflows.

What it does

GPT Image 1.5 handles two main workflows: creating images from text descriptions and editing existing images with specific instructions. It’s designed to follow your directions closely while keeping the parts you want unchanged.

The model runs up to 4x faster than its predecessor and costs 20% less per image. It’s particularly good at maintaining consistency across edits, change a person’s outfit without altering their face, adjust lighting without recomposing the scene, or translate text in an image while preserving the original layout.

Key capabilities

Photorealism and detail: The model produces natural-looking images with accurate lighting, believable materials, and rich textures. You can get everything from honest, unposed photography to polished commercial visuals.

Text rendering: Need readable text in your images? GPT Image 1.5 handles dense text, small lettering, and complex layouts like infographics, UI mockups, and marketing materials.

Precise editing: Make targeted changes without reinterpreting the entire image. The model preserves identity, composition, and lighting while you adjust specific elements.

Style control: Apply consistent visual styles across different subjects, or transfer the look of one image to another with minimal prompting.

World knowledge: Ask for a scene set in “Bethel, New York in August 1969” and the model understands you want Woodstock, it has reasoning built in.

Use cases

Image generation: Create infographics, logos, UI mockups, photorealistic scenes, comic strips, and marketing visuals. The model works well for both creative exploration and production-ready outputs.

Image editing: Style transfer, virtual clothing try-ons, product mockups, text translation in images, lighting adjustments, object removal, and scene compositing. You can insert people into new scenes while preserving their likeness, or swap furniture in room photos without changing the camera angle.

Character consistency: Build multi-page illustrations where characters look the same across different scenes, useful for children’s books, storyboards, and campaigns.

Example outputs

Infographic

Technical diagrams created from a text prompt explaining how an automatic coffee machine works.

Photorealistic image

Honest, unposed photography with natural skin texture and worn materials.

Logo generation

Multiple logo variations for a bakery called “Field & Flour” with clean, scalable designs.

Virtual try-on

Changed clothing while maintaining the person’s face, body shape, and pose.

UI mockup

Realistic mobile app design that looks like a shipped product.

Drawing to photo

Turned a rough drawing into a photorealistic scene while preserving the layout.

Scene compositing

Inserted a person into a new environment while maintaining their identity and realistic lighting.

Children’s book illustration

Character illustration with consistent appearance that can be reused across story pages.

How to get good results

Be specific: Describe what you want clearly. Instead of “make it better,” say “add soft coastal daylight” or “change the red hat to light blue velvet.”

Use photo language for realism: Mention lens type, lighting quality, and framing when you want photorealistic results. “Shot with a 50mm lens, soft daylight, shallow depth of field” gets you closer to real photography than generic quality terms.

Lock what shouldn’t change: When editing, explicitly state what must stay the same. “Change only the lighting, preserve the subject’s face, pose, and clothing” prevents unwanted alterations.

Put text in quotes: For readable text in images, put the exact copy in “quotes” and describe the typography. “Bold sans-serif, centered, high contrast” helps ensure legibility.

Iterate with small changes: Start with a base image, then make one adjustment at a time rather than rewriting everything.

Reference multiple images clearly: When working with several input images, label them by number and describe how they relate. “Apply the style from image 1 to the subject in image 2.”

Technical details

GPT Image 1.5 supports both generation (creating images from text) and editing (modifying existing images with text instructions). You can provide multiple input images for compositing and style reference.

The model includes quality and fidelity settings—use lower quality for faster generation in high-volume use cases, or higher quality when visual precision matters. For most tasks, the default settings provide good results.

For more prompting strategies and production workflows, see OpenAI’s prompting guide.


You can try this model on the Replicate Playground at replicate.com/playground

Model created
Model updated