Official

batchbinary / flux-qfaces

Flux trained on QFACES of High quality facial textures of diverse people, age, ethnicity. Mainly used for image to image (img2img) enhancement. See Readme for ideas & settings.

  • Public
  • 2.4K runs
  • H100

Input

*string
Shift + Return to add a new line

Prompt for generated image. If you include the `trigger_word` used in the training process you are more likely to activate the trained object, style, or concept in the resulting image.

file
Preview
image

Input image for image to image or inpainting mode. If provided, aspect_ratio, width, and height inputs are ignored.

number
(minimum: 0, maximum: 1)

Prompt strength when using img2img. 1.0 corresponds to full destruction of information in image

Default: 0.8

string

Which model to run inference with. The dev model performs best with around 28 inference steps but the schnell model only needs 4 steps.

Default: "dev"

integer
(minimum: 1, maximum: 50)

Number of denoising steps. More steps can give more detailed images, but take longer.

Default: 28

number
(minimum: 0, maximum: 10)

Guidance scale for the diffusion process. Lower values can give more realistic images. Good values to try are 2, 2.5, 3 and 3.5

Default: 3

integer

Random seed. Set for reproducible generation

string

Format of the output images

Default: "webp"

integer
(minimum: 0, maximum: 100)

Quality when saving the output images, from 0 to 100. 100 is best quality, 0 is lowest quality. Not relevant for .png outputs

Default: 80

boolean

This model’s safety checker can’t be disabled when running on the website. Learn more about platform safety on Replicate.

Disable safety checker for generated images.

Default: false

boolean

Run faster predictions with model optimized for speed (currently fp8 quantized); disable to run in original bf16

Default: false

string

Approximate number of megapixels for generated image

Default: "1"

number
(minimum: -1, maximum: 3)

Determines how strongly the main LoRA should be applied. Sane results between 0 and 1 for base inference. For go_fast we apply a 1.5x multiplier to this value; we've generally seen good performance when scaling the base value by that amount. You may still need to experiment to find the best value for your particular lora.

Default: 1

string
Shift + Return to add a new line

Load LoRA weights. Supports Replicate models in the format <owner>/<username> or <owner>/<username>/<version>, HuggingFace URLs in the format huggingface.co/<owner>/<model-name>, CivitAI URLs in the format civitai.com/models/<id>[/<model-name>], or arbitrary .safetensors URLs from the Internet. For example, 'fofr/flux-pixar-cars'

number
(minimum: -1, maximum: 3)

Determines how strongly the extra LoRA should be applied. Sane results between 0 and 1 for base inference. For go_fast we apply a 1.5x multiplier to this value; we've generally seen good performance when scaling the base value by that amount. You may still need to experiment to find the best value for your particular lora.

Default: 1

Including mask and 4 more...
file

Image mask for image inpainting mode. If provided, aspect_ratio, width, and height inputs are ignored.

string

Aspect ratio for the generated image. If custom is selected, uses height and width below & will run in bf16 mode

Default: "1:1"

integer
(minimum: 256, maximum: 1440)

Height of generated image. Only works if `aspect_ratio` is set to custom. Will be rounded to nearest multiple of 16. Incompatible with fast generation

integer
(minimum: 256, maximum: 1440)

Width of generated image. Only works if `aspect_ratio` is set to custom. Will be rounded to nearest multiple of 16. Incompatible with fast generation

integer
(minimum: 1, maximum: 4)

Number of outputs to generate

Default: 1

Output

output
Generated in

Run time and cost

This model costs approximately $0.016 to run on Replicate, or 62 runs per $1, but this varies depending on your inputs. It is also open source and you can run it on your own computer with Docker.

This model runs on Nvidia H100 GPU hardware. Predictions typically complete within 11 seconds.

Readme

Flux trained on ‘QFACES’ = High Quality Facial textures and photos of diverse people, age, ethnicity. Use for improving existing photos with Flux, while maintaining the core details of the image and people within it (img2img, image to image). Best with images of people close up but can generate content, by adding a content description to the prompt. If you try to upscale an image with multiple people at a distance, then increase the ‘Prompt Strength’ and ‘Guidance Scale’ and add a better description of what the photo contains in the prompt. Remember the higher these values the more the model will generate infill content (i.e. ‘make things up’). If you want to upscale an image of someone you or others can identify ( a family member ) then keep these values low and use a close up picture.
If you are looking to improve the facial textures of a picture of someone you know & recognise, then copy the following settings into the JSON tab. (Yes, you can set the Guidance Scale to 0, this works well to keep the person looking familiar), but prime the model with a good prompt and use the training keyword QFACES in the prompt to get better likeness:


{
“model”: “dev”,
“prompt”: “photo, high resolution, faces, people, high quality, QFACES, photograph, “,
“go_fast”: false,
“lora_scale”: 0.5,
“megapixels”: “1”,
“num_outputs”: 1,
“aspect_ratio”: “1:1”,
“output_format”: “png”,
“guidance_scale”: 0,
“output_quality”: 80,
“prompt_strength”: 0.3,
“extra_lora_scale”: 1,
“num_inference_steps”: 50
}
Note: if you copy paste this JSON code and it has BR commands in (i.e. HTML styling is copied in), you need to remove the BR codes first.
Important: the images used to fine tune (QFACE) this model are all copyright free public domain.