You're looking at a specific version of this model. Jump to the model overview.

edenartlab /sdxl-lora-trainer:95542d9a

Input

string
Shift + Return to add a new line

Name of new LORA concept

string
Shift + Return to add a new line

Training images for new LORA concept (can be images or a .zip file of images)

string
Shift + Return to add a new line

face / style or concept (default)

Default: "concept"

integer

Random seed for reproducible training. Leave empty to use a random seed

integer

Square pixel resolution which your images will be resized to for training

Default: 896

integer

Batch size (per device) for training

Default: 2

integer

Number of epochs to loop through your training dataset

Default: 10000

integer

Number of individual training steps. Takes precedence over num_train_epochs

Default: 600

integer

Number of steps between saving checkpoints. Set to very very high number to disable checkpointing, because you don't need one.

Default: 10000

boolean

Whether to use LoRA training. If set to False, will use Full fine tuning

Default: true

number

Learning rate for the U-Net. We recommend this value to be somewhere between `1e-6` to `1e-5`.

Default: 0.000001

number

Scaling of learning rate for training textual inversion embeddings. Don't alter unless you know what you're doing.

Default: 0.0003

number

Scaling of learning rate for training LoRA embeddings. Don't alter unless you know what you're doing.

Default: 0.0001

number

weight decay for textual inversion embeddings. Don't alter unless you know what you're doing.

Default: 0.00001

number

weight decay for LoRa. Don't alter unless you know what you're doing.

Default: 0.0001

integer

Rank of LoRA embeddings. For faces 4 is good, for complex objects you might try 6 or 8

Default: 4

string

Learning rate scheduler to use for training

Default: "constant"

integer

Number of warmup steps for lr schedulers with warmups.

Default: 50

string
Shift + Return to add a new line

A unique string that will be trained to refer to the concept in the input images. Can be anything, but TOK works well

Default: "TOK"

string
Shift + Return to add a new line

Text which will be used as prefix during automatic captioning. Must contain the `token_string`. For example, if caption text is 'a photo of TOK', automatic captioning will expand to 'a photo of TOK under a bridge', 'a photo of TOK holding a cup', etc.

Default: "a photo of TOK, "

string
Shift + Return to add a new line

Prompt that describes part of the image that you will find important. For example, if you are fine-tuning your pet, `photo of a dog` will be a good prompt. Prompt-based masking is used to focus the fine-tuning process on the important/salient parts of the image

boolean

If you want to crop the image to `target_size` based on the important parts of the image, set this to True. If you want to crop the image based on face detection, set this to False

Default: true

boolean

If you want to use face detection instead of CLIPSeg for masking. For face applications, we recommend using this option.

Default: false

number

How blurry you want the CLIPSeg mask to be. We recommend this value be something between `0.5` to `1.0`. If you want to have more sharp mask (but thus more errorful), you can decrease this value.

Default: 1

boolean

Add left-right flipped version of each img to the training data, recommended for most cases. If you are learning a face, you prob want to disable this

Default: true

boolean

verbose output

Default: true

string
Shift + Return to add a new line

Subdirectory where all files will be saved

Default: "1693687184"

boolean

for debugging locally

Default: false

Output

No output yet! Press "Submit" to start a prediction.