Good

LoRA Setting1 => Setting2 (1, char.) Setting3
dim 128 128 128
alpha 1 1 64
scheduler cosine with restarts cosine with restarts ---
cosine restarts 5 or 10 5 ---
warmup ratio 0.1 0.1 ---
learning rate 3e-3 or 4e-3 4e-3 6e-5
text encoder 8e-5 8e-5 3e-5
unet lr 3e-3 or 4e-3 4e-3 16e-5 or 15e-5
num_workers --- 1 ---
batch size 3 3 2
ag_dropout --- 0.1 ---
caption_dropout --- 0.03 ---
res 512 --- 960
Images --- --- ---
epoch --- --- ---
repeats by image --- --- ---
Preview MEGA image Used for characters. With very good results image

Anon1's note

UNCONFIRMED: multiply 1e-4 by your batch_size to get your unet_lr set learning_rate equal to unet_lr multiply 1e-5 by your batch_size to get your text_encoder_lr If this math scares you, just type it into google or w/e to get the answer EX: 1e-4 * 12

Anon2's note

From my experience as I train lora for a while I feel like any parameter value will be okay as long as is not absurd as the priorities are: 1. your dataset 2. your tags 3. your parameter setting

Not as good

LoRA Setting2
dim 128
alpha 128
scheduler constant
cosine restarts ---
warmup ratio ---
learning rate 1e-4
text encoder 5e-5
unet lr 1e-4
num_workers ---
batch size 2
ag_dropout ---
caption_dropout ---
res 512
Images 35
epoch 1
repeats by image 100
Preview Anus Image Apparently not satisfactory.
Edit Report
Pub: 11 Feb 2023 22:45 UTC
Edit: 12 Feb 2023 14:29 UTC
Views: 962