An experimental LoCon trained on outputs from my MIX-GEM-T2_2 model (and a few other MIX-GEM outputs to make up the gap). I spent a lot of time finetuning that model to my ideal aesthetic and I'd rather try to retrieve the style directly from the model than try to remix on a new SDXL base from scratch. Outputs are not very clean and this LoCon has a lot of issues. I will likely have to regenerate the dataset a couple of times with cleaner outputs. Likely there will be a lot of versions of hits LoCon, this will be an iterative process with a lot of rebakes.
Insights gleaned from prototyping:
Prodigy is worse than AdamW8bit at training style LoCons on PonyXL, even at a higher learning rate it retains a lot less than AdamW8bit. But it also destroys the base model's posing a lot faster, whereas the prodigy tends to keep a lot better with the original posing.
LoCons are better at training for styles than LoRAs.
Style retention comes hand in hand with magnifying small mistakes. This isn't a huge issue with ordinary style training, but is extremely problematic when training on SD1.5 outputs because of the way that unnecessary noise gets diffused into random elements which don't really makes aesthetic sense. Case has to be put into selecting only clean outputs.
Things to try in the future:
White background regularization images
Hiding hands as much as possible
Using copyright characters as part of the dataset
After testing, for some reason this LoCon works poorly on autismMixSDXL which washes out a lot of the details, but works extremely well on 4th tail.
Description
Extremely messy dataset, just a collection of MIX-GEM-T2_2, T3, and T2A images taken off aibooru
network_dim = 16
network_alpha = 8.0
min_timestep = 0
max_timestep = 1000
network_train_unet_only = true
conv_dim = 8
conv_alpha = 4.0
FAQ
Comments (13)
Hello! I noticed that you used Lora in your latest generation with PonyXL, however, your generations use <lyco:MIX-GEM-D1 adamw8bit-P1:1> and not <lora:Prodigy:1>. This means that we do not get the same result as you at all. In the download, only <lora:Prodigy:1> is available. Is it possible for you to provide us with the version <lyco:MIX-GEM-D1 adamw8bit-P1:1> that you used in all your image generations?
My apologies, I meant to upload the AdamW8bit and accidentally uploaded the Prodigy one instead. The correct version should be up now.
@Jemnite Thank you very much! I have another question, what is the difference between <lyco:MIX-GEM-D1 adamw8bit-P1:1> and <lora:locon-adamw8bit-r2:1> ? Could you also possibly upload <lyco:MIX-GEM-D1 adamw8bit-P1:1>?"
@GiggleGuru There's no difference, it's just renamed.
Would it be possible for you to reupload the prodigy version as well? I liked both a lot
I don't have any examples images, but it's here if you want it.
https://huggingface.co/Jemnite/experiments/blob/main/ponyfuckery/MIX-GEM-D1%20prodigy-P1.safetensors
Tell me your HF username when you request access.
@Jemnite Thank you so much!!! Sorry for the late reply just saw this now!
@Jemnite Hey, found your stuff recently and liked the works someone I follow did with your Prodigy version. I tried clicking the link, but it's 403'd. Do you still have it so that I could download it too?
hi, just came across your LoCon and is truly amazing! hope you don't mind telling more about the training process, what Alpha/DIM values did you use? how many epoch/repeats and on how many training images? what Learning Rate? thanks for your hard work!
I ran for 20 epochs with 2 repeats, but the one I uploaded was e08 or e12 iirc. It overtrained a bit in the end.
[network_args.args]
network_dim = 16
network_alpha = 8.0
min_timestep = 0
max_timestep = 1000
network_train_unet_only = true
[optimizer_args.args]
optimizer_type = "AdamW8bit"
lr_scheduler = "cosine"
learning_rate = 0.0005
max_grad_norm = 1.0
min_snr_gamma = 8
lr_scheduler_type = "LoraEasyCustomOptimizer.CustomOptimizers.CosineAnnealingWarmupRestarts"
lr_scheduler_num_cycles = 4
unet_lr = 0.0005
text_encoder_lr = 0.00025
warmup_ratio = 0.1
The training data is zip'ed up and available for download.
@Jemnite hi, I just downloaded the training data, there are NPZ files for each image, do I need those for the training as well? by the way, did you use PonyDifussionXL as the base? or which model it was? thanks a lot!
@pohttems The base is Pony Diffusion. You probably don't need the NPZ files, just the captions. I had latent caching on so that's what those are.
Can you update this to be closer to the sd 1.5 version?



















