zImageBase - V1.0
After quite a few test with different settings, including theory's without any proof, decided to just change all training settings till something works. Did a small test with only 6 images, different styles, photo, illustration etc. After 500 steps it pretty much nailed all images, low style bleeding if any, no flip-flopping of characters changing position, reduced hallucinations, learns faster ( well, it's slow but constant ) and the Lora has more influence even with lower strength ( in some cases you had to crank up the strength over 1.0 which induces artifacts )
This one is just another test with 20 images. Old images with a lot of junk in them, which was wanted to see if it tries to compensate mistakes or learns what is given.
This is a version with lower step count. The versions with higher steps where a bit too close to the source images and made the output very rigged even with different seeds. Even this one is kinda strong.
I don't trust it fully yet. Currently training stuff with a bigger dataset which takes some time.
AI-Toolkit - Prodigy ( non 8bit, at least i think, if i read the optimizer files correctly )
Rank 64 ( haven't tested lower ranks yet, just went with it )
LR: 1 ( Prodigy changes the learning rate dynamically )
noise_scheduler: "flowmatch"
optimizer: "prodigy"
timestep_type: "shift"
content_or_style: "balanced"
optimizer_params:
weight_decay: 0.01 ( still needs testing if needed at all, changed it a few times )
d_coef: 1.0 ( also needs testing, changed it to 1.5 for the bigger dataset for now )
decouple: true
use_bias_correction: false
Take all of this with a big grain of salt ... actually buy a whole salt-mine. With AI it seems no one knows how and why shit works or not. Even the people who made the models don't know what the fuck is going on, all just theory and math. High chance that i'm on the wrong path too, wouldn't be the first time. It did what it's supposed to do for now, so i'll stick with it till problems occur.
V.2.0
Added more pictures and higher step count.
Recommended to use a low strength or you turn everything into a tentacle monster for whatever reason.
Really wish they would put out the Pro version of Flux since this distilled stuff is kinda hard to control and also limited in its flexibility. If you play around with some prompts it always goes to a certain image or something it knows good and is trained on, which results in Flux typical steril look ( cinematic, photo, certain animals or people etc )
I also highly doubt that any of the full checkpoints out there that are trained will ever work. Went trough pretty much every single one of them, but it always comes down to fucked up anatomy or a hardcore bias which you can't negate with negative prompts or prompt weighting since Flux doesn't use this stuff. A picture of a person that is always naked, even if you ask for clothes for example. Doubtful if that even did/does anything in SDXL, never really used that much negative prompts, but it was good enough to get rid of some things you tried to avoid.
Strangely enough, all checkpoints that are trained or LoRAfied with a style ( like anime ) work great, even for realistic images, which i pretty much use for all of my pictures ( not here, basic FP8 checkpoint for showcase reason )... so confusing.
Trained on a dataset which i planned to use for SDXL, but i never got satisfying results . Just a small test with a few images ( basic capitations for now ) and only 800 steps. Will change that later to natural language.
Most used words should be atmospheric, moody, calm, soothing, serene, mysterious ... and assorted ( the images without capitation ... well, that is a capitation, forgot that Kohya picks the name of the folder if no .txt file is present :D )
Order: First 2 Images with LoRA / without LoRA, after that, reversed order
Has more/less impact in specific cases ( for now )
Did quite some testing with Flux LoRA's I've made and got really crazy results. Even after only like 100 - 200 steps it got the concept and lower/higher strength from base 1 has always a huge impact, but something is always left and it feels like you can kinda pick specific parts from the images it was trained on without it using the whole image ( like you only want the the yellow clothing but nothing else )
Made like 30 LoRA's so far ( only for testing purposes ) and what you can do with just a few images is baffling.
Might be just a fluke, who knows.
Description
FAQ
Comments (7)
I'm not sure why, but some Loras cause my generation time in Forge to go from 30 seconds a generation to 3-4 minutes per generation. This is one of the Loras that makes Forge almost grind to a halt.
I had the same problem with a few LoRA's, but mostly with NF4 models. I was watching the VRAM usage and it went trough the roof, from like 13.7 GB to 22 GB or something, which means it was compensating with system memory... slow as fuck of course.
Currently using a merged FP8 checkpoint ( clip + text encoder inside), no NF4 or anything else, just normal Flux Dev, no problems with any LoRA, no matter how many at once.
@TijuanaSlumlord Intersting. I'm using the FP8 version in forge. I should test it in Comfy to test if it's just a forge issue.
I like the style of this lora though! Very cool. Thanks for making it.
Make sure your "Diffusion in Low Bits" is set to Automatic (fp16 LORA) or bnb-nf4 fp16 LORA if you are using nf4.
@TijuanaSlumlord I was having the same slow issue with this and other loras I was using flux1-dev-bnb-nf4-v2. thanks for the tip I will try out an FP8 merge. Could you possibly recommend the current best one to use? Thank you for this amazing lora by the way!
@rlewisfr346 oh great tip I never even noticed this setting was even there! Thx
@StecFX Stupid Civit not showing comment notifications again.
Not sure but i think it was this one https://civitai.com/models/637170?modelVersionId=712441 , but i haven't used it for quite some time after the model post. Also, it seems like NF4 is fixed for the moment in Forge, but i usually stay with normal FP8 checkpoints because apart from memory usage i don't see the benefit of NF4 ( speed increase would be nice ). The whole NF4 or QQUF.. q4.....QQ 0.432463455 .... %#!?!& or wtf ever is just extremely confusing and i see a lot of people always asking on other model post's if something works or not, but with Flux being a resource hog, testing other peoples LoRA's and training there is no time to test every single model variation there is.
Details
Files
Available On (3 platforms)
Same model published on other platforms. May have additional downloads or version variants.



















