I have trained Flux Dev on my SDXL dataset and merged loras, correcting anatomy censorship and excessive bokeh/blurred backgrounds.
V12 - Merge with SRPO model and some loras, a bit cleaner looking than previous models, might want to use some grain loras if you want that amateur look.
I think not as good at NSFW so might want to use my lora for that: https://civarchive.com/models/652791/jibs-flux-nipple-fix
Jib Mix Flux v8-Flash! SVDQuant-4bit
Update 06/04/2025 - NOW WITH CONTROLNET AND NATIVE LORA SUPPORT!
This new SVDQuant format currently requires ComfyUI and an Nvidia 2000 series+ GPU.
I recommend;
Guidance scale = 2.5-3.5
Sampler = dpmpp_2m.
Scheduler = Sgm_uniform, Beta or Custom Stigmas.
This model is a new format that is a little tricky to first get setup but it is worth it as you can make flux images super fast (5 seconds on a 3090, 2.5 sevonds on a 4090 and 0.8 seconds on 5090! at 10 steps) ,
You need to install the nunchaku project following the instructions there.
and the nunchaku ComfyUI custom nodes to get it to work.
Download and unzip the archive from Civitai to: \comfyui.git\app\models\diffusion_models\jib-mix-svdq\
This is the Nunchaku workflow I am using: https://civarchive.com/models/617562
Thanks a lot to theunlikely for running the Quantization which takes an H100 GPU 6 hours!
Unfortunately, it seems the NSFW capabilities are reduced by the quantisation but can be bought back with my NSFW Lora you need to use the Nunchaku lora loader in the node pack and it will auto convert the lora for use the first time (and optionally can be saved).
There is currently a size limit above 2 million pixels 1024 x 2048 that causes a crash in ComfyUI, so if you get that lower your gen/upscale resolution, The Developers say a fix will be released this week.
Jib Mix Flux Version 8 AccentuEight
Much better Skin texture than Jib Mix Flux V7 but without the bad Flux Lines of Jib Mix V6.
NSFW anatomy is slightly lacking so I have uploaded a separate NSFW model that has very slightly reduced details / artstyles or loras can be used.
The V8 Pruned Model nf4 (6.33 GB) model is actually the Q4_0.gguf
The V8 Pruned Model bf16 (11.84 GB) model is actually the Q8_0.gguf
15/04/2025 - The new fp16 version of V8 provides very slightly better details and anatomical and image consistency, if you have the VRAM for it....
Jib Mix Flux Version 8 AccentuEight NSFW
Better female anatomy with a slight loss in skin details.
The V8 NSFW Pruned Model nf4 (6.33 GB) model is actually a Q4_0.gguf
Jib Mix Flux Version 7.8 Clear Text Focus
This version focuses on having more readable text generation.
The Skin texture may not be generally as realistic as Jib Mix v6 or v7.
It has less nipple flashing through clothes.
The 2000s Analog Core @0.6 weight, removes some of the plastic Fluxness of this merge without hurting the text.
Jib Mix Flux Version 7.2 Pixel Heaven
7/7.2 mainly fixes "Flux Line" caused by merging some effected lora's, this has caused some drop in photo realism but an increase in drawn/concept ability and general details.
It also tones down obsessive amounts of freckles (especially red freckles).
The V7.2 Pruned Model nf4 (6.33 GB) model is actually the Q4_0.gguf
The V7.2 Pruned Model bf16 (11.84 GB) model is actually the Q8_0.gguf
Jib Mix Flux version 7 PixelHeaven - beta
The main change is it removes "Flux lines" that plagued V6 and the original Flux Dev to some extent.
It may be overdoing freckles a lot but I wanted to see what people think of it. hence the Beta name.
I really recommend using Movie Portrait Lora on quite a high weight for a less plastic look, but I couldn't merge it into the model as in testing it was a Lora that can cause "Flux Lines"
Jib Mix Flux Version 6.1 Real Pix Fixed
6.1 mainly tries to fix small stubby hands or massively distorted hands/arms.
(If you still have problems with hands and you are using a low step count around 8 then increasing the step count usually fixes it alternatively applying a low weight (< 0.10) of the Hyper Flux lora pretty much always fixes hands, although you may find it lowers details).
V6.1 is more realistic on the CFG / more natural faces.
I think it actually does art/cartoon styles a bit better than the original v6 as well.
The V6.1 Pruned Model nf4 (6.33 GB) model is actually the Q4_0.gguf
The V6.1 Pruned Model bf16 (11.84 GB) model is actually the Q8_0.gguf
V6 still has the most detailed backgrounds in my testing.
The V6 Pruned Model nf4 (11.84 GB) model is actually Q8_0.gguf
The V6 Pruned Model bf16 (6.33 GB) model is actually a Q4_0.gguf
Jib Mix Flux Version 5 - It's Alive:
Improved photorealism. (Less likely to default to painting styles)
Fixed issues with wonky text.
More detailed backgrounds
Reconfigured NSWF slightly
fp8 V4 Canvas Galore:
better fine details and much better artistic styles, and improved NSFW capabilities.
fp8 V3.0 V3.1 - Clarity Key
I initially uploaded the wrong model file on the 21/10/2024, it was very similar but the new file since 22/10/2024 has slightly better contrast and was used for the sample images.
This version Improves detail levels and has a more cinematic feel like the original flux dev.
reduced the "Flux Chin"
Settings - I use a Flux Guidance of 2.9
Sampler = dpmpp_2m.
Scheduler = Beta or Custom Stigmas.
FP8 V2 - Electric Boogaloo: Better NSFW and skin/image quality.
Settings:
I find the best settings are a guidance and 2.5 and a CFG of 2.8 (although CFG does slow down the generation).
When using Loras these values may/will change.
Version: mx5 GGUF 7GB v1
This is a quantized version of my Flux model to run on lower-end graphics cards.
Thanks to @https://civarchive.com/user/chrisgoringe243 for quantizing this, it is really good quality for such a small model.
There are larger-sized GGUF versions available here: https://huggingface.co/ChrisGoringe/MixedQuantFlux/tree/main
for mid-range graphics cards.
Version 2 - fp16:
For those with high Vram Cards who want maximum quality I have created this merge with the full fp16 Flux model. if you "only have 24GB of Vram" you will need to force the T5 text encoder onto the CPU/System RAM with this force node on this pack:
https://github.com/city96/ComfyUI_ExtraModels
Those people waiting for a smaller quantized model I am still looking into it.
Version 2:
Merged in 8 Step Hyper Lora and some others.
Settings:
I like a Guidance of 2 and 8-14 steps.
Resolution: I like a around 1280x1344
Version 1 : brings some of the benefits and look of SDXL with the massive prompt adherence benefits of Flux.
Settings:
I like a Guidance of 2 and 20-40 steps.
Description
Better dynamic poses.
A more natural amateur photographic look by default.
More detailed skin texture.
FAQ
Comments (24)
Is it on tensor? God oh God. Just tried it in comfyui, the images are INSANE. but a little bit of feedback: the images look a lot realistic than your previous model and some other models, but when talking about photorealism, it is essential for models to generate images with:
1) sharp or smooty skin textures (Based on prompt)
2) amateur or "professional " feel (based on prompt)
3) that "real-lifeliness".
Actually your model lacks in producing "real life like" images although textures are incredible. A tip: if you merge "UltraReal Finetune" with your next model, the results would be unmatched.
Thanks, yes, I guess my worry is that if I tune it to look amateur and graining/noise simple images that it will no longer be able to do the proffessional looking images.
There are loads of great amateur Loras that can bring down the image quality like xlabs Realisim https://civitai.com/models/631986?modelVersionId=706528 but I don't know of anything that can easily increase the image quality on a checkpoint that doesn't do that by default.
@J1B you are right! But actually, it is possible (have a look at Verus Vision). If you include a fixed no of amateurish, professionally cinematic images, that can help.
And also, it would be better to finetune "Dev Destilled" model, it will be much better.
How much Vram to run this?
12GB to run this FP8 model if you force the T5 Text encoder to run on the CPU in comfyUI (Advised).
I have a 6.4GB https://civitai.com/api/download/models/1028151?type=Model&format=GGUF&size=pruned&fp=nf4 version of my V5 (So likely 8GB Vram needed for that one)
Plz put it on tensorart
Amazing model, have used the v5 GGUF version in Forge on a RTX 2060. Works beautifully! Any chance you'll make GGUF of v6?
Yes, on RTX 2600 SUPER the v5 GGUF works great. And from the quality standpoint it is the best checkpoint I ever testest so far for lower cards. It creates professional pictures. Behold this unique line please! Just Loras seems not to work. At least not all. Would it work as NF4 output?
@Wadka Thanks. I think all Quantized modele types struggle a bit with lora compatibility, have you tried setting the lora wieght higher to 1.5-2 ? I have heard that can help get more accurate results when using with .GGUF or .NF4 models.
Loving v6 so far!
Also, just have to say that I list JibMix in almost every single one of my LoRAs' "suggested resources" sections because it pretty much always makes everything better 👍
What is the recommended number of steps for v5 Q4 GGUF? I saw earlier versions was 8-14?
Yes it can go as low as 5 steps (but will not be very consistently good) and the image can look a little noisy/amateur (fine if that is the look of image you are going for) to get best quality I use 14, or 20 if you are willing to wait longer.
I am in love with pictures, that can make this model. You are great! But unfortunately I can't use it cuz of bug with Apple M1. It works only with Qx_0 (not _1 or K_S). It will be great, if you make next model quantization with Qx_0 parameter
Thanks, Ok, I will give that a try, unfortunately the script I used to make the .GGUF for V5 isn't working on my PC right now. But I will see what I can do in the next few days.
I have added a Q8_0 Under NF4 for version 6 now, let me know if you want another sized .gguf _0 quant.
@J1B finally it works! Big thanks man! 👍🏼️ Q4_0 will be great. Q8 a bit oversized 😅️ Thank you again!
A Q4_0 is here https://civitai.com/api/download/models/1096166?type=Model&format=GGUF&size=pruned&fp=bf16 under bf16.
Can you also add V6 in GGUF format ?
hello, is the v6 has been fixed ?
I have got something ready without the possibly offending lora mixed in, but my extensive hand testing showed it wasn't really just an issue with Jib Mix V6 but most Flux models have issues with hands on some particular seeds (V6 exaggerated it slightly) , one of the best ways I found to fix it was to add in some hyper lora but that drastically reduced overall detail, which I cannot do. Another way to fix it nearly all the time was just to increase the number of steps.
V1 is still cencored
Still got Flux chin, plastic skin and bad unrealistic nipples
All those things are a lot better than in the original Flux Dev model.
Details
Files
Available On (1 platform)
Same model published on other platforms. May have additional downloads or version variants.











