CivArchive
    Preview undefined

    Full Checkpoint with improved TE do not load additional CLIP/TE

    FLUX.1 (Base UNET) + Google FLAN

    • All uploaded models sourced from 65GB Full FP32

    Per the Apache 2.0 license FLAN is attributed to Google

    Description

    FAQ

    Comments (31)

    RainM4kerOct 28, 2024· 1 reaction
    CivitAI

    This checkpoint doesn't work in diffusion loader? It doesn't work in the unet folder of models in comfyui, is there something else I need to set up?

    Felldude
    Author
    Oct 28, 2024

    You need the NF4 checkpoint loader if your using comfy, forge would be native. Forge also has Lora support for NF4 comfy does not right now

    RainM4kerOct 28, 2024

    @Felldude thx!

    RainM4kerOct 28, 2024

    @Felldude you said Do not load VAE, CLIP or TE - FP32 Quantized versions baked in Does this mean that I shouldn't load VAE, but only load one of CLIP or TE?

    Felldude
    Author
    Oct 28, 2024

    I am adding a FP8 version for easier Comfy UI use

    jervis314Oct 28, 2024· 4 reactions
    CivitAI

    What's the benefit of using this Google FLAN encoder over the normal t5xxl_fp8_e4m3fn?

    Felldude
    Author
    Oct 28, 2024

    Much higher accuracy by quantizing from FP32 on a model Google finetuned

    925_StudioOct 28, 2024· 3 reactions
    CivitAI

    Looks nice. Can you tell me how much vram we are talking about here? Perhaps you should upload more samples to show that flan is better than t5xxl.

    Felldude
    Author
    Oct 28, 2024

    Google themselves published the stats on the finetune of T5xxl, I am quantizing with the highest publicly available quality FP32

    Felldude
    Author
    Oct 28, 2024· 1 reaction

    Comfy and Forge have CPU offloading so VRAM shouldn't be an issue. I wouldn't attempt to use all the models in FP32 as that would be close to 100GB...it would be funny to post that checkpoint though. (Unfortunately Blackforest has never published a FP32 version of FLUX)

    theunlikelyOct 28, 2024· 2 reactions
    CivitAI

    Did you remove the decoder part of the FLAN encoder? If not, that would definite decrease the file size.

    Felldude
    Author
    Oct 28, 2024· 1 reaction

    Edit: The decoder block layer yes that gets trimmed, its reduces the weights by 11GB the models would be quite large otherwise

    Felldude
    Author
    Oct 28, 2024· 1 reaction

    The CLIP is trimmed to the model weights when merged.

    favouriteuser955Oct 28, 2024· 4 reactions
    CivitAI

    Can you share only the Google FLAN FP32/NF4/FP8 so we can test it also on other models?

    Felldude
    Author
    Oct 29, 2024

    So currently the model gets some block errors when extracting, I will have to look into if it is being pruned incorrectly or extracted incorrectly. I am leaning toward extracting incorrectly as I have no clip errors in the merged models

    jayepe6189480Oct 29, 2024

    @Felldude Getting full black image when using this clip only model (dual clip loader)

    Felldude
    Author
    Oct 29, 2024

    @jayepe6189480 If you will note in bold you should not be loading any clips

    ericreatorNov 3, 2024

    @Felldude Hey there Felldude, I'd like to try this out but it seems to have been "archived" - I can't download it. Currently testing off the model loader

    Felldude
    Author
    Nov 4, 2024

    @ericreator I released a simple tool to extract the T5 Flan

    sevenof9247Oct 28, 2024· 4 reactions
    CivitAI

    quality i will see

    but fits perfect my 16GB VRAM
    and its 25% faster than normal models (sec/it)

    hmmm seems a bit less details and for long prompts it dont get all objects like the original FLUX_DEVfp8 tested on 3 prompts a 10 image

    Felldude
    Author
    Oct 28, 2024

    The UNET is the same, regarding prompt adherence the logic should be mathematically better per google but that is meaningless if the results don't look aesthetically pleasing.

    It did appear faster but it can be hard to tell with CPU offloading

    StablediffusionloverOct 28, 2024· 2 reactions
    CivitAI

    well downloading now, going to test it on 8gb vram and 32gb of ram. Let you all know if it works or not (forgeUI)

    Felldude
    Author
    Oct 28, 2024

    I recommend the NF4 for forge

    LiteSoulHDOct 30, 2024

    How are you planning to fit a 15gb file into 8gb vram so it runs at decent speed? I don't see how's that possible

    Felldude
    Author
    Oct 30, 2024· 1 reaction

    @LiteSoulHD With CPU offloading, before that it was not possible at all

    devon_rex18406May 28, 2025

    It does using a 2070 it takes a good 90 to 120 seconds per image though.

    @LiteSoulHD with multigpu node gguf its possible!

    jazara930Jun 2, 2025

    both models work with Forge for me, I have 128Gb Ram and 3090 24Gb, but I prefer to use NF4 with Forge.

    GitarooManOct 29, 2024· 2 reactions
    CivitAI

    Is there a way to make photos look less comic/hdr? Like a natural photograph, soft lighting

    Felldude
    Author
    Oct 29, 2024

    The Dev/NF4 model will have the highest quality apart from Full FP16/BF16

    Checkpoint
    Flux.1 D

    Details

    Downloads
    216
    Platform
    CivitAI
    Platform Status
    Deleted
    Created
    10/28/2024
    Updated
    7/9/2025
    Deleted
    6/3/2025

    Files

    fluxDevSchnellBaseUNET_fluxDevFLANFP8.safetensors