CivArchive
    FLAN-T5-XXL (Text-Encorder only) - Q4_K_M
    Preview 64230350
    Preview 64230356

    FLAN-T5-XXL (Text-Encoder Only)

    The FP8 and GGUF format is distributed as a compressed ZIP file. Please unzip it using any decompression software of your choice before use, or download from Hugging Face page.

    FLAN-T5-XXL is a fine-tuned version of T5-XXL v1.1, designed to improve accuracy and performance.

    The original FLAN-T5-XXL model is available on Google's Hugging Face page.

    When used with Flux.1, SD3.5 and HiDream, replacement for T5-XXL v1.1 to FLAN-T5-XXL offers improved prompt comprehension and enhanced image quality.

    This model has been streamlined by extracting only the text encoder portion, making it optimized for image generation workflows.

    Model Variants

    +------------+--------+----------+-----------+
    | Flan-T5xxl |  Size  | Accuracy | Reccomend |
    +------------+--------+----------+-----------+
    |    FP32    |  19 GB |  100.0 % |     🌟    |
    |    FP16    |  9.6 GB|   98.0 % |     ✅    |
    |    FP8     |  4.8 GB|   95.3 % |     🔺    |
    |    Q8_0    |  5.1 GB|   97.6 % |     ✅    |
    |    Q6_K    |  4.0 GB|   97.3 % |     🔺    |
    |   Q5_K_M   |  3.4 GB|   94.8 % |           |
    |   Q4_K_M   |  2.9 GB|   96.4 % |           |
    +------------+--------+----------+-----------+

    Usage Instructions

    Place the downloaded model files in one of the following directories:

    • models/text_encoder

    • models/clip

    • Models/CLIP

    Select this model in place of the standard T5-XXL v1.1 model in your workflow.

    FP32 format

    The FP32 format provides the highest image quality.

    Stable Diffusion webUI Forge

    To use the text encoder in FP32 format, launch Stable Diffusion WebUI Forge with the --clip-in-fp32 argument.

    ComfyUI

    You can also use FP32 text encoders for optimal results by enabling the --fp32-text-enc argument at startup.

    GGUF format in ComfyUI

    As of April 13, 2025, ComfyUI's default DualClipLoader node includes a device selection option, allowing you to choose where to load the model:

    • cuda → VRAM

    • cpu → System RAM

    Since Flux.1’s text encoder is large, setting the device to cpu and storing the model in system RAM often improves performance.

    Unless your system RAM is 16GB or less, keeping the model in system RAM is more effective than GGUF quantization. Thus, GGUF formats offer limited benefits in ComfyUI for most users due to sufficient RAM availability.

    For running Flux.1 in ComfyUI, use the FP16 or FP32 text encoder.

    More about ComfyUI settings.

    Comparisons

    Tip: Upgrade CLIP-L Too

    For even better results, consider pairing FLAN-T5-XXL with an upgraded CLIP-L text encoder:

    Combining FLAN-T5-XXL with an enhanced CLIP-L model can further boost image quality.


    License

    This model is based on Google's FLAN-T5-XXL, also licensed under Apache 2.0.


    Update History

    August 22, 2025

    Add Why Use FP32 Text Encoder?

    July 24, 2025

    Re-upload of the GGUF model, reduction in model size, and correction of metadata.

    Description

    FAQ

    Comments (10)

    timtimberJun 3, 2025
    CivitAI

    interesting.
    but i don´t see a huge difference compared to t5xxl_fp16.
    did i miss something?

    easygoing0114
    Author
    Jun 12, 2025· 1 reaction

    The difference in image quality is not significant and can only be distinguished using tools.

    For the differences between t5xxl and flan-t5xxl, matataByy has written an article in Japanese, so please check it using your browser's translation function.

    https://note.com/matatabyy/n/n26385b13782d

    zerocool22Jul 1, 2025
    CivitAI

    The Q8 is from the FP32?

    If i'm using FP16, it makes sense trying the Q8?

    easygoing0114
    Author
    Jul 2, 2025· 1 reaction

    The Q8 model was created from the FP32.gguf file.

    If FP16 is working well for you, there’s no need to switch to Q8.

    zczcgJul 14, 2025
    CivitAI

    It`s a big different between fp16 and fp32?

    easygoing0114
    Author
    Jul 24, 2025

    There is a difference between FP16 and FP32 — but it's subtle.

    When used together with the improved CLIP-L (FP32) encoder mentioned in the model description, you may observe better consistency and detail in generation quality.

    https://www.ai-image-journey.com/2024/12/image-difference-t5xxl-clip-l.html

    xiongvang12851285Sep 16, 2025

    Personally there's no different.

    FranciscoMSep 8, 2025
    CivitAI

    the sha256 from zip will never be the same as the model so the images cant be link to this model

    easygoing0114
    Author
    Sep 9, 2025

    Civitai does not support uploading GGUF format, and FP8 format also could not be uploaded in this case, so I have compressed it into ZIP format.

    The original model is uploaded to Hugging Face, where you can verify the hash value.

    Example: flan_t5_xxl_TE-only_FP8.safetensors model

    https://huggingface.co/easygoing0114/flan-t5-xxl-fused/blob/main/flan_t5_xxl_TE-only_FP8.safetensors

    FranciscoMSep 9, 2025· 1 reaction

    @easygoing0114 thank you. It's just that when we click the little magnifying glass in Reforge, it opens the Civitai page directly. I've seen other models in GGUF and q8 or fp8 formats; I didn't know some couldn't be uploaded. Your guides helped me a lot to get started with Flux thanks!

    Other
    Other

    Details

    Downloads
    280
    Platform
    CivitAI
    Platform Status
    Available
    Created
    3/18/2025
    Updated
    5/14/2026
    Deleted
    -

    Files

    flanT5XXLTextEncorder_q4KM.zip

    Mirrors

    flanT5XXLTextEncorder_q4KM.zip

    Mirrors