CivArchive
    AnimEasy Flux - AnimEasy Flux v1.1
    NSFW
    Preview 37856139
    Preview 37861971
    Preview 34539752
    Preview 34539755
    Preview 34539563
    Preview 34539557
    Preview 34532330
    Preview 34539567
    Preview 34532308
    Preview 34532269
    Preview 34532272
    Preview 34532296
    Preview 34532285
    Preview 34879411
    Preview 34887186
    Preview 34887497
    Preview 34896488
    /###### Now available in 4bit and even 2bit quantization! ######

    Want to generate anime style images but aren't a prompt wizard? AnimEasy only needs a couple of words to generate decent images (sometimes even one is enough). Capable of generating both semi-realistic and anime style images along with decent text performance. See the show reel for some examples.

    Using the DDIM sampler, a guidance factor of 3.5, and 10 steps works well enough - if you find some images are a bit half-baked, increase the number of steps and let them cook a bit more (more steps rarely hurts the image and even at 50 steps, there's still an improvement). To improve performance, I recommend the flux version of TAESD VAE. All examples were generated in ComfyUI with the fp16 version of this checkpoint and the stock FLUX CLIP models.

    If you don't explicitly use a keyword like "anime", "manga", or "cartoon", you may find some images take on a more... illustration-y look to them. One way to counter this without actually changing the prompt is to increase the number of steps, this tends to sharpen the line-art and flatten the colours a bit without any major stylistic or compositional changes (although it will take proportionally longer to finish cooking).

    N.B. Even using the fp8 checkpoint with fp8 CLIP models still needs a good 13-14GB of VRAM to run. If you want to have everything run native (fp16/bf16) you'll need a good 23-24GB VRAM. FLUX is also a very compute-intensive model, as a point of reference, each fp16 step takes ~1.2s on an RTX3090.

    If you don't have the VRAM to run the fp16 model natively, I'd recommend getting the GGUF loader (if you use ComfyUI), the Q8 model seems a lot closer to the original than the fp8 model, and the loader also works with Q4 and Q2 quantized models. The only advantage fp8 has is that it has native support in a lot more programs whereas GGUF sometimes needs third-party plugins to work. Just be aware that any quantization has a bit of a compute penalty, I saw an ~30% increase in step time on an RTX3090 no matter whether I was using Q8, Q4 or Q2.

    "backyard anime picnic" - steps:50, cfg:3.5, seed:918081932971977

    Description

    bf16, fp8_e5m2, Q8, Q4 & Q2 versions of an anime-optimized FLUX.1 Dev checkpoint model, If you have ComfyUI, you can load a basic workflow from the showreel images.

    FAQ

    Comments (7)

    wzwowzwOct 23, 2024· 4 reactions
    CivitAI

    nf4 version pls

    DamacyNov 10, 2024
    CivitAI

    Interesting to see that Q4_K_M looks pretty similar to Q8.

    And I guess, there will be better results with IQ_2 quants.

    MultitudinousNov 14, 2024
    CivitAI

    I have a question! I need to use GGUF loader as I only have 16GB of VRAM so FP16 is a bit of a stretch. I downloaded Flux_Q6_K model, as it was suggested, but how would I use that together with your checkpoint? I may have missed something entirely aha

    traugdor90Feb 18, 2025

    use the GGUF loader as the model and then import your clip models using GGUF's clip loader. Google it if you're unsure how to use GGUF models.

    accesdJun 16, 2025
    CivitAI

    For some reason, when I use this model with Invoke UI in Linux, my RAM gets filled up to a point my computer freezes. It looks like there's some kind of memory leak with this model or it causes one for some reason?

    frikitinOct 18, 2025
    CivitAI

    ¿What is the difference between Pruned Model nf4 (6.46 GB) and Pruned Model fp32 (3.76 GB)??
    ¿who is the Q4 version ??

    emris93Dec 9, 2025
    CivitAI

    Hello,

    tried all version but nothing is working on swarmUI or Stable diffusion.

    Stable diffusion : TypeError: 'NoneType' object is not iterable

    SwarmUI : Possible reason: ComfyUI execution error: Error while deserializing header: incomplete metadata, file not fully covered.

    Checkpoint
    Flux.1 D

    Details

    Downloads
    1,759
    Platform
    CivitAI
    Platform Status
    Available
    Created
    10/14/2024
    Updated
    5/13/2026
    Deleted
    -

    Files

    animeasyFlux_animeasyFluxV11.gguf

    Mirrors

    animeasyFlux_animeasyFluxV11.gguf

    Mirrors

    animeasyFlux_animeasyFluxV11.gguf

    Mirrors

    Available On (1 platform)

    Same model published on other platforms. May have additional downloads or version variants.