CivArchive
    TXT to VIDEO simple workflow WAN2.2 | GGUF | LoRA | UPSCALE | Fast LoRA - v1.2
    Preview 92316589
    Preview 92316615

    I'm offering you a first version of my workflow adapted to WAN2.2. There's still a lot to improve, so don't hesitate to give me feedback.

    Resources you need:

    馃搨Files :

    For base version
    T2V Model : wan2.2_t2v_high_noise_14B_fp8_scaled.safetensors and wan2.2_t2v_low_noise_14B_fp8_scaled.safetensors
    In models/diffusion_models

    For GGUF version
    T2V Quant Model : wan2.2_t2v_high_noise_14B_QX.gguf and wan2.2_t2v_low_noise_14B_QX.gguf
    In models/unet

    For lightx2V version
    lightx2V LoRA : Wan21_T2V_14B_lightx2v_cfg_step_distill_lora_rank64.safetensors

    CLIP: umt5_xxl_fp8_e4m3fn_scaled.safetensors
    in models/clip

    VAE: wan_2.1_vae.safetensors
    in models/vae

    ANY upscale model:

    in models/upscale_models

    馃摝Custom Nodes :

    Description

    fixed wrong workflow upload (sorry i dont know how it happened)

    FAQ

    Comments (20)

    sonofabeanAug 5, 20251 reaction
    CivitAI

    Workflow is nice and simple as always, but I'm having issues with the LightX GGUF one.

    Firstly MagCache throws an error if enabled. "NoneType" attribute "to" apparently. But whatever, disabling that and enabling TeaCache allowed a nice generation, pretty fast too, faster than my 2.1 WF.

    However subsequent generations with different prompts and seeds threw up the same video as the first (woman holding an umbrella in the rain) but grainy/blotchy. Seems like TeaCache isn't properly clearing itself. I've updated all nodes to no avail. Disabling TeaCache lets it do another generation no issues. I might just leave it off for now, 190 seconds for a video is still an improvement over 2.1 for me.

    Thanks!

    ViciousCalamariAug 13, 2025

    Exact same issue here

    CyberoAug 5, 2025
    CivitAI

    Awesome workflow! Thanks! The only thing I personally felt was missing is the option to manually choose the sampler and maybe a live preview, so it鈥檚 easier to get a sense of what to expect during generation.

    seductivelyai695Aug 6, 20252 reactions
    CivitAI

    do you know why the videos are always washed out? the video is not sharp and the colors seem faded. maybe its just a setting somewhere...

    fox23vang226Aug 11, 20252 reactions

    maybe resolution size, anything below 720p will be bad quality, processing the video through an upscaler may also disturb the video quality. For example, if you have a decent video of a person and throw it through a bad or incompatible upscaler your original video could come out with the persons face blurry and with artifacts due to oversharpening.

    Light6969Aug 13, 20251 reaction

    Make sure you actually load the light loras, it can seem like it's loaded since it shows the name, but you have to select it in the list to make sure it loads your one.

    ApchXiAug 7, 2025
    CivitAI

    Hi! Problem!!

    WanVideoNAG

    mat1 and mat2 shapes cannot be multiplied (385x768 and 4096x5120)

    grahama3927Aug 15, 20252 reactions
    CivitAI

    Using GGUF+lightx2v json: Sampler High gives error "'NoneType' object has no attribute 'to'". 15th Aug. 2025.

    sonofabeanAug 17, 20252 reactions

    I fixed this by turning off Mag Cache in the options. Unfortunately Tea Cache is broken too, it will work for one generation then it won't clear. Works quite well with neither cache, but still waiting for a fix as TeaCache really speeds up the generations.

    ViciousCalamariAug 16, 2025
    CivitAI

    Wildcard integration would be awesome here

    SysDeepAug 19, 2025

    I install comfyui-dynamicprompts then add the node and connect it to normal prompt node, works fine :)

    ViciousCalamariAug 19, 2025

    SysDeep聽Thank you, workflow wizard!

    tiskloteweer588Aug 17, 2025
    CivitAI

    This workflow crashes my Comfyui every time.
    No error in the logs.
    RTX3090, 64GB ram.

    drd1esel273Aug 24, 2025
    CivitAI

    Can you tell me how to reduce the speed of the video? I can't figure out where to change the settings so that the video becomes slower.

    redSamredAug 25, 2025
    CivitAI

    Does making the video longer use more VRAM with a 5080 I can get 4 seconds but 6 crashes. Anyway to make a longer video with the same VRAM?

    UmeAiRT
    Author
    Aug 25, 2025

    You can use blockswap to tranfer some to RAM or use gguf models

    adn7Sep 23, 2025
    CivitAI

    Is for wan2.1 in description ?

    For lightx2V version
    lightx2V LoRA : Wan21_T2V_14B_lightx2v_cfg_step_distill_lora_rank64.safetensors

    adamdeepg991Oct 22, 2025
    CivitAI

    I'm getting this error when I run this workflow, any recommendations please.

    "RuntimeError: Expected all tensors to be on the same device, but got weight is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA__native_layer_norm)"

    lkimDec 2, 2025
    CivitAI

    In v.0.3.76+v1.2,

    gguf works fine.

    gguf+lightx2v seems to have some issues. I'm investigating the cause.

    Thank you.

    Workflows
    Wan Video 14B t2v

    Details

    Downloads
    2,628
    Platform
    CivitAI
    Platform Status
    Available
    Created
    8/4/2025
    Updated
    5/14/2026
    Deleted
    -

    Files

    txtToVIDEOSimpleWorkflowWAN22_v12.zip

    Mirrors