CivArchive
    Wan 2.2 - SVI Pro 2.0 - I2V for 12GB VRAM (Different Loras Per Stage)(Optimized for Speed) - v1.2
    NSFW
    Preview 125650554

    WAN 2.2 / SVI Pro 2 / I2V for 12GB VRAM

    v1.2 - Please update to this version.

    Seed per Sample - fixed.

    1. Seed arrangement were out of order. (I'm not sure how this slipped through.)

    2. Added labels on seed per sample.

    3. Tested each seed on each clip and changed seed to make sure its working correctly.

    Added a preview image after Resize Image so you can check and compare your image to the original.

    For 832x1216 Images use 576x832 resolution if you're using the crop option on resize.

    Modified version of [SVI Pro 2.0 for Low VRAM (8GB)]

    And [Wan2.2 SVI Pro Example KJ]

    • 7 Stage Sample Setup, with each Stage having their own Loras, combined with Sage Attention Cuda for faster speeds.

    • Can save each stage clip if needed.

    • Final Output w/ Upscaler + RIFE for smooth 60FPS.

    • Fast Group Bypasser - for quick access.

    ### Required Models & LoRAs

    GGUF Main Models:

    * [DaSiWa-Wan 2.2 I2V] or

    * [Smooth Mix Version] or

    * [Enhanced NSFW Camera Prompt Adherence]

    > Note: Use a suitable quantization (e.g., Q4 or Q5) based on your available VRAM. I highly recommend DaSiWa-Wan high/low Models, as the Lightning Loras are BAKED in, leaving you only with SVI Loras being required.

    SVI PRO LoRAs (Wan2.2-I2V-A14B):

    * Both Required

    [SVI PRO - HIGH (Rank 128)]

    [SVI PRO - LOW (Rank 128)]

    Text Encoders:

    [WAN UMT5] or

    [NSFW WAN UMT5]

    VAE:

    [Wan 2.1 VAE]

    The following is for Speed Boosts for nVidia Cards - If its already working then skip this!

    Patch Sage Attention Node (sageattn_qk_int8_pv_fp16_cuda) + Model Patch Torch Settings Node (Faster Speed Times):

    Prompt executed in 136.56 seconds <- Sage Attention Disable/FP16 Accumulation = Disable/Allow Compile = False

    Prompt executed in 104.38 seconds <- Sage Attention Enabled/FP16 Accumulation = Enabled/Allow Compile = False

    Prompt executed in 96.26 seconds <-- Sage Attention Enabled/FP16 Accumulation = True/Allow Compile = True

    With this setup you can save a massive 40+ seconds just for one Stage!

    If Sage Attention is not working/crashing comfyui then do the following or use (CTRL+B to bypass the nodes but I highly recommend getting it working for massive speed boost):

    • The following is for Comfyui_windows_portable, do not do it this way if you are using a different setup!

      • Step 1 — Check your PyTorch + CUDA version

    Open CMD in your ComfyUI Portable folder (SAME directory as run_nvidia_gpu.bat) and run the following command:

    .\python_embeded\python.exe -c "import torch; print(torch.__version__, torch.version.cuda)"

    output = 2.9.1+cu130 13.0

    check Python embeded version:

    .\python_embeded\python.exe -V

    output = Python 3.13.9

    Which Means:

    Python: 3.13 (embeded)

    PyTorch: 2.9.1

    CUDA: 13.0

    Warning! If you are unsure how to proceed with the following steps, then paste your error code into Grok/ChatGPT

    for a more detailed analysis.

    Pick the wheel that matches your Python + PyTorch + CUDA output from Step 1.

    That means the correct SageAttention wheel for your setup would be something like this:

    sageattention-2.2.0.post3+cu130torch2.9.0-cp313-cp313-win_amd64.whl

    download the correct wheel for your setup from:

    [List of Wheels]

    It matches Python 3.13 (cp313-cp313), PyTorch 2.9.x, and CUDA 13.0.

    The slight difference in patch version (2.9.1 vs 2.9.0) is fine — this wheel works with PyTorch 2.9.x.

    • Step 2 — Install Wheel (make sure the file is in \ComfyUI_windows_portable, same directory as run_nvidia_gpu.bat)

    Open CMD in your ComfyUI Portable folder and run with the correct wheel file (example below):

    .\python_embeded\python.exe -m pip install "sageattention-2.2.0.post3+cu130torch2.9.0-cp313-cp313-win_amd64.whl"

    • Step 3 — How to check if it works:

    Open CMD in your ComfyUI Portable folder and run:

    .\python_embeded\python.exe -c "import sageattention; print('SageAttention import successful!'); print(dir(sageattention))"

    You should see:

    SageAttention import successful!

    ['__builtins__', '__cached__', '__doc__', '__file__', '__loader__', '__name__', '__package__', '__path__', '__spec__', '_fused', '_qattn_sm80', '_qattn_sm89', '_qattn_sm90', 'core', 'quant', 'sageattn', 'sageattn_qk_int8_pv_fp16_cuda', 'sageattn_qk_int8_pv_fp16_triton', 'sageattn_qk_int8_pv_fp8_cuda', 'sageattn_qk_int8_pv_fp8_cuda_sm90', 'sageattn_varlen', 'triton']

    • Step 4 — confirm if triton attention mode is available:

    Open CMD in your ComfyUI Portable folder and run:

    .\python_embeded\python.exe -c "import sageattention; print('SageAttention import successful!'); print('Triton mode available:' , hasattr(sageattention, 'sageattn_qk_int8_pv_fp16_triton'))"

    You should see:

    SageAttention import successful!

    Triton mode available: True

    if any triton errors run this command:

    .\python_embeded\python.exe -m pip install triton

    Step 5 - now you should be able to use "sageattn_qk_int8_pv_fp16_cuda" with Patch Sage Attention + Model patch Torch Settings Nodes properly.

    Description

    Seed per Sample - fixed.

    1. Seed arrangement were out of order. (I'm not sure how this slipped through.)

    2. Added labels on seed per sample.

    3. Tested each seed on each clip and changed seed to make sure its working correctly.

    Added a preview image after Resize Image so you can check and compare your image to the original.

    For 832x1216 Images use 576x832 resolution if your using the crop option on resize.

    FAQ

    Workflows
    Wan Video 2.2 I2V-A14B

    Details

    Downloads
    571
    Platform
    CivitAI
    Platform Status
    Available
    Created
    3/28/2026
    Updated
    4/26/2026
    Deleted
    -

    Files

    wan22SVIPro20I2VFor12GBVRAM_v12.zip

    Mirrors