WAN 2.2 / SVI Pro 2 / I2V for 12GB VRAM
v1.2 - Please update to this version.
Seed per Sample - fixed.
Seed arrangement were out of order. (I'm not sure how this slipped through.)
Added labels on seed per sample.
Tested each seed on each clip and changed seed to make sure its working correctly.
Added a preview image after Resize Image so you can check and compare your image to the original.
For 832x1216 Images use 576x832 resolution if you're using the crop option on resize.
Modified version of [SVI Pro 2.0 for Low VRAM (8GB)]
And [Wan2.2 SVI Pro Example KJ]
7 Stage Sample Setup, with each Stage having their own Loras, combined with Sage Attention Cuda for faster speeds.
Can save each stage clip if needed.
Final Output w/ Upscaler + RIFE for smooth 60FPS.
Fast Group Bypasser - for quick access.
### Required Models & LoRAs
GGUF Main Models:
* [DaSiWa-Wan 2.2 I2V] or
* [Smooth Mix Version] or
* [Enhanced NSFW Camera Prompt Adherence]
> Note: Use a suitable quantization (e.g., Q4 or Q5) based on your available VRAM. I highly recommend DaSiWa-Wan high/low Models, as the Lightning Loras are BAKED in, leaving you only with SVI Loras being required.
SVI PRO LoRAs (Wan2.2-I2V-A14B):
* Both Required
Text Encoders:
[WAN UMT5] or
VAE:
The following is for Speed Boosts for nVidia Cards - If its already working then skip this!
Patch Sage Attention Node (sageattn_qk_int8_pv_fp16_cuda) + Model Patch Torch Settings Node (Faster Speed Times):
Prompt executed in 136.56 seconds <- Sage Attention Disable/FP16 Accumulation = Disable/Allow Compile = False
Prompt executed in 104.38 seconds <- Sage Attention Enabled/FP16 Accumulation = Enabled/Allow Compile = False
Prompt executed in 96.26 seconds <-- Sage Attention Enabled/FP16 Accumulation = True/Allow Compile = True
With this setup you can save a massive 40+ seconds just for one Stage!
If Sage Attention is not working/crashing comfyui then do the following or use (CTRL+B to bypass the nodes but I highly recommend getting it working for massive speed boost):
The following is for Comfyui_windows_portable, do not do it this way if you are using a different setup!
Step 1 — Check your PyTorch + CUDA version
Open CMD in your ComfyUI Portable folder (SAME directory as run_nvidia_gpu.bat) and run the following command:
.\python_embeded\python.exe -c "import torch; print(torch.__version__, torch.version.cuda)"
output = 2.9.1+cu130 13.0
check Python embeded version:
.\python_embeded\python.exe -V
output = Python 3.13.9
Which Means:
Python: 3.13 (embeded)
PyTorch: 2.9.1
CUDA: 13.0
Warning! If you are unsure how to proceed with the following steps, then paste your error code into Grok/ChatGPT
for a more detailed analysis.
Pick the wheel that matches your Python + PyTorch + CUDA output from Step 1.
That means the correct SageAttention wheel for your setup would be something like this:
sageattention-2.2.0.post3+cu130torch2.9.0-cp313-cp313-win_amd64.whl
download the correct wheel for your setup from:
It matches Python 3.13 (cp313-cp313), PyTorch 2.9.x, and CUDA 13.0.
The slight difference in patch version (2.9.1 vs 2.9.0) is fine — this wheel works with PyTorch 2.9.x.
Step 2 — Install Wheel (make sure the file is in \ComfyUI_windows_portable, same directory as run_nvidia_gpu.bat)
Open CMD in your ComfyUI Portable folder and run with the correct wheel file (example below):
.\python_embeded\python.exe -m pip install "sageattention-2.2.0.post3+cu130torch2.9.0-cp313-cp313-win_amd64.whl"
Step 3 — How to check if it works:
Open CMD in your ComfyUI Portable folder and run:
.\python_embeded\python.exe -c "import sageattention; print('SageAttention import successful!'); print(dir(sageattention))"
You should see:
SageAttention import successful!
['__builtins__', '__cached__', '__doc__', '__file__', '__loader__', '__name__', '__package__', '__path__', '__spec__', '_fused', '_qattn_sm80', '_qattn_sm89', '_qattn_sm90', 'core', 'quant', 'sageattn', 'sageattn_qk_int8_pv_fp16_cuda', 'sageattn_qk_int8_pv_fp16_triton', 'sageattn_qk_int8_pv_fp8_cuda', 'sageattn_qk_int8_pv_fp8_cuda_sm90', 'sageattn_varlen', 'triton']
Step 4 — confirm if triton attention mode is available:
Open CMD in your ComfyUI Portable folder and run:
.\python_embeded\python.exe -c "import sageattention; print('SageAttention import successful!'); print('Triton mode available:' , hasattr(sageattention, 'sageattn_qk_int8_pv_fp16_triton'))"
You should see:
SageAttention import successful!
Triton mode available: True
if any triton errors run this command:
.\python_embeded\python.exe -m pip install triton
Step 5 - now you should be able to use "sageattn_qk_int8_pv_fp16_cuda" with Patch Sage Attention + Model patch Torch Settings Nodes properly.
Description
updated default settings. If you're a beginner please read the tips at the bottom of notes.