π Ultimate AI Video Production Workflow (ComfyUI)
This is a complete end-to-end AI video pipeline that combines:
π§© Qwen-Edit 2509 β Image Editing
π Wan Animate 2.2 β Character Animation
πΊ SeedVR2 Upscaler β Final 4K Enhancement
Everything is packed into a clean, modular, subgraphed workflow that removes the headache of giant node spaghetti.
This workflow allows you to:
Edit any reference image
Animate the edited character using a real motion video
Upscale and clean the final output
Produce 4K smooth AI videos
Run everything inside ComfyUI with optimized VRAM usage
Perfect for:
AI character reels
Cosplay transformations
Cinematic AI edits
TikTok / IG Reels
YouTube shorts
Film-style previews
ποΈ Features
β¨ 1. Qwen-Edit 2509 FP8 (Image Editing Engine)
Outfit changes
Face cleanup / enhancement
Makeup edits
Background removal
Style injection
Very VRAM-friendly
π¬ 2. Wan Animate 2.2 FP8 (Character Animation)
Robust identity preservation
Smooth motion from reference video
Pose & face control
Windowed batching for long videos
Subgraphed for easy navigation
π 3. SeedVR2 Upscaler (Video Restore + 4K)
2Γ / 4Γ upscaling
Face enhance option
Removes noise and artifacts
Makes AI videos look real and cinematic
π§ Bonus Modules
VRAM Cleaner
Resolution Selector
Reference Routing System
Organized Subgraphs
Cleaner wiring = easier editing
π§° How to Use This Workflow
1οΈβ£ Load your reference image
Use Qwen-Edit to perform any image cleanup or enhancement before animation.
2οΈβ£ Load your motion video
Wan Animate will replicate the motion on your edited character.
3οΈβ£ Adjust resolution & pre-process settings
Resolution presets included (320 β 1280).
4οΈβ£ Run Wan Animate 2.2
Generates stable character animation frames.
5οΈβ£ Pass output into SeedVR2
Upscale to 2K / 4K depending on hardware.
6οΈβ£ Save final video
Both preview & HQ versions included.
πΌοΈ Qwen Image Edit FP8 (Diffusion Model, Text Encoder, and VAE)
These are hosted on the Comfy-Org Hugging Face page.
Diffusion Model (qwen_image_edit_fp8_e4m3fn.safetensors):
https://huggingface.co/Comfy-Org/Qwen-Image-Edit_ComfyUI/blob/main/split_files/diffusion_models/qwen_image_edit_fp8_e4m3fn.safetensorsText Encoder (qwen_2.5_vl_7b_fp8_scaled.safetensors):
https://huggingface.co/Comfy-Org/Qwen-Image-Edit_ComfyUI/blob/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensorsVAE (qwen_image_vae.safetensors):
https://huggingface.co/Comfy-Org/Qwen-Image-Edit_ComfyUI/blob/main/split_files/vae/qwen_image_vae.safetensors
π Wan 2.2 Animate 14B FP8 (Diffusion Model, Text Encoder, and VAE)
The components are spread across related community repositories.
Diffusion Model (Wan2_2-Animate-14B_fp8_e4m3fn_scaled_KJ.safetensors):
https://huggingface.co/Kijai/WanVideo_comfy_fp8_scaled/blob/main/Wan22Animate/Wan2_2-Animate-14B_fp8_e4m3fn_scaled_KJ.safetensorsText Encoder (umt5_xxl_fp8_e4m3fn_scaled.safetensors):
https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/blob/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensorsVAE (wan2.2_vae.safetensors):
https://huggingface.co/Comfy-Org/Wan_2.2_ComfyUI_Repackaged/blob/main/split_files/vae/wan2.2_vae.safetensorsLightx2v : https://huggingface.co/Kijai/WanVideo_comfy/tree/main/Lightx2v
πΎ SeedVR2 Diffusion Model (FP8)
Diffusion Model (seedvr2_ema_3b_fp8_e4m3fn.safetensors):
https://huggingface.co/numz/SeedVR2_comfyUI/blob/main/seedvr2_ema_3b_fp8_e4m3fn.safetensors
π§± Why This Workflow is Easy to Use
I wrapped the heavy sections inside subgraphs, so you donβt have to scroll through 200 tangled nodes.
Subgraphs included:
Wan Animate Engine
Qwen-Edit Module
SeedVR2 Upscaler
VRAM Cleaner
Resolution Manager
Makes editing, debugging, and expanding the workflow extremely clean and fast.
β€οΈ Credits
Kijai for FP8 Wan Animate
Comfy-Org for Qwen-Edit
ByteDance for SeedVR2
Community testers & contributors
Description
FAQ
Comments (7)
Seems Great Workflow, it would be nice , if you can make wf for Qwen-Edit 2509 + Wan 2.2 I2V (Long Length) Ultimate AI Video Workflow.
Thank you
Soon. What exactly are you looking for when it comes to long video generation? WAN I2V has several methods for doing long sequences
Great! Thank you
Am I wrong, or is there no upscale here?
No, I didnβt upscale the videos β they are the raw Animate outputs. You can try upscaling them, and Iβll be posting another set of generations with upscale soon.
Quick Q - where can I go to get the lightx2v lora you have in your WF? wasn't quite sure which one it was when looking at the huggingface page. thanks!!
