CivArchive

    โœจOne-click Pod available on:โœจ

    ๐ŸŸฃ Runpod ComfyUI 0.19.0 CUDA 12.8 for 5090

    ๐ŸŸฃ Runpod ComfyUI 0.19.0 CUDA 12.4 for 4090

    ๐ŸŸก VastAI ComfyUI 0.19.0 CUDA 13.0 for 5090

    ๐ŸŸก VastAI ComfyUI 0.19.0 CUDA 13.0 for 4090

    Just click on the link, choose a Video Card and the Template will install all you need + all my workflows.

    Wan 2.2 Models are not included, you can install it using Civicomfy or ComfyUI-HuggingFace directly inside ComfyUI


    OneClick-I2V-Story on Runpod Basic Tutorial

    https://limewire.com/d/2TAty#No5eoN7WhU


    โ˜•๏ธ buymeacoffee


    IMPORTANT:

    If you install RES4LYF node it will broke the MoEKSampler, to use it you have to use the KSampler included in that node.


    ๐Ÿ”ฅ 02/28/26 UPDATE ๐Ÿ”ฅ

    โœจ NEW NODES โœจ

    EASY MODEL DOWNLOAD FROM HUGGINGFACE

    cd ComfyUI/custom_nodes
    git clone https://github.com/huchukato/ComfyUI-HuggingFace.git

    RIFE INTERPOLATION WITH TENSORRT

    with Auto Install (CUDA 12/CUDA 13) and Auto Model Download

    cd ComfyUI/custom_nodes
    git clone https://github.com/huchukato/ComfyUI-RIFE-TensorRT-Auto.git

    UPSCALER WITH TENSORRT

    with Auto Install (CUDA 12/CUDA 13) and Auto Model Download

    cd ComfyUI/custom_nodes
    git clone https://github.com/huchukato/ComfyUI-Upscaler-TensorRT-Auto.git

    โœจMY QWEN3-VL NODE IS FINALLY IN THE MANAGERโœจ

    Just search for "QwenVL-Mod" to install it

    or dowload it from GitHub: QwenVL-Mod: Enhanced Vision-Language

    ComfyUI-QwenVL-Mod - Enhanced Vision-Language with WAN 2.2

    Version 2.2.4 (2026/03/13) - ๐ŸŽฌ Critical I2V Timeline Fixes & NSFW Presets Optimization


    ๐ŸŒŸ What is ComfyUI-QwenVL-Mod?

    A powerful enhanced vision-language node for ComfyUI that combines Qwen3-VL models with professional WAN 2.2 video generation workflows. Features multilingual support, visual style detection, and NSFW capabilities for professional AI content creation.

    Think: "Your all-in-one solution for intelligent prompt enhancement and video generation with cutting-edge AI models!"


    ๐ŸŽฌ Key Features

    ๐Ÿš€ WAN 2.2 Video Generation

    • Text-to-Video (T2V): Professional 5-second video generation

    • Image-to-Video (I2V): Advanced image animation with style detection

    • Story Generation: 20-second continuous videos with 4 narrative segments

    • Storyboard Workflows: Seamless storyboard-to-storyboard generation

    • Cinematic Video: Professional cinematography specifications

    ๐ŸŒ Enhanced Capabilities

    • Multilingual Support: Process prompts from any language (Italian, English, etc.)

    • Visual Style Detection: 12+ artistic styles (anime, 3D, pixel art, puppet animation, etc.)

    • Smart Prompt Caching: Performance optimization with Fixed Seed Mode

    • GGUF Backend: Efficient local model inference with quantization support

    • NSFW Support: Comprehensive content generation without restrictions

    ๐Ÿง  Intelligent Features

    • Auto-Prompt Enhancement: Automatically enhance user prompts for optimal generation

    • Professional Cinematography: Built-in specifications for lighting, camera angles, shot types

    • Timeline Structure: Precise 5-second timeline with frame-by-frame descriptions

    • Keep Last Prompt: Generate once, preserve results while changing inputs


    ๐ŸŽฏ What's New in v2.2.4 - CRITICAL I2V TIMELINE FIXES

    ๐Ÿšจ Major I2V Timeline (20s) Fixes

    • โœ… Style Coherence: Fixed AI changing animeโ†’realism mid-sequence

    • โœ… Character Stability: Fixed characters disappearing/appearing incorrectly

    • โœ… Natural Lighting: Fixed AI adding artificial lights not in image

    • โœ… Timeline Structure: Fixed continuous numbering (6,7,8...) instead of 0-5 restart

    • โœ… Format Consistency: Fixed missing parentheses and unwanted labels

    • โœ… Output Format: Each prompt starts directly with timeline markers

    ๐Ÿ”ง NSFW Presets Optimization

    • โœ… Complete Specifications: All 8 NSFW presets now include full NSFW descriptions

    • โœ… Emoji Display: Restored proper emoji rendering (๐Ÿฟ๐ŸŽฅ๐ŸŽฌ๐Ÿ“–)

    • โœ… Clear Instructions: Removed confusing recommendations from presets

    • โœ… User Guide: Token settings guide created for workflow optimization

    ๐Ÿ“‹ Technical Improvements

    • โœ… Timeline Markers: Correct (At X seconds: ...) format for all 4 prompts

    • โœ… Character Continuity: Natural progression without forced artificial presence

    • โœ… Lighting Rules: Logical progression instead of absolute prohibitions

    • โœ… Style Detection: Consistent style application across all timeline segments

    ๐ŸŽฏ Model Recommendations

    • Qwen3-VL-8B: Recommended for I2V Timeline (20s) complex sequences

    • Qwen3-VL-4B: Sufficient for I2V Scene (5s) single prompts

    • Token Settings: 2048+ for 20s timeline, 1024+ for 5s prompts

    ๐ŸŽฏ What's New in v2.2.3

    • CUDA 13 Compatibility: Fixed crashes caused by conflicting unload operations

    • Parameter Cleanup: Removed redundant unload_after_run from all nodes

    • Bug Fixes: Resolved "missing required positional argument" errors

    • Memory Management: Streamlined VRAM cleanup with VRAM Cleanup node

    • Documentation: Updated all README files with new memory features

    • Credits: Added community credits for feedback and testing

    ๐ŸŽฏ What's New in v2.2.2

    ๐Ÿš€ Critical T2V/I2V Workflow Fixes

    • Batch Processing: Fixed critical T2V โ†’ GGUF issue with batch images

    • Frame Detection: Added automatic batch detection and individual frame processing

    • Video Support: Enhanced video frame processing with proper shape handling

    • Debug Enhanced: Comprehensive logging for batch processing troubleshooting

    ๐Ÿ”„ Same Model Reuse Fix

    • Conflict Resolution: Fixed crash when using same model between T2V and I2V nodes

    • Memory Management: Enhanced cleanup with CUDA synchronization and timing

    • Signature Mismatch: Resolved different signature patterns between nodes

    • Aggressive Cleanup: Forced complete VRAM cleanup before model reload

    ๐Ÿ”ง keep_model_loaded Enhancement

    • Missing Parameter: Added keep_model_loaded to PromptEnhancer node

    • Consistent Behavior: Both GGUF and PromptEnhancer now have identical memory management

    • Conditional Cleanup: Proper cleanup based on keep_model_loaded setting

    • User Control: Full control over memory usage vs performance


    ๐Ÿšจ CRITICAL BUG FIXES - v2.2.4

    ๐ŸŽฌ I2V Timeline (20s) - COMPLETELY FIXED

    Before v2.2.4:

    • โŒ Anime style changed to realism mid-sequence

    • โŒ Characters disappeared/appeared randomly

    • โŒ AI added artificial lights not in image

    • โŒ Timeline numbering: 6,7,8... instead of 0-5 restart

    • โŒ Missing parentheses and unwanted "Prompt 1:" labels

    After v2.2.4:

    • โœ… Perfect Style Coherence: Anime stays anime, realism stays realism

    • โœ… Character Stability: Same characters throughout all 4 prompts

    • โœ… Natural Lighting: Only lights visible in image, logical progression

    • โœ… Correct Timeline: Each prompt uses 0-5 seconds format

    • โœ… Clean Output: Proper (At X seconds: ...) format, no labels

    ๐Ÿ”ฅ NSFW Presets - ENHANCED & FIXED

    • โœ… Complete Specifications: All 8 presets with full explicit descriptions

    • โœ… Emoji Display: Proper ๐Ÿฟ๐ŸŽฅ๐ŸŽฌ๐Ÿ“– icons (no more unicode codes)

    • โœ… User-Friendly: Removed confusing technical recommendations

    • โœ… Token Guide: Workflow note for optimal settings

    ๐ŸŽฏ Result: Perfect I2V Timeline generation every time!


    ๐ŸŽฌ WAN 2.2 Story Workflow - Revolutionary AI Storytelling

    ๐Ÿ“– AI Story Generation

    • 4-Segment Videos: Automatic 20-second videos (4 ร— 5-second segments)

    • Narrative Continuity: Perfect story flow between segments

    • NSFW Support: Enhanced adult content generation

    • Timeline-Free: Natural storytelling without time markers

    ๐Ÿ”„ Smart Auto-Split

    • Story Split Node: Intelligent prompt separation technology

    • Auto-Detection: Handles any separator format automatically

    • 4-Output Guarantee: Always produces exactly 4 prompts

    • Debug Mode: Built-in troubleshooting information


    ๐Ÿ“ฆ Installation

    Requirements

    • ComfyUI: v0.13.0+

    • GPU: 8GB+ VRAM (16GB+ recommended)

    • System: Windows/Linux/Mac

    • Python: 3.10+ (or use provided Docker environment)

    Docker/Cloud Ready

    • RunPod: Pre-configured templates available

    • VastAI: Optimized instances ready

    • Local: Docker support included

    Quick Install

    1. Download: ComfyUI-QwenVL-Mod (latest version)

    2. Extract to ComfyUI/custom_nodes/ComfyUI-QwenVL-Mod

    3. Restart ComfyUI

    4. Load included workflows


    ๐ŸŽฎ Usage Examples

    Basic Image-to-Video

    1. Load WAN2.2-I2V-AutoPrompt.json

    2. Upload your image

    3. Select model (HF or GGUF)

    4. Generate enhanced video

    Basic Text-to-Video

    1. Load WAN2.2-T2V-AutoPrompt.json

    2. Input your text prompt

    3. Select model (HF or GGUF)

    4. Generate enhanced video

    Image-to-Video with Style

    1. Load WAN2.2-I2V-AutoPrompt.json

    2. Upload your image

    3. Enable style detection

    4. Generate animated video

    AI Story Generation

    1. Load WAN2.2-I2V-AutoPrompt-Story.json

    2. Input your story idea

    3. Auto-split into 4 segments

    4. Generate 20-second story video


    ๐Ÿ”ง Technical Specifications

    โšก Performance

    • Context: 65,536 tokens (8B models)

    • Memory: Optimized VRAM usage

    • Stability: Crash-free operation

    • Speed: Fast generation times

    ๐ŸŽจ Model Support

    • Qwen3-VL 4B: 7 GGUF variants (2.38GB-4.28GB)

    • Qwen3-VL 8B: 7 GGUF variants (4.8GB-8.71GB)

    • HF Models: Josiefed and official variants

    • Quantization: Q4_K_S, Q5_K_S for VRAM efficiency

    ๐ŸŒ Multilingual Capabilities

    • Input Languages: Any language supported

    • Auto-Translation: Automatic translation to optimized English

    • Style Detection: Works with multilingual prompts

    • Cultural Adaptation: Context-aware prompt enhancement


    ๐ŸŽฏ Included Workflows

    ๐Ÿฟ WAN 2.2 Presets

    • ๐Ÿฟ Wan 2.2 I2V Timeline: Image-to-video with timeline structure

    • ๐ŸŽฅ Wan 2.2 I2V Scene: Image-to-video with professional specs

    • ๐Ÿฟ Wan 2.2 T2V Timeline: Text-to-video with timeline structure

    • ๐ŸŽฅ Wan 2.2 T2V Scene: Text-to-video with professional specs

    • ๐ŸŽฌ Wan 2.2 I2V Timeline: 20-second multi-segment continuity

    • ๐Ÿ“– Wan 2.2 I2V Scene: 20-second single scene with cinematography

    ๐Ÿ”ฅ Advanced Features

    • NSFW Enhancement: Uncensored content generation

    • Professional Lighting: 8 light types + 9 qualities

    • Camera Control: 6 shot types + 5 compositions

    • Color Grading: 4 tone options


    ๐ŸŽจ Visual Style Detection

    Automatically detects and enhances:

    • Photorealistic style - Realistic lighting, natural textures, lifelike details

    • Anime style - Japanese animation aesthetics with vibrant colors

    • Cartoon style - Bold outlines, flat colors, exaggerated expressions

    • 3D animation style - Computer-generated 3D rendering characteristics

    • Fantasy style - Magical elements, ethereal lighting, imaginative atmosphere

    • Artistic portrait style - Painterly qualities, artistic brushwork, stylized composition


    ๐Ÿ”ฅ NSFW Content Support

    Enhanced Generation

    • Explicit Content: Uncensored adult descriptions

    • Detailed Scenes: 8-12 sentences per segment

    • Natural Progression: Smooth story flow

    • Style Adaptation: Automatic visual style matching

    • Quality: Consistent characters & scenes

    Professional Applications

    • Adult Content: Industry-standard generation

    • Artistic Nudity: Classical art styles

    • Educational: Anatomy and artistic study

    • Creative: Artistic expression


    ๐Ÿš€ Why Choose ComfyUI-QwenVL-Mod?

    ๐ŸŽฌ For Content Creators

    • Storytelling: Create compelling narratives

    • Efficiency: One prompt โ†’ complete video

    • Quality: Professional video output

    • Flexibility: Any genre, any style

    ๐Ÿ”ฅ For NSFW Content

    • Explicit: Uncensored generation

    • Detailed: Rich scene descriptions

    • Continuous: Smooth story flow

    • Natural: Realistic progression

    โšก For Power Users

    • Customizable: Easy to modify

    • Extendable: Add more segments

    • Integrable: Works with existing setups

    • Optimized: Maximum performance


    ๐ŸŒŸ What Makes This Special?

    • First: Complete AI story system with vision enhancement

    • Smart: Intelligent prompt splitting and enhancement

    • Complete: End-to-end solution from text to video

    • Optimized: Performance-tuned for professional use

    • Ready: Works out-of-the-box with included workflows


    ๐ŸŽฌ Create Amazing AI Videos Today!

    Transform your ideas into stunning videos with the power of Qwen3-VL vision enhancement and WAN 2.2 video generation.

    Perfect for creators, artists, and professionals looking for the ultimate AI video enhancement tool! ๐ŸŒŸ


    Built with โค๏ธ for the ComfyUI community


    The new version is out today, with a brand new anime style โœจ

    WORKFLOWS TESTED ON:

    • ComfyUI 0.17.1

    • Python 3.12.12

    • Pytorch 2.9.1 + CUDA 13.0


    ๐ŸŸฃ Credits

    This workflows are intended to be used with the models by taek75799 as they follow the structure of Dynamic Prompts you can find under these models:

    WAN 2.2 Enhanced NSFW | SVI | camera prompt adherence (Lightning Edition) I2V and T2V fp8 GGUF

    ๐ŸŸฃ Other tested models:

    Thanks to all the users who are commenting and helping me improve the workflows โค๏ธ


    ๐Ÿ”– T2V-I2V AUTOPROMPT

    ๐Ÿ›‘ Experimental WF ๐Ÿ›‘

    Start with a T2V prompt and extend the generated video with I2V

    This workflow requires both T2V and I2V Wan 2.2 Models

    ๐Ÿ”– SVI I2V AUTOPROMPT

    Thanks to taek75799 for his models โค๏ธ

    SVI LORAS:

    LIGHTX2V LoRaS are not included in the model

    ๐Ÿ”– FULL I2V AUTOPROMPT

    Complete workflow that includes:

    • Long Video Generation [from 5 to 20 seconds]

    • Auto Prompting [Qwen3-VL]

    • Upscale [2xLexicaRRDBNet and Tensorrt]

    • Frame Interpolation [30fps and 60fps for img2vid | 24fps and 50fps for MMAudio]

    • MMAudio [NSFW Unlocked]

    ๐Ÿ“ AUTO PROMPT

    • Prompt Description Box [Multilanguage]: Just write your idea and the LLM will do the rest, formatting the prompt in the dynamic prompt format used in the Wan 2.2 models

    • Final Prompt Preview: Shows up the final prompt

    ๐Ÿ”ถ QWEN3-VL NODE FOR GGUF MODELS ๐Ÿ”ถ

    ๐Ÿ”– To use the Qwen3-VL GGUF Quantized models you have to install llama-cpp-python

    If you are not comfortable in manual install the llama, just go with the normal version inside Full-I2V-LongVideo

    1. ๐Ÿ›‘ STOP COMFYUI

    1. ๐Ÿ“‚ Activate the ComfyUI Virtual Enviroment

    In your ComfyUI root installation folder type:

    on Windows:

    Command Prompt:

    \venv\Scripts\activate.bat

    or Power Shell:

    \venv\Scripts\Activate.ps1

    on Linux:

    . /venv/bin/activate

    If you use ComfyUI Desktop:

    Click on Console and then on Terminal

    1. โฌ‡๏ธ Install llama-cpp-python

    pip install --upgrade --force-reinstall --no-cache-dir "llama-cpp-python @ git+https://github.com/JamePeng/llama-cpp-python.git"

    1. ๐Ÿ”„ Restart ComfyUI and enjoy

    ๐Ÿ“ SWITCH CLIP

    By default the WF uses the GGUF node to load quantized Clip, if you wanna switch to the NSFW Clip model, you have to bypass the GGUF node and connect the other Clip loader to the "Set_CLIP"

    ACCELERATION:

    Triton is disabled by default, you can enable it by opening the first Subgraph

    Inside the workflow you will find all the links to download the models you will need

    That's all, hope you enjoy ^^

    Description

    Removed the Tensorrt Upscaler and replaced with 2xLexicaRRDBNet

    FAQ

    Comments (41)

    xuf129750620Jan 26, 2026
    CivitAI

    AILab_QwenVL_GGUF_Advanced

    [QwenVL] llama_cpp is not available. Install the GGUF vision dependency first. See docs/GGUF_MANUAL_INSTALL.md

    huchukato
    Author
    Jan 26, 2026ยท 1 reaction

    I wrote the guide to install it here and also in the WF...

    meritrash6350Jan 26, 2026

    @huchukatoย Yeah,the guide isn't geared toward everyone. You might have come up with something cool here, but the explanations on how to get llama going are extremely lacking.

    huchukato
    Author
    Jan 27, 2026

    @meritrash6350ย I wrote all the steps ypu have to do to install llama, the only thing that lack is how to activate the virtual enviroment on Windows coz I don't use Windows, I'm on Mac and Linux

    meritrash6350Jan 27, 2026

    @huchukatoย So, you get my point then?

    huchukato
    Author
    Jan 28, 2026

    @meritrash6350ย I got it but you don't have to install the GGUF version of the WF in every cases, go with the normal one where you don't have to install llama python, I cannot tell you how to get in a venv in Windows not having Windows on my PC :\

    evantopsmithJan 28, 2026ยท 1 reaction

    @meritrash6350 In your ComfyUI root installation folder activate enviornment using either script:

    Command Prompt: \venv\Scripts\activate.bat

    PowerShell: \venv\Scripts\Activate.ps1

    meritrash6350Jan 28, 2026

    @evantopsmithย Sorry, wasn't ignoring you, I got too busy. I'll take a look at it, and I appreciate the extra effort. I personally find the effort of dealing with venv to be the thing I hate most about Python. That and the fact it insists on caching on the C: even though that's not where I told it to install.

    huchukato
    Author
    Jan 28, 2026

    @evantopsmithย Thanks a lot

    lanceshockerJan 26, 2026
    CivitAI

    I am so confused on what you mean to install. The guide isn't clear, whad you you even mean by start a ComfyUI virtual environment???

    huchukato
    Author
    Jan 27, 2026

    When you install ComfyUI the installer create a virtual enviroment to run it, usually thers is a "venv" folder inside ComfyUI. This allows Comfy to run on a specific Python version and to install all the dependecies that it needs not for all you system but just for Comfy, to avoid comflicts in you system

    qwe246Jan 27, 2026
    CivitAI

    Hello, in my QWENVL node, the generated prompt are completely unrelated to the prompt I entered. The generated prompt seem to only describe the content of the image itself. Why is this happening? Full i2v longvideo gguf

    Adam_LolFeb 2, 2026

    I am having the same issue

    evantopsmithJan 28, 2026
    CivitAI

    I am about to lose my fucking mind with this Qwen3 VL GGUF auto prompt bullshit. git cloned QwenVL-Mod (no QwenVL node from manager) start ComfyUI install dependencies, quit. download llama_cpp_python-0.3.23+cu128.basic-cp312-win_amd64.whl (Python 3.12.10, CUDA 2.8.0cu128) cmd in python_embedded (equivalent to being in active virtual environment using comfyui-easy-install) pip install --upgrade --force-reinstall llama_cpp_python-0.3.23+cu128.basic-cp312-win_amd64.whl (whl file is in same folder cmd was started in) restart comfyui and try to use Huihui-Qwen3-VL-4B-Instruct-abliterated-Q8_0.gguf.....

    AILab_QwenVL_GGUF_Advanced

    [QwenVL] Missing Qwen VL chat handler in llama_cpp. Install the correct fork/wheel. See docs/GGUF_MANUAL_INSTALL.md

    huchukato
    Author
    Jan 28, 2026

    :\

    huchukato
    Author
    Jan 28, 2026

    @gregariousbuttons257ย This is the original node but you have to manual install the abliterated Qwen models, the only difference with mine is that I added the uncensored models

    evantopsmithJan 28, 2026

    @huchukatoย but I followed all of that to the tee. which part of what I described am I doing wrong? do I have the wrong whl for my python cuda versions?

    yamyproJan 28, 2026
    CivitAI

    Was so hyped for this workflow but I have been trying to get it to work for the last 3 hours. I have yet to make even one generation! something with the Ksampers? Which I don't understand because they work normally in other workflows but for some reason this one just refuses to work for me!

    this is what i keep getting:
    Failed to validate prompt for output 1431:

    * KSamplerAdvanced 1252:1269:

    - Return type mismatch between linked nodes: scheduler, received_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent']) mismatch input_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent', 'beta57'])

    * KSamplerAdvanced 1252:1270:

    - Return type mismatch between linked nodes: scheduler, received_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent']) mismatch input_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent', 'beta57'])

    Output will be ignored

    Failed to validate prompt for output 1252:1259:

    Output will be ignored

    Failed to validate prompt for output 1327:

    Output will be ignored

    [QwenVL] Flash-Attn auto mode: dependency not ready, using SDPA

    Interrupting prompt 1cc5223a-200f-42bd-908f-13af58061592

    Interrupting prompt 1cc5223a-200f-42bd-908f-13af58061592

    Interrupting prompt 1cc5223a-200f-42bd-908f-13af58061592

    Interrupting prompt 1cc5223a-200f-42bd-908f-13af58061592

    Interrupting prompt 1cc5223a-200f-42bd-908f-13af58061592

    Processing interrupted

    Prompt executed in 524.73 seconds

    huchukato
    Author
    Jan 28, 2026

    Seems there is a problem with the setnode with the scheduler, which version are you using?

    jarigoni949Jan 28, 2026

    @huchukato, i have the same problem, what do you mean by Version? Comfy?ย 

    huchukato
    Author
    Jan 28, 2026

    @jarigoni949ย Comfy and also the WF version, coz for me all versions works on Comfy 0.10.0, Python 3.12.12 and CU13.0 with Pytorch 2.9.1

    jarigoni949Jan 29, 2026

    @huchukatoย Hey :)!
    I'am on: Comfy 0.11.0 / Python 3.12.10 / pytorch version: 2.9.1+cu130 /

    huchukato
    Author
    Jan 29, 2026

    @jarigoni949ย I'm testing the WF on 0.11.0 now and works :\ You can try to do a thing: disconnect the scheduler and sampler from the subgraph and manually set them inside it, maybe it's a problem of the Selectors node

    SolHelJan 29, 2026ยท 1 reaction

    I agree. If it's not about sampler issue, it's about QWEN abridged. The model doesn't download automatically, and when you do manually, the node won't detect it and download other models.

    yamyproJan 31, 2026

    im using comfy11.1, Python 3.12.11, 2.10.0+cu130 and CUDA13.0

    huchukato
    Author
    Jan 31, 2026

    @yamyproย The WF is tested on Python 3.12.12, update your py version maybe

    yamyproFeb 1, 2026

    @huchukatoย i know this is a dumb question but how do i upgrade it?

    jetpisces619Jan 30, 2026
    CivitAI

    I still have the error: The checkpoint you are trying to load has model type qwen3_vl but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.

    huchukato
    Author
    Jan 30, 2026

    You updated my Qwen node? Now it supports both transformers>5.0 and <5.0, I fixed the deprecated syntax

    redlucario1735Jan 31, 2026
    CivitAI

    Ive got it working with no major issues, the only problem that i have is that the generated prompts are very censored, it doesnt describe any nsfw behaviors and at most it implies vaguely that some "intimacy" might be going on. Any idea on how to make it as detailed and nsfw as possible?

    huchukato
    Author
    Jan 31, 2026

    If you are using my modified node and one of the Qwen Abliterated models it should work, for me it works

    dukefan6842872Jan 31, 2026
    CivitAI

    [QwenVL] Failed to apply SageAttention patch: module 'transformers.models.qwen2.modeling_qwen2' has no attribute 'F'

    [QwenVL] SageAttention patch failed, continuing with SDPA

    Is this a SageAttention // Transformers version issue?

    huchukato
    Author
    Jan 31, 2026

    you have sageattention==2.2.0? If it fail to load it it uses SDPA by default

    dukefan6842872Feb 1, 2026

    Name: sageattention

    Version: 2.2.0+cu130torch2.9.0andhigher.post4

    I do yeah, and SDPA still only takes a moment to run QwenVL, but it's still bothering me.

    dukefan6842872Feb 1, 2026ยท 1 reaction

    I've got things tuned so that this workflow is working really well for me. My main issue now is that UpscalerTensorRT at 2x upscale on a 10 second clip is using an insane amount of vram and spilling into shared memory causing it to take upwards of 30min to run, when the rest of the generation only takes ~10 minutes. (RTX 4090) Trying to figure out if it's possible to optimize UpscalerTensorRT's memory usage or if some models loaded that don't need to be when the workflow gets there.
    Edit: I moved the post-processing (UpscalerTensorRT and Interpolation) into a separate workflow and it finishes in around a minute on my video output, so there's something that I don't have the understanding to fix regarding memory management to run your entire workflow efficiently.

    dukefan6842872Feb 1, 2026

    Oh, and thanks for your responses and thanks very much for sharing the workflow!!!

    huchukato
    Author
    Feb 1, 2026

    @dukefan6842872ย no problem, btw now I'm getting the same error LOL, I will take a look tomorrow. Regarding Tensorrt, try the normal Upscale and check if it takes less time. PS. I just update again the node, now have more aderence in NSFW prompting, in minutes I also release the SVI version of the WF

    huchukato
    Author
    Feb 1, 2026

    @dukefan6842872ย I think I fixed the problem, update the node and let me know thank u

    Workflows
    Wan Video 2.2 I2V-A14B

    Details

    Downloads
    119
    Platform
    CivitAI
    Platform Status
    Deleted
    Created
    1/26/2026
    Updated
    4/27/2026
    Deleted
    1/31/2026

    Files

    WAN22NSFWI2VT2VWorkflowsAutoPrompt_fullI2VLVGGUF.zip