CivArchive
    Preview 121143811

    ✨One-click Pod available on:✨

    🟣 Runpod ComfyUI 0.19.0 CUDA 12.8 for 5090

    🟣 Runpod ComfyUI 0.19.0 CUDA 12.4 for 4090

    🟡 VastAI ComfyUI 0.19.0 CUDA 13.0 for 5090

    🟡 VastAI ComfyUI 0.19.0 CUDA 13.0 for 4090

    Just click on the link, choose a Video Card and the Template will install all you need + all my workflows.

    Wan 2.2 Models are not included, you can install it using Civicomfy or ComfyUI-HuggingFace directly inside ComfyUI


    OneClick-I2V-Story on Runpod Basic Tutorial

    https://limewire.com/d/2TAty#No5eoN7WhU


    ☕️ buymeacoffee


    IMPORTANT:

    If you install RES4LYF node it will broke the MoEKSampler, to use it you have to use the KSampler included in that node.


    🔥 02/28/26 UPDATE 🔥

    ✨ NEW NODES ✨

    EASY MODEL DOWNLOAD FROM HUGGINGFACE

    cd ComfyUI/custom_nodes
    git clone https://github.com/huchukato/ComfyUI-HuggingFace.git

    RIFE INTERPOLATION WITH TENSORRT

    with Auto Install (CUDA 12/CUDA 13) and Auto Model Download

    cd ComfyUI/custom_nodes
    git clone https://github.com/huchukato/ComfyUI-RIFE-TensorRT-Auto.git

    UPSCALER WITH TENSORRT

    with Auto Install (CUDA 12/CUDA 13) and Auto Model Download

    cd ComfyUI/custom_nodes
    git clone https://github.com/huchukato/ComfyUI-Upscaler-TensorRT-Auto.git

    ✨MY QWEN3-VL NODE IS FINALLY IN THE MANAGER✨

    Just search for "QwenVL-Mod" to install it

    or dowload it from GitHub: QwenVL-Mod: Enhanced Vision-Language

    ComfyUI-QwenVL-Mod - Enhanced Vision-Language with WAN 2.2

    Version 2.2.4 (2026/03/13) - 🎬 Critical I2V Timeline Fixes & NSFW Presets Optimization


    🌟 What is ComfyUI-QwenVL-Mod?

    A powerful enhanced vision-language node for ComfyUI that combines Qwen3-VL models with professional WAN 2.2 video generation workflows. Features multilingual support, visual style detection, and NSFW capabilities for professional AI content creation.

    Think: "Your all-in-one solution for intelligent prompt enhancement and video generation with cutting-edge AI models!"


    🎬 Key Features

    🚀 WAN 2.2 Video Generation

    • Text-to-Video (T2V): Professional 5-second video generation

    • Image-to-Video (I2V): Advanced image animation with style detection

    • Story Generation: 20-second continuous videos with 4 narrative segments

    • Storyboard Workflows: Seamless storyboard-to-storyboard generation

    • Cinematic Video: Professional cinematography specifications

    🌐 Enhanced Capabilities

    • Multilingual Support: Process prompts from any language (Italian, English, etc.)

    • Visual Style Detection: 12+ artistic styles (anime, 3D, pixel art, puppet animation, etc.)

    • Smart Prompt Caching: Performance optimization with Fixed Seed Mode

    • GGUF Backend: Efficient local model inference with quantization support

    • NSFW Support: Comprehensive content generation without restrictions

    🧠 Intelligent Features

    • Auto-Prompt Enhancement: Automatically enhance user prompts for optimal generation

    • Professional Cinematography: Built-in specifications for lighting, camera angles, shot types

    • Timeline Structure: Precise 5-second timeline with frame-by-frame descriptions

    • Keep Last Prompt: Generate once, preserve results while changing inputs


    🎯 What's New in v2.2.4 - CRITICAL I2V TIMELINE FIXES

    🚨 Major I2V Timeline (20s) Fixes

    • ✅ Style Coherence: Fixed AI changing anime→realism mid-sequence

    • ✅ Character Stability: Fixed characters disappearing/appearing incorrectly

    • ✅ Natural Lighting: Fixed AI adding artificial lights not in image

    • ✅ Timeline Structure: Fixed continuous numbering (6,7,8...) instead of 0-5 restart

    • ✅ Format Consistency: Fixed missing parentheses and unwanted labels

    • ✅ Output Format: Each prompt starts directly with timeline markers

    🔧 NSFW Presets Optimization

    • ✅ Complete Specifications: All 8 NSFW presets now include full NSFW descriptions

    • ✅ Emoji Display: Restored proper emoji rendering (🍿🎥🎬📖)

    • ✅ Clear Instructions: Removed confusing recommendations from presets

    • ✅ User Guide: Token settings guide created for workflow optimization

    📋 Technical Improvements

    • ✅ Timeline Markers: Correct (At X seconds: ...) format for all 4 prompts

    • ✅ Character Continuity: Natural progression without forced artificial presence

    • ✅ Lighting Rules: Logical progression instead of absolute prohibitions

    • ✅ Style Detection: Consistent style application across all timeline segments

    🎯 Model Recommendations

    • Qwen3-VL-8B: Recommended for I2V Timeline (20s) complex sequences

    • Qwen3-VL-4B: Sufficient for I2V Scene (5s) single prompts

    • Token Settings: 2048+ for 20s timeline, 1024+ for 5s prompts

    🎯 What's New in v2.2.3

    • CUDA 13 Compatibility: Fixed crashes caused by conflicting unload operations

    • Parameter Cleanup: Removed redundant unload_after_run from all nodes

    • Bug Fixes: Resolved "missing required positional argument" errors

    • Memory Management: Streamlined VRAM cleanup with VRAM Cleanup node

    • Documentation: Updated all README files with new memory features

    • Credits: Added community credits for feedback and testing

    🎯 What's New in v2.2.2

    🚀 Critical T2V/I2V Workflow Fixes

    • Batch Processing: Fixed critical T2V → GGUF issue with batch images

    • Frame Detection: Added automatic batch detection and individual frame processing

    • Video Support: Enhanced video frame processing with proper shape handling

    • Debug Enhanced: Comprehensive logging for batch processing troubleshooting

    🔄 Same Model Reuse Fix

    • Conflict Resolution: Fixed crash when using same model between T2V and I2V nodes

    • Memory Management: Enhanced cleanup with CUDA synchronization and timing

    • Signature Mismatch: Resolved different signature patterns between nodes

    • Aggressive Cleanup: Forced complete VRAM cleanup before model reload

    🔧 keep_model_loaded Enhancement

    • Missing Parameter: Added keep_model_loaded to PromptEnhancer node

    • Consistent Behavior: Both GGUF and PromptEnhancer now have identical memory management

    • Conditional Cleanup: Proper cleanup based on keep_model_loaded setting

    • User Control: Full control over memory usage vs performance


    🚨 CRITICAL BUG FIXES - v2.2.4

    🎬 I2V Timeline (20s) - COMPLETELY FIXED

    Before v2.2.4:

    • ❌ Anime style changed to realism mid-sequence

    • ❌ Characters disappeared/appeared randomly

    • ❌ AI added artificial lights not in image

    • ❌ Timeline numbering: 6,7,8... instead of 0-5 restart

    • ❌ Missing parentheses and unwanted "Prompt 1:" labels

    After v2.2.4:

    • Perfect Style Coherence: Anime stays anime, realism stays realism

    • Character Stability: Same characters throughout all 4 prompts

    • Natural Lighting: Only lights visible in image, logical progression

    • Correct Timeline: Each prompt uses 0-5 seconds format

    • Clean Output: Proper (At X seconds: ...) format, no labels

    🔥 NSFW Presets - ENHANCED & FIXED

    • Complete Specifications: All 8 presets with full explicit descriptions

    • Emoji Display: Proper 🍿🎥🎬📖 icons (no more unicode codes)

    • User-Friendly: Removed confusing technical recommendations

    • Token Guide: Workflow note for optimal settings

    🎯 Result: Perfect I2V Timeline generation every time!


    🎬 WAN 2.2 Story Workflow - Revolutionary AI Storytelling

    📖 AI Story Generation

    • 4-Segment Videos: Automatic 20-second videos (4 × 5-second segments)

    • Narrative Continuity: Perfect story flow between segments

    • NSFW Support: Enhanced adult content generation

    • Timeline-Free: Natural storytelling without time markers

    🔄 Smart Auto-Split

    • Story Split Node: Intelligent prompt separation technology

    • Auto-Detection: Handles any separator format automatically

    • 4-Output Guarantee: Always produces exactly 4 prompts

    • Debug Mode: Built-in troubleshooting information


    📦 Installation

    Requirements

    • ComfyUI: v0.13.0+

    • GPU: 8GB+ VRAM (16GB+ recommended)

    • System: Windows/Linux/Mac

    • Python: 3.10+ (or use provided Docker environment)

    Docker/Cloud Ready

    • RunPod: Pre-configured templates available

    • VastAI: Optimized instances ready

    • Local: Docker support included

    Quick Install

    1. Download: ComfyUI-QwenVL-Mod (latest version)

    2. Extract to ComfyUI/custom_nodes/ComfyUI-QwenVL-Mod

    3. Restart ComfyUI

    4. Load included workflows


    🎮 Usage Examples

    Basic Image-to-Video

    1. Load WAN2.2-I2V-AutoPrompt.json

    2. Upload your image

    3. Select model (HF or GGUF)

    4. Generate enhanced video

    Basic Text-to-Video

    1. Load WAN2.2-T2V-AutoPrompt.json

    2. Input your text prompt

    3. Select model (HF or GGUF)

    4. Generate enhanced video

    Image-to-Video with Style

    1. Load WAN2.2-I2V-AutoPrompt.json

    2. Upload your image

    3. Enable style detection

    4. Generate animated video

    AI Story Generation

    1. Load WAN2.2-I2V-AutoPrompt-Story.json

    2. Input your story idea

    3. Auto-split into 4 segments

    4. Generate 20-second story video


    🔧 Technical Specifications

    ⚡ Performance

    • Context: 65,536 tokens (8B models)

    • Memory: Optimized VRAM usage

    • Stability: Crash-free operation

    • Speed: Fast generation times

    🎨 Model Support

    • Qwen3-VL 4B: 7 GGUF variants (2.38GB-4.28GB)

    • Qwen3-VL 8B: 7 GGUF variants (4.8GB-8.71GB)

    • HF Models: Josiefed and official variants

    • Quantization: Q4_K_S, Q5_K_S for VRAM efficiency

    🌐 Multilingual Capabilities

    • Input Languages: Any language supported

    • Auto-Translation: Automatic translation to optimized English

    • Style Detection: Works with multilingual prompts

    • Cultural Adaptation: Context-aware prompt enhancement


    🎯 Included Workflows

    🍿 WAN 2.2 Presets

    • 🍿 Wan 2.2 I2V Timeline: Image-to-video with timeline structure

    • 🎥 Wan 2.2 I2V Scene: Image-to-video with professional specs

    • 🍿 Wan 2.2 T2V Timeline: Text-to-video with timeline structure

    • 🎥 Wan 2.2 T2V Scene: Text-to-video with professional specs

    • 🎬 Wan 2.2 I2V Timeline: 20-second multi-segment continuity

    • 📖 Wan 2.2 I2V Scene: 20-second single scene with cinematography

    🔥 Advanced Features

    • NSFW Enhancement: Uncensored content generation

    • Professional Lighting: 8 light types + 9 qualities

    • Camera Control: 6 shot types + 5 compositions

    • Color Grading: 4 tone options


    🎨 Visual Style Detection

    Automatically detects and enhances:

    • Photorealistic style - Realistic lighting, natural textures, lifelike details

    • Anime style - Japanese animation aesthetics with vibrant colors

    • Cartoon style - Bold outlines, flat colors, exaggerated expressions

    • 3D animation style - Computer-generated 3D rendering characteristics

    • Fantasy style - Magical elements, ethereal lighting, imaginative atmosphere

    • Artistic portrait style - Painterly qualities, artistic brushwork, stylized composition


    🔥 NSFW Content Support

    Enhanced Generation

    • Explicit Content: Uncensored adult descriptions

    • Detailed Scenes: 8-12 sentences per segment

    • Natural Progression: Smooth story flow

    • Style Adaptation: Automatic visual style matching

    • Quality: Consistent characters & scenes

    Professional Applications

    • Adult Content: Industry-standard generation

    • Artistic Nudity: Classical art styles

    • Educational: Anatomy and artistic study

    • Creative: Artistic expression


    🚀 Why Choose ComfyUI-QwenVL-Mod?

    🎬 For Content Creators

    • Storytelling: Create compelling narratives

    • Efficiency: One prompt → complete video

    • Quality: Professional video output

    • Flexibility: Any genre, any style

    🔥 For NSFW Content

    • Explicit: Uncensored generation

    • Detailed: Rich scene descriptions

    • Continuous: Smooth story flow

    • Natural: Realistic progression

    ⚡ For Power Users

    • Customizable: Easy to modify

    • Extendable: Add more segments

    • Integrable: Works with existing setups

    • Optimized: Maximum performance


    🌟 What Makes This Special?

    • First: Complete AI story system with vision enhancement

    • Smart: Intelligent prompt splitting and enhancement

    • Complete: End-to-end solution from text to video

    • Optimized: Performance-tuned for professional use

    • Ready: Works out-of-the-box with included workflows


    🎬 Create Amazing AI Videos Today!

    Transform your ideas into stunning videos with the power of Qwen3-VL vision enhancement and WAN 2.2 video generation.

    Perfect for creators, artists, and professionals looking for the ultimate AI video enhancement tool! 🌟


    Built with ❤️ for the ComfyUI community


    The new version is out today, with a brand new anime style ✨

    WORKFLOWS TESTED ON:

    • ComfyUI 0.17.1

    • Python 3.12.12

    • Pytorch 2.9.1 + CUDA 13.0


    🟣 Credits

    This workflows are intended to be used with the models by taek75799 as they follow the structure of Dynamic Prompts you can find under these models:

    WAN 2.2 Enhanced NSFW | SVI | camera prompt adherence (Lightning Edition) I2V and T2V fp8 GGUF

    🟣 Other tested models:

    Thanks to all the users who are commenting and helping me improve the workflows ❤️


    🔖 T2V-I2V AUTOPROMPT

    🛑 Experimental WF 🛑

    Start with a T2V prompt and extend the generated video with I2V

    This workflow requires both T2V and I2V Wan 2.2 Models

    🔖 SVI I2V AUTOPROMPT

    Thanks to taek75799 for his models ❤️

    SVI LORAS:

    LIGHTX2V LoRaS are not included in the model

    🔖 FULL I2V AUTOPROMPT

    Complete workflow that includes:

    • Long Video Generation [from 5 to 20 seconds]

    • Auto Prompting [Qwen3-VL]

    • Upscale [2xLexicaRRDBNet and Tensorrt]

    • Frame Interpolation [30fps and 60fps for img2vid | 24fps and 50fps for MMAudio]

    • MMAudio [NSFW Unlocked]

    📝 AUTO PROMPT

    • Prompt Description Box [Multilanguage]: Just write your idea and the LLM will do the rest, formatting the prompt in the dynamic prompt format used in the Wan 2.2 models

    • Final Prompt Preview: Shows up the final prompt

    🔶 QWEN3-VL NODE FOR GGUF MODELS 🔶

    🔖 To use the Qwen3-VL GGUF Quantized models you have to install llama-cpp-python

    If you are not comfortable in manual install the llama, just go with the normal version inside Full-I2V-LongVideo

    1. 🛑 STOP COMFYUI

    1. 📂 Activate the ComfyUI Virtual Enviroment

    In your ComfyUI root installation folder type:

    on Windows:

    Command Prompt:

    \venv\Scripts\activate.bat

    or Power Shell:

    \venv\Scripts\Activate.ps1

    on Linux:

    . /venv/bin/activate

    If you use ComfyUI Desktop:

    Click on Console and then on Terminal

    1. ⬇️ Install llama-cpp-python

    pip install --upgrade --force-reinstall --no-cache-dir "llama-cpp-python @ git+https://github.com/JamePeng/llama-cpp-python.git"

    1. 🔄 Restart ComfyUI and enjoy

    📝 SWITCH CLIP

    By default the WF uses the GGUF node to load quantized Clip, if you wanna switch to the NSFW Clip model, you have to bypass the GGUF node and connect the other Clip loader to the "Set_CLIP"

    ACCELERATION:

    Triton is disabled by default, you can enable it by opening the first Subgraph

    Inside the workflow you will find all the links to download the models you will need

    That's all, hope you enjoy ^^

    Description

    null

    FAQ

    Comments (125)

    fatberg_slimFeb 15, 2026
    CivitAI

    Hey, thank you for your work. I noticed that llama.cpp does not seem to work with Cuda 12.9 installed. Can you confirm? The QwenVL node silently falls back to CPU and generating a prompt takes a very long time.

    huchukato
    Author
    Feb 15, 2026

    the wheels are for 12.8 or 13.0 but you can compile it for your hardware, for Windows with CUDA (for example) $env:CMAKE_ARGS = "-DGGML_CUDA=on" pip install "llama-cpp-python @ git+https://github.com/JamePeng/llama-cpp-python.git"

    fatberg_slimFeb 15, 2026· 1 reaction

    @huchukato Thank you! ❤️

    dirtysemFeb 15, 2026
    CivitAI

    QwenVL-Mod is not in the repository. You used to have a link in the description on GITHUB. Not right now. The comfyui manager does not search.

    huchukato
    Author
    Feb 15, 2026· 1 reaction
    huchukato
    Author
    Feb 15, 2026· 1 reaction

    BTW you are right I removed the github link coz I was too happy to be in the Manager, I was supposed to leave also the link xDDD

    nailywhite172Feb 15, 2026· 1 reaction
    CivitAI

    Add “IP adapter for the face,” and you'll be counted among the saints.

    huchukato
    Author
    Feb 15, 2026

    Noted for the next update xD

    CrackOutFeb 15, 2026
    CivitAI

    Just a heads up - seems your RunPod container doesn't include a terminal in the Juypeter Lab setup, so awkward pulling anything outside of Civit e.g SVI loras etc. Good job though.

    huchukato
    Author
    Feb 15, 2026

    Thanks for let me know this I will take a look later

    CrackOutFeb 15, 2026· 1 reaction

    @huchukato No worries!

    huchukato
    Author
    Feb 15, 2026

    @CrackOut I just noticed that in the Connect tab there is "Enable Web Terminal" that let you use the RunPod integrated WebTerminal in the browser, BTW in the next build I added A custom webterminal on port 8081, you will find it in the Connect tab <3

    CrackOutFeb 15, 2026· 1 reaction

    @huchukato Ah yes, I've not used the webterminal before, usually used the Juypeter one opened in each file to quickly run a wget command for huggingface whilst using the civic node for the odd hosted model to play with if needed. I'll take a look at the webterminal now, I got by using a HF-Downloader haha!

    FferrettFeb 16, 2026· 1 reaction
    CivitAI

    The T2V/I2V WF is outstanding. I've put the standard KJ FP8 models back into the beginning to get rid of the bias of smoothmix and remix. I'd be interested in seeing a T2V/I2V SVI that can handle a full minute of video and the autoprompt to support it. Excellent work. (Running on 5090 FE, 3.13 cu130, py2.9.1)

    twopoints225221Feb 16, 2026
    CivitAI

    Hi, this looks like an amazing workflow, but I am unfortunately facing the exact same issue as other commenters.

    When I use the "qwen3-vl gguf" node (AILab_QwenVL GGUF) within the Autoprompt group, the node successfully triggers the automatic download of both the GGUF and mmproj files. However, even after the download is complete, I consistently get the following error:

    AILab_QwenVL_GGUF_Advanced: Failed to load model from file: D:\AI\StabilityMatrix-win-x64\Data\Packages\ComfyUI\models\LLM\GGUF\mradermacher\Qwen3-VL-8B-Instruct-c_abliterated-v3-GGUF\Qwen3-VL-8B-Instruct-c_abliterated-v3.Q4_K_M.gguf

    I have tried restarting ComfyUI multiple times, but the error persists. I've spent nearly 12 hours troubleshooting this with Gemini and ChatGPT, but nothing has worked so far. It seems like a persistent bug on Windows environments.

    I am using Windows 11 and Stability Matrix. I suspect it might be related to the long file path or how the node handles GGUF loading on Windows. I would truly appreciate it if you could look into this bug.

    huchukato
    Author
    Feb 17, 2026

    you tried another model? The 4B or another quantization of the 8B?

    twopoints225221Feb 17, 2026

    @huchukato Yes.
    In the model_name (in the qwen3-vl gguf node), I tried:

    qwen3-vl-8b-instruct-c_abliterated-v3.q4_k_m.gguf

    qwen3-vl-8b-instruct-c_abliterated-v3.q6_k.gguf

    qwen3-vl-4b-instruct-c_abliterated-v2.q4_k_m.gguf

    However, I encountered the same error with all of them.

    And following Gemini’s advice, I installed CUDA, the NVIDIA CUDA Toolkit, and Microsoft Visual Studio. I also showed the ComfyUI CMD (console) window to Gemini and ChatGPT, but there was no problem and the result was the same.

    Also the manager inside ComfyUI did not detect any missing files. It seems that all the required files are installed.

    twopoints225221Feb 17, 2026

    @huchukato Also, I have a question. In the "qwen3-vl autoprompt" group, there are four prompts (prompt1–4). I noticed that I can’t directly edit the text written inside them. Is that how it’s supposed to be?

    (I’m referring to the text that says “two woman with vibrant green eyes~”.)

    (I already have the “comfyui-easy-use” node.)

    GooddoggyFeb 17, 2026

    The QwenVL GGUF nodes need llama-cpp-python with vision support. A normal pip install llama-cpp-python often doesn’t include handlers.. Install the vision-capable llama-cpp-python
    python -m pip install --upgrade --no-cache-dir --force-reinstall ` "llama-cpp-python @ git+https://github.com/JamePeng/llama-cpp-python.git"

    twopoints225221Feb 17, 2026

    @drernestbrown474 After following your instructions and reinstalling llama-cpp-python with vision support, qwen3-vl seems to be working now. Thank you for that!

    huchukato
    Author
    Feb 18, 2026

    @twopoints225221 the prompts are generated by Qwen following your prompt and the image, that's why I called it "OneClick", you can also try just with the image and no prompt and Qwen will generate a story following the initial reference image

    twopoints225221Feb 19, 2026

    @huchukato I didn’t know that—what an incredible feature!

    I’m using it well, but I have a question.

    When writing a prompt, how can I make the 'qwen3-vl gguf' node structure the prompt by time segments?

    For example, how would I give a command like, “From 0 to 5 seconds, walk and from 5 to 10 seconds, run”?

    huchukato
    Author
    Feb 20, 2026

    @twopoints225221 if you use the story wf just try Prompt 1: (something) prompt 2 etc, you you use the full i2v autoprompt, you can write a different prompt every 5 seconds by your own and it generates 1 prompt every 5 seconds of video

    twopoints225221Feb 21, 2026· 1 reaction

    @huchukato Thank you!

    DaDomFeb 17, 2026· 1 reaction
    CivitAI

    I see you are the author of the qwen-vl-mod for comfyui. As there is no possibility to make an issue on your github: can you please add ltx2 prompt presets? because now it will make prompts for 5 seconfs, but ltx2 can make 20 secs....

    GooddoggyFeb 18, 2026· 1 reaction
    CivitAI

    Ok, sorry but I am daft. I don't understand how the Qwen3-VL Autoprompt is supposed to work in the workflow. What is the format for what I am supposed to add to the text field. I disable it because it generates things in the prompt that I did not ask for.

    draxskoal608Mar 6, 2026

    t. A simillar work flow called Smooth uses florence2run. Init's case to try to make video from a picture and the ai tries to figure out what'd be going on. In it's case you can disable those nodes manually and describe action yourself. The also did a all in one wokflow kind of thing where you can pick what you want to do. One I grabbed didn't pick up on what ever the author has set up for dependencies, and my reaction is: I have better things to do then fuck around with this.

    twopoints225221Feb 18, 2026
    CivitAI

    I am currently using OneClick-I2v-Svi-Story, and it is working very well — thank you for the great workflow!

    tonedef99Feb 18, 2026
    CivitAI

    Ran into issues with the Qwen-VL GGUF Advanced node where it wasn't cleaning up vram correctly and would fail after 2 runs.

    https://github.com/1038lab/ComfyUI-QwenVL/issues/104

    Ended up swapping the GGUF py file from your repo for the ones here and it hasn't been an issue since. However I do lose all the additions you made to your fork.

    It's a great workflow and runs perfectly on my 4070. Takes about 18 minutes for a full 20 second clip at 640x960, upscaled 30fps, FP8.

    huchukato
    Author
    Feb 18, 2026

    Thanks for this I'll check ASAP

    huchukato
    Author
    Feb 18, 2026· 2 reactions

    Fixed, I will remove the "Clean VRAM Used" nodes after testing the fix, thanks a lot

    pyeeater283Feb 18, 2026
    CivitAI

    having the following error when i try to run this

    SyntaxError: Unexpected non-whitespace character after JSON at position 4 (line 1 column 5)

    huchukato
    Author
    Feb 19, 2026

    which WF?

    fopof4264449Feb 19, 2026· 2 reactions

    I encountered the same error when trying the OneClick-I2V-Story workflow. The latest version of the ComfyUI desktop app is currently 0.14.1 (it doesn't automatically update, even though 0.14.2 is out). I encountered the same issue with that version, and the error disappeared when I unpacked all subgraphs. Try unpacking all subgraphs.

    According to ChatGPT, if you're not using the latest ComfyUI version, something is causing the issue with the subgraphs.

    durachellFeb 22, 2026

    @fopof4264449 what do you mean by "unpacked" all subgraphs?

    huchukato
    Author
    Feb 22, 2026

    Guys I'm using 0.14.2

    durachellFeb 24, 2026

    @huchukato Me too. Didn't fiddle too much, got the same error message.

    PreviousScheme9737506Feb 20, 2026
    CivitAI

    Very nice enjoyable workflow! The autoprompt works great. I wish there was any way to avoid the quality drop in later half of longer videos though, it feel inevitable

    huchukato
    Author
    Feb 20, 2026

    I have to work on the SVI workflow for that but I got issues with slow motion

    @huchukato Yeah I experience same, longer videos loses its point when one have to speed it up , hence again creating a relatively short video😅 I hope you figure out something clever🙌🏻

    ssouthwizzFeb 21, 2026
    CivitAI

    so this can use taek's fp8 nsfw low/high v2? would you recommend high motion or the normal version for realistic scenes?

    phara000868Feb 21, 2026
    CivitAI

    I would really appreciate a little help.

    I have installed everything, followed every step, downloaded the neccessary stuff. I tried to use the mmost simple I2V workflow, without promp help, but my MoE KSampler just refuses to work. It doesn't give me an error, just highlighted in red, and the proccess wont go above like 6%.

    Any ideas?

    i_m_l_OFeb 22, 2026· 1 reaction

    I have the same issue, it lights up red almost instantly when generation starts, the prompt isn't even generated and still it's red, and in console I can see at least "Failed to validate prompt for output..." error.

    i_m_l_OFeb 22, 2026· 1 reaction

    I spent a moment more looking into why this doesn't work in my case;

    * WanMoeKSamplerAdvanced ...

    - Return type mismatch between linked nodes: scheduler, received_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent']) mismatch input_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent', 'beta57'])

    Output will be ignored

    Looks like at least in my case there's something changed somewhere, and the enum doesn't anymore match - there's beta57 in another of the enums, which comes from RES4LYF I think. But if change the sampler to be manually configured within the subgraph, at least this error/issue is gone.

    huchukato
    Author
    Feb 22, 2026· 1 reaction

    the scheduler issue is due to RES4LYF

    emotionaldreams2Feb 22, 2026
    CivitAI

    how would i setup Sageattention to speed things along? or do you have worflows with it already?

    emotionaldreams2Feb 22, 2026

    @huchukato

    also, where the negative prompt box?

    kaibakaiyden413Feb 22, 2026· 1 reaction
    CivitAI

    Hello, I tried to "go with the normal version Full-I2V-LongVideo-v1.7". Yet I am still prompted by ComfyUI to install the QWEN-stuff. Is the workflow in the file correct?

    huchukato
    Author
    Feb 22, 2026

    With the last update this WF contains both normal and GGUF Long Video WFs coz I made too many variants and I was not able to manage them anymore, so I grouped them. To use the normal Qwen3-VL node you have to load WAN2.2-I2V-Full-AutoPrompt-MMAudio-v1-9.json from here https://civitai.com/models/2320999?modelVersionId=2613591

    i_m_l_OFeb 22, 2026
    CivitAI

    The scheduler selection causes currently issues, which stops the video subgraphs from functioning, specifically the Wan MoE KSampler, as it expects different enum (missing at least one scheduler name, in my case beta57. There is a mismatch in the enum with latest ComfyUI node versions of the used custom nodes as of today (23.2)

    huchukato
    Author
    Feb 22, 2026

    I don't have beta57 in latest ComfyUI 0.14.2, maybe you installed it with RES4LYF?

    huchukato
    Author
    Feb 22, 2026

    I tryed to install RES4LYF and it brokes everything coz the KSampler does not recognize the scheduler, if you want to use that node you have to change the KSampler.

    i_m_l_OFeb 23, 2026

    @huchukato Yes, it's not a problem for me, I can get around it easy enough but just thought to point it out that there's possible incompatibilities.

    terrosaurxFeb 26, 2026

    Same here, but I'm on 0.15 which probably is troublesome.

    huchukato
    Author
    Feb 26, 2026

    I wrote about RES4LYF in the readme, didn't tested 0.15 yet

    SmurfypieFeb 27, 2026

    @i_m_l_O how do you get around it easily? do you replace the ksampler? do I need 2 ksamplers now cause the regular clownsharksampler only got 1 model connection.

    huchukato
    Author
    Feb 27, 2026

    @Smurfypie I think you need 2 clownshark, I didn't tested it till now

    draxskoal608Mar 7, 2026

    ok cock fucker. Prompt outputs failed validation: AILab_QwenVL_GGUF_Advanced: - Value not in list: preset_prompt: '🍿 Wan 2.2 NSFW I2V' not in ['🖼️ Tags', '🖼️ Simple Description', '🖼️ Detailed Description', '🖼️ Ultra Detailed Description', '🎬 Cinematic Description', '🖼️ Detailed Analysis', '📹 Video Summary', '📖 Short Story', '🪄 Prompt Refine & Expand'] WanMoeKSamplerAdvanced: - Return type mismatch between linked nodes: scheduler, received_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent']) mismatch input_type(['simple', 'sgm_uniform', 'karras', 'exponential', 'ddim_uniform', 'beta', 'normal', 'linear_quadratic', 'kl_optimal', 'bong_tangent', 'beta57']) . what piece of shit.

    draxskoal608Mar 7, 2026

    @huchukato huchukato

    OP

    9 days ago

    " I wrote about RES4LYF in the readme, didn't tested 0.15 yet" wtf are you even trying to say?

    SmurfypieFeb 27, 2026
    CivitAI

    QwenVLGGUFBase.run() missing 1 required positional argument: 'unload_after_run' I get this error and cant seem to fix it.

    huchukato
    Author
    Feb 27, 2026· 1 reaction

    Let me check, I did some changes yesterday, I fix it ASAP

    huchukato
    Author
    Feb 28, 2026· 1 reaction

    fixed, update my node and download the WF again

    cielfyreMar 1, 2026
    CivitAI

    I am confused about the upscalers. If I want to use tensorrt upscaler do I turn on both Enable Upscaler and Enable Upscaler TENSORRT, or just Enable Upscaler TENSORRT?

    huchukato
    Author
    Mar 3, 2026

    Just the Tensorrt one and link the "setupscale" setnode the the one you are using

    FreshLobbyMar 1, 2026
    CivitAI

    I'm getting this error when the workflow moves onto the 10s subgraph.

    AILab_QwenVL_Advanced

    t:1 must be larger than temporal_factor:2

    This is on the I2V-Full-AutoPrompt-MMAudio workflow. Any ideas?

    FreshLobbyMar 1, 2026

    I'm also stuck at "Loading checkpoint shards: 100%" when using the Qwen3-VL-8B-Instruct-Ablitered model. Waited over 20 minutes. I'm able to generate a 5s video with Qwen3-VL-8B-Instruct and Qwen3-VL-8B-Instruct-FP8 models however.

    evantopsmithMar 4, 2026
    CivitAI

    Is it supposed to build a rife engine every single time I run the workflow?

    draxskoal608Mar 7, 2026

    The shitstain didn't even bother to test some stuff it's just a broken cunt ass add.

    draxskoal608Mar 7, 2026

    Probably. the shit stain has no idea wtf he's doing

    pipindirovskigorjo318Mar 5, 2026
    CivitAI

    I am getting an error about OOM when using the I2V prompt workflow, specifically when using Qwen ( any model, doesn't matter if it fits GPU ):
    AILab_QwenVL_Advanced

    Allocation on device 0 would exceed allowed memory. (out of memory) Currently allocated : 0 bytes Requested : 1.16 GiB Device limit : 31.37 GiB Free (according to CUDA): 14.46 GiB PyTorch limit (set by user-supplied memory fraction) : 17179869184.00 GiB This error means you ran out of memory on your GPU. TIPS: If the workflow worked before you might have accidentally set the batch_size to a large number.

    What the hell is going on? Has someone else had this issue?

    huchukato
    Author
    Mar 5, 2026

    which WF? the Story one or the Full one?

    @huchukato Specifically WAN2.2-I2V-AutoPrompt.json, it's not story or full, I just want the basic Image to Video, but it keeps throwing that weird memory error.

    huchukato
    Author
    Mar 5, 2026

    @pipindirovskigorjo318 i check and let you know ASAP

    huchukato
    Author
    Mar 6, 2026

    Updated both normal and GGUF Single Video WFs, download them again and let me know https://civitai.com/models/2320999?modelVersionId=2624175

    @huchukato Same error unfortunately :
    [QwenVL] Loading Qwen3-VL-4B-Instruct-Abliterated (8-bit (Balanced), attn=sdpa)

    Loading weights: 0%| | 0/713 [00:00<?, ?it/s]

    !!! Exception during processing !!! Allocation on device 0 would exceed allowed memory. (out of memory)

    Currently allocated : 9.12 MiB

    Requested : 741.88 MiB

    Device limit : 31.37 GiB

    Free (according to CUDA): 22.41 GiB

    PyTorch limit (set by user-supplied memory fraction)

    : 17179869184.00 GiB

    Traceback (most recent call last):

    File "/ComfyUI/execution.py", line 524, in execute

    output_data, output_ui, has_subgraph, has_pending_tasks = await get_output_data(prompt_id, unique_id, obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, v3_data=v3_data)

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/ComfyUI/execution.py", line 333, in get_output_data

    return_values = await asyncmap_node_over_list(prompt_id, unique_id, obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, v3_data=v3_data)

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/ComfyUI/execution.py", line 307, in asyncmap_node_over_list

    await process_inputs(input_dict, i)

    File "/ComfyUI/execution.py", line 295, in process_inputs

    result = f(**inputs)

    ^^^^^^^^^^^

    File "/ComfyUI/custom_nodes/ComfyUI-QwenVL-Mod/AILab_QwenVL.py", line 706, in process

    return self.run(model_name, quantization, preset_prompt, custom_prompt, image, video, frame_count, max_tokens, temperature, top_p, num_beams, repetition_penalty, seed, keep_model_loaded, attention_mode, use_torch_compile, device, keep_last_prompt)

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/ComfyUI/custom_nodes/ComfyUI-QwenVL-Mod/AILab_QwenVL.py", line 585, in run

    self.load_model(

    File "/ComfyUI/custom_nodes/ComfyUI-QwenVL-Mod/AILab_QwenVL.py", line 468, in load_model

    self.model = AutoModelForVision2Seq.from_pretrained(model_path, **load_kwargs).eval()

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/opt/comfyui-env/lib/python3.12/site-packages/transformers/models/auto/auto_factory.py", line 374, in from_pretrained

    return model_class.from_pretrained(

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/opt/comfyui-env/lib/python3.12/site-packages/transformers/modeling_utils.py", line 4137, in from_pretrained

    loading_info, disk_offload_index = cls._load_pretrained_model(model, state_dict, checkpoint_files, load_config)

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/opt/comfyui-env/lib/python3.12/site-packages/transformers/modeling_utils.py", line 4256, in loadpretrained_model

    loading_info, disk_offload_index = convert_and_load_state_dict_in_model(

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/opt/comfyui-env/lib/python3.12/site-packages/transformers/core_model_loading.py", line 1212, in convert_and_load_state_dict_in_model

    realized_value = mapping.convert(

    ^^^^^^^^^^^^^^^^

    File "/opt/comfyui-env/lib/python3.12/site-packages/transformers/core_model_loading.py", line 678, in convert

    collected_tensors = self.materialize_tensors()

    ^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/opt/comfyui-env/lib/python3.12/site-packages/transformers/core_model_loading.py", line 654, in materialize_tensors

    tensors = [future.result() for future in tensors if future.result() is not None]

    ^^^^^^^^^^^^^^^

    File "/usr/lib/python3.12/concurrent/futures/_base.py", line 449, in result

    return self.__get_result()

    ^^^^^^^^^^^^^^^^^^^

    File "/usr/lib/python3.12/concurrent/futures/_base.py", line 401, in __get_result

    raise self._exception

    File "/usr/lib/python3.12/concurrent/futures/thread.py", line 58, in run

    result = self.fn(*self.args, **self.kwargs)

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/opt/comfyui-env/lib/python3.12/site-packages/transformers/core_model_loading.py", line 800, in _job

    return materializecopy(tensor, device, dtype)

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    File "/opt/comfyui-env/lib/python3.12/site-packages/transformers/core_model_loading.py", line 789, in materializecopy

    tensor = tensor.to(device=device, dtype=dtype)

    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

    torch.OutOfMemoryError: Allocation on device 0 would exceed allowed memory. (out of memory)

    Currently allocated : 9.12 MiB

    Requested : 741.88 MiB

    Device limit : 31.37 GiB

    Free (according to CUDA): 22.41 GiB

    PyTorch limit (set by user-supplied memory fraction)

    : 17179869184.00 GiB

    Memory summary: |===========================================================================|

    | PyTorch CUDA memory summary, device ID 0 |

    |---------------------------------------------------------------------------|

    | CUDA OOMs: 0 | cudaMalloc retries: 0 |

    |===========================================================================|

    | Metric | Cur Usage | Peak Usage | Tot Alloc | Tot Freed |

    |---------------------------------------------------------------------------|

    | Allocated memory | 83589 KiB | 8418 MiB | 0 B | 0 B |

    | from large pool | 0 KiB | 0 MiB | 0 B | 0 B |

    | from small pool | 0 KiB | 0 MiB | 0 B | 0 B |

    |---------------------------------------------------------------------------|

    | Active memory | 83589 KiB | 8418 MiB | 0 B | 0 B |

    | from large pool | 0 KiB | 0 MiB | 0 B | 0 B |

    | from small pool | 0 KiB | 0 MiB | 0 B | 0 B |

    |---------------------------------------------------------------------------|

    | Requested memory | 0 B | 0 B | 0 B | 0 B |

    | from large pool | 0 B | 0 B | 0 B | 0 B |

    | from small pool | 0 B | 0 B | 0 B | 0 B |

    |---------------------------------------------------------------------------|

    | GPU reserved memory | 131072 KiB | 8480 MiB | 0 B | 0 B |

    | from large pool | 0 KiB | 0 MiB | 0 B | 0 B |

    | from small pool | 0 KiB | 0 MiB | 0 B | 0 B |

    |---------------------------------------------------------------------------|

    | Non-releasable memory | 0 B | 0 B | 0 B | 0 B |

    | from large pool | 0 B | 0 B | 0 B | 0 B |

    | from small pool | 0 B | 0 B | 0 B | 0 B |

    |---------------------------------------------------------------------------|

    | Allocations | 0 | 0 | 0 | 0 |

    | from large pool | 0 | 0 | 0 | 0 |

    | from small pool | 0 | 0 | 0 | 0 |

    |---------------------------------------------------------------------------|

    | Active allocs | 0 | 0 | 0 | 0 |

    | from large pool | 0 | 0 | 0 | 0 |

    | from small pool | 0 | 0 | 0 | 0 |

    |---------------------------------------------------------------------------|

    | GPU reserved segments | 0 | 0 | 0 | 0 |

    | from large pool | 0 | 0 | 0 | 0 |

    | from small pool | 0 | 0 | 0 | 0 |

    |---------------------------------------------------------------------------|

    | Non-releasable allocs | 0 | 0 | 0 | 0 |

    | from large pool | 0 | 0 | 0 | 0 |

    | from small pool | 0 | 0 | 0 | 0 |

    |---------------------------------------------------------------------------|

    | Oversize allocations | 0 | 0 | 0 | 0 |

    |---------------------------------------------------------------------------|

    | Oversize GPU segments | 0 | 0 | 0 | 0 |

    |===========================================================================|

    huchukato
    Author
    Mar 6, 2026

    @pipindirovskigorjo318 after a loooong night of debugging I finally fixed it xD Download the WF again and update my Qwen Node (just do an Update All in ComfyUI Manager) and it will work

    malurMar 13, 2026

    i'm not crashing with an OOM message, but my 5090 runs full, and then it just stops processing, the 4B or 8B model doesn't matter.

    huchukato
    Author
    Mar 13, 2026

    @malur You are running the last WFs version and the last Qwen node version? coz I fixed that

    malurMar 13, 2026

    @huchukato i downloaded them literally earlier today, after some more testing, i guess only the 8B-MAX VL Version doesn't like to run consistently. fresh launch and it runs, next run it just just stops or takes ages, idk, i stopped it after like 10 minutes last time. can i somehow convert it to gguf or nvfe4 to save on some vram? i mean, 8B is huge anyway.

    huchukato
    Author
    Mar 14, 2026

    @malur mmm I disable the quantization in the normal node coz it caused a lot of problems so in that node all the models runs only in FP16, if you wanna use the quantized models you have to go for the GGUF node and install llama cpp

    ThunderOakMar 6, 2026
    CivitAI

    Looking very much to trying these workflows. Do you have a Url for the StoryTeller repository? The one I googled is no longer available.

    Thanks

    xgdxyzs944Mar 8, 2026
    CivitAI

    I noticed that some parameters in Oneclick and fulllongvideo seem to be misaligned. For example, it shows max token 0, temperature 512, top p 0.6, but normally it should be max token 512, temperature 0.6, and so on. The ComfyUI version is 0.16.3. I tried switching versions, but the issue persists. The OneVideo workflow parameters are normal and functioning correctly.

    huchukato
    Author
    Mar 8, 2026

    I did some updated to the node and maybe the parameters are not aligned, you can reload the node to fix this, I will update the WFs asap

    huchukato
    Author
    Mar 8, 2026

    Updated almost all the WFs

    yafkinzebi617Mar 8, 2026
    CivitAI

    Hello, i dont understand how im supposed to install qwen LV model, can someone help me please ?

    huchukato
    Author
    Mar 9, 2026· 1 reaction

    Just select a model in the node and it will auto download it

    domhillMar 9, 2026
    CivitAI

    One day ill find a work flow that works for me, every single one always has a point where something will just refuse the install, this time its tensorrt

    huchukato
    Author
    Mar 9, 2026

    My Tensorrt custom node have autoinstall, strange

    domhillMar 9, 2026

    @huchukato mine makes me restart the backend, the backend never restarts and when i manually restart it need installing again. Maybe Comfyui just isn't for me lol

    tanyakkkMar 26, 2026

    lol, same for me

    xsulaiMar 9, 2026
    CivitAI

    Why does any model loaded to qwinVL node if it is Q4 it gives me uncensored response but Q8 never give me uncensored response?

    xgdxyzs944Mar 10, 2026
    CivitAI

    Found a bug in qwenvl-mod: NSFW content is missing from the generated prompts. I found on GitHub that the AILab_System_Prompts.json file was modified on Mar 5, 2026, and the NSFW-related descriptions were removed from the prompt.

    Just got home and added the NSFW prompt back in—tested it and everything works fine now.

    huchukato
    Author
    Mar 13, 2026

    I checked the preset now and there is

    WHEN there are NSFW images or text, provide an NSFW description consistent with the requested artistic style

    as always, I just changed the presets names

    huchukato
    Author
    Mar 13, 2026

    Ok I see the problem, the extended NSFW check was only in 1 preset. Fixed the others right now

    FferrettMar 13, 2026· 1 reaction
    CivitAI

    Now we just need a "OneClick-T2V I2V SVI -Story

    huchukato
    Author
    Mar 13, 2026

    I planned to make it but I got slow motion issues with SVI and I was trying to fix that issues before

    bugsbelightyear730Mar 14, 2026

    @huchukato make sure to hardcode more repos that are half baked in while your at it. oh and more cunt-um nodes. I don't think people are trolled enough make sure it never garabage collects to while you vibe code.

    huchukato
    Author
    Mar 15, 2026

    @bugsbelightyear730 Man, if you don't like what I did, just don't use it. Easy peasy.

    domcashiki636Mar 15, 2026· 1 reaction
    CivitAI

    when i try to use any kind of wan2.2 lora

    it seems to turn into a blurry mess, had to adjust a few settings to keep moving people from becoming blurry. however couldn t get loras working. are there specific lora type to be used for this?

    frugaltechMar 17, 2026

    Having the same issue

    huchukato
    Author
    Mar 17, 2026

    Depends on which Wan 2.2 Model are you running

    johnn3101Mar 25, 2026

    your step size is too low. You need to adjust your stepsize to higher like 20, or use a lightning LORA

    rolthof89Mar 20, 2026
    CivitAI

    You should add a referral code for Vast.AI and RunPod, happy to have used it!

    lighthorsexajz830Mar 20, 2026
    CivitAI

    WE'LL TEST TO SEE AND THEN I'LL MAKE A RETURN

    oldirtysancho452Mar 21, 2026
    CivitAI

    There are issues and conflicts with this setup and the nodes it's looking for, at least on Linux. I can't fully explain them because I had to poke at it a bit but first I thought there was an issue with the ~/models/llm directory because the NSFW Qwen LLM wasn't showing up and something is looking for /LLM and something else is looking in /llm which are separate directories in Linux but not in Windows. Then I realized that if /ComfyUI-QwenVL exists then the node picks that instead of the intended /ComfyUI-QwenVL-Mod. So I had to remove /ComfyUI-QwenVL and then the node showed the expected gguf listings. I would assume this issue would have been recognized in the runpod setup since that is Linux, I'm running it locally on a Linux box.

    lighthorsexajz830Mar 21, 2026
    CivitAI

    UNFORTUNATELY, NOTHING IS WORKING PROPERLY. WHAT A SHAME.

    shapeshifter83Mar 21, 2026· 2 reactions
    CivitAI

    If you think you came here to find workflows to get running WAN right now, you came to the wrong place. At least at this moment in time. All of these are currently impossibly broken.

    hobo006Mar 22, 2026
    CivitAI

    shits taking 7min to forge a 5sec video on a 5080 on 832*480 res. no tutorial no nothing

    komemo46713Mar 27, 2026
    CivitAI

    you speak about motion frames in imtovid Description: Number of trailing frames from previous_video used as motion reference
    Default: 5 where isi it? i don't understand that ? can you help me?

    burneremail392776Mar 28, 2026
    CivitAI

    I keep getting image token error. I believe all the nodes are up to date. Anyone have any advice on the issue?

    [load_hparams: Qwen-VL models require at minimum 1024 image tokens to function correctly on grounding tasks

    load_hparams: if you encounter problems with accuracy, try adding --image-min-tokens 1024

    load_hparams: more info: https://github.com/ggml-org/llama.cpp/issues/16842]

    CowgaryMar 30, 2026· 1 reaction
    CivitAI

    Using I2V AutoPrompt Story and it's throwing TypeError: VRAMCleanup.empty_cache() got an unexpected keyword argument 'input'. Swap out to LayerUtility VRAMCleanup and same error.

    messinajonathan2459Apr 2, 2026· 2 reactions

    because the stupid guy on the node 5s / 10s / 15s / 20s . Is sending wan moe decode latent *purplee dot to > vram cleanup input.

    When he should be sending latent to > wan vae decode > samples.

    from there send wan vae decode output to image

    delete vram cleanup.

    Because i dont know where in the hell i should connect that

    also disconnect the node in wan moe decode > sample name > just leave it without the green cable. Otherwise will thow error of mismatch.

    and now it works. Repeat this in all others nodes, meaning 10s 15s 20s.

    mkbmbt126com649Apr 20, 2026

    @messinajonathan2459 Thank you for the solution. I delete the vramcleanup node and it works fine. Still a wonderful workflow.

    SirManateeMar 31, 2026
    CivitAI

    For anyone experiencing Qwen prompt issues on the "Story" workflows. Qwen3-VL-4B will not be able to understand the Timeline (20s) instructions. Had to switch to a quantized 8B model for it to understand and generate the prompts correctly.

    messinajonathan2459Apr 2, 2026
    CivitAI

    is totally broken given errors of mismatch

    [QwenVL] GPU memory after load: 16.3GB / 31.8GB

    got prompt

    Failed to validate prompt for output 1327:

    * VAEDecode 1252:1658:

    - Return type mismatch between linked nodes: samples, received_type(*) mismatch input_type(LATENT)

    Output will be ignored

    Failed to validate prompt for output 1252:1657:

    * (prompt):

    - Return type mismatch between linked nodes: input, received_type(LATENT) mismatch input_type(*)

    * VRAMCleanup 1252:1657:

    - Return type mismatch between linked nodes: input, received_type(LATENT) mismatch input_type(*)

    Output will be ignored

    Failed to validate prompt for output 1314:

    * AutoRifeTensorrt 1596:

    - Return type mismatch between linked nodes: frames, received_type(*) mismatch input_type(IMAGE)

    Output will be ignored

    Failed to validate prompt for output 1608:

    * (prompt):

    - Return type mismatch between linked nodes: input, received_type(IMAGE) mismatch input_type(*)

    * VRAMCleanup 1608:

    - Return type mismatch between linked nodes: input, received_type(IMAGE) mismatch input_type(*)

    Output will be ignored

    Failed to validate prompt for output 1596:

    * (prompt):

    - Return type mismatch between linked nodes: frames, received_type(*) mismatch input_type(IMAGE)

    Output will be ignored

    Failed to validate prompt for output 1547:

    Output will be ignored

    got prompt

    Failed to validate prompt for output 1327:

    * VAEDecode 1252:1658:

    - Return type mismatch between linked nodes: samples, received_type(*) mismatch input_type(LATENT)

    Output will be ignored

    Failed to validate prompt for output 1252:1657:

    * (prompt):

    - Return type mismatch between linked nodes: input, received_type(LATENT) mismatch input_type(*)

    * VRAMCleanup 1252:1657:

    - Return type mismatch between linked nodes: input, received_type(LATENT) mismatch input_type(*)

    Output will be ignored

    Failed to validate prompt for output 1314:

    * AutoRifeTensorrt 1596:

    - Return type mismatch between linked nodes: frames, received_type(*) mismatch input_type(IMAGE)

    Output will be ignored

    Failed to validate prompt for output 1608:

    * (prompt):

    - Return type mismatch between linked nodes: input, received_type(IMAGE) mismatch input_type(*)

    * VRAMCleanup 1608:

    - Return type mismatch between linked nodes: input, received_type(IMAGE) mismatch input_type(*)

    Output will be ignored

    Failed to validate prompt for output 1596:

    * (prompt):

    - Return type mismatch between linked nodes: frames, received_type(*) mismatch input_type(IMAGE)

    Output will be ignored

    Failed to validate prompt for output 1547:

    Output will be ignored

    got prompt

    Failed to validate prompt for output 1327:

    * VAEDecode 1252:1658:

    - Return type mismatch between linked nodes: samples, received_type(*) mismatch input_type(LATENT)

    Output will be ignored

    Failed to validate prompt for output 1547:

    Output will be ignored

    Failed to validate prompt for output 1252:1657:

    * (prompt):

    - Return type mismatch between linked nodes: input, received_type(LATENT) mismatch input_type(*)

    * VRAMCleanup 1252:1657:

    - Return type mismatch between linked nodes: input, received_type(LATENT) mismatch input_type(*)

    messinajonathan2459Apr 2, 2026· 1 reaction
    CivitAI

    the fucking stupid guy sent wan moe decoder latent to vram clean up
    when you should be sending latent to > van decode samples.

    vram cleanup is fucked up or something is missing

    ilo444Apr 6, 2026
    CivitAI

    is nsfw blocked in this flow? Can't seem to get any nsfw content

    dantufis664Apr 10, 2026
    CivitAI

    auto prompt works fine, but no video is generated. i don't know what i am doing wrong, no errors i receive

    hardwire666Apr 14, 2026

    Same. Just keep spitting out empty video files and two png's.

    dantufis664Apr 15, 2026· 2 reactions

    i discover that i have some issues with sage, i bypass the node and woks fine now.

    hardwire666Apr 15, 2026

    @dantufis664 Good to know. Thanks!

    fivebigbooms935509Apr 10, 2026
    CivitAI

    The workflow works great, but I am having an issue with the Qwen3VL nodes - they seem to be running on CPU even though I have cuda selected as the device to use. Takes a very long time to load each prompt. Any known fix?

    stoneroses721Apr 19, 2026
    CivitAI

    I watched the tutorial video and followed it as it was, but most of the models didn't install, and I couldn't. Just wasted time and runpod credits.

    zyeesiApr 19, 2026
    CivitAI

    Using your WAN2.2-I2V-AutoPrompt with 60FPS enabled, I get this error:

    File "/ComfyUI/execution.py", line 308, in _async_map_node_over_list await process_inputs(input_dict, i) File "/ComfyUI/execution.py", line 296, in process_inputs result = f(**inputs) ^^^^^^^^^^^ File "/ComfyUI/custom_nodes/ComfyUI-RIFE-TensorRT-Auto/__init__.py", line 420, in load_rife_tensorrt_model engine.build( File "/ComfyUI/custom_nodes/ComfyUI-RIFE-TensorRT-Auto/trt_utilities.py", line 338, in build raise RuntimeError("TensorRT not available - please install TensorRT first")

    I installed requirements_cu12.txt to fix it, but I feel like I shouldn't have really done that.

    Did I miss a config setting, picked a wrong pod? Don't want to reinstall again the next time

    persindon644Apr 24, 2026
    CivitAI

    The battle with TensorRT lasted half a day :) But I emerged victorious. I'll describe this using the RTX5060ti drv 595.79 torch as an example: 2.10.0+cu130 Python 3.13.11 ComfyUI 0.19.4

    The problem is that it's looking for a dll in a folder that doesn't exist - ComfyUI_windows_portable\python_embeded\Lib\site-packages\tensorrt\tensorrt.libs

    But simply creating it doesn't help.

    1. Download the one you need from https://developer.nvidia.com/tensorrt/download/10x

    2. Unzip the TensorRT-10.16.1.11.Windows.amd64.cuda-13.2.zip archive to G:\TensorRT). You should now have a TensorRT folder containing python, lib, bin, etc. The python folder contains the required whl (do not use the dispatch and lean versions).

    3. Install from ComfyUI_windows_portable via CMD - .\python_embeded\python.exe -m pip install G:\TensorRT\python\tensorrt-10.16.1.11-cp313-none-win_amd64.whl - substitute the required .whl here.

    4. Create a tensorrt.libs folder in the G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\tensorrt folder.

    5. Copy all files from the bin folder of the TensorRT-10.16.1.11.Windows.amd64.cuda-13.2.zip archive to the folder. G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\tensorrt\tensorrt.libs

    Open the Environment Variables window:

    Press Win + R, type sysdm.cpl, and press Enter. Go to the "Advanced" → "Environment Variables..." tab. Find and edit the Path variable: Under "System Variables," double-click Path, click New

    And paste the path:

    G:\ComfyUI_windows_portable\python_embeded\Lib\site-packages\tensorrt\tensorrt.libs

    Save the changes. Click "OK" in all open windows. Afterwards, be sure to restart your computer for the changes to take effect. Done! TensorRT is installed. Now restart ComfyUI. The ComfyUI-RIFE-TensorRT-Auto plugin should load without errors, and the AutoLoadRifeTensorrtModel and AutoRifeTensorrt nodes will become available.

    Workflows
    Wan Video 2.2 I2V-A14B

    Details

    Downloads
    7,506
    Platform
    CivitAI
    Platform Status
    Available
    Created
    2/15/2026
    Updated
    4/30/2026
    Deleted
    -

    Files

    WAN22NSFWI2VT2VWorkflowsAutoPrompt_OneclickI2VStory.zip

    WAN22NSFWI2VT2VWorkflowsAutoPrompt_oneclickI2VStory.zip

    WAN22NSFWI2VT2VWorkflowsAutoPrompt_OneclickI2VStory.zip

    WAN22NSFWI2VT2VWorkflowsAutoPrompt_OneclickI2VStory.zip

    WAN22NSFWI2VT2VWorkflowsAutoPrompt_oneclickI2VStory.zip

    WAN22NSFWI2VT2VWorkflowsAutoPrompt_OneclickI2VStory.zip

    WAN22NSFWI2VT2VWorkflowsAutoPrompt_OneclickI2VStory.zip