CivArchive
    WAN 2.1 IMAGE to VIDEO with Caption and Postprocessing - v3.0
    NSFW
    Preview 79186984

    Workflow: Image -> Autocaption (Prompt) -> WAN I2V with Upscale and Frame Interpolation and Video Extension

    • Creates Video Clips with up to 480p resoltion (720p with corresponding model)

    There is a Florence Caption Version and a LTX Prompt Enhancer (LTXPE) version. LTXPE is more heavy on VRAM

    LTX Prompt Enhancer (LTXPE) might have issues with latest Comfy and Lightricks update


    MultiClip: Wan 2.1. I2V Version supporting Fusion X Lora to create clips with 8 steps and extend up to 3 times, see examples posted with 15-20sec of length.

    Workflow will create a clip on Input Image and extends it with up to 3 clips/sequences. It uses a colormatch feature to ensure consistency in color and light in most cases. See the notes in worflow with full details.

    There is a normal version which allows to use own prompts and a version using LTXPE for autoprompting. Normal version works well for specific or NSFW clips with Loras and the LTXPE is made to just drop an image, set width/height and hit run. The clips are combined to one full video at the end.

    update 16th of July 2025: A new Lora "LightX2v"has been released as an alternative to Fusion X Lora. To use, switch Lora in black "Lora Loader" node. It can create great motion with only 4-6 steps. : https://huggingface.co/lightx2v/Wan2.1-I2V-14B-480P-StepDistill-CfgDistill-Lightx2v/tree/main/loras

    More info/tips & help: https://civarchive.com/models/1309065/wan-21-image-to-video-with-caption-and-postprocessing?dialog=commentThread&commentId=869306


    V3.1: Wan 2.1. I2V Version supporting Fusion X Lora for fast processing

    Fusion X Lora: process the video with just 8 Steps (or lower, see notes in workflow). It does not have the issues like the CausVid Lora from V3.0 and does not require a color match correction.

    Fusion X Lora can be downloaded here: https://civarchive.com/models/1678575?modelVersionId=1900322 (i2V)


    V3.0: Wan 2.1. I2V Version supporting Optimal Steps Scheduler (OSS) and CausVid Lora

    • OSS is a newer comfy core node to allow lower no. of steps with a boost in quality. Instead of using 50+ steps you can receive same result with like 24 steps. https://github.com/bebebe666/OptimalSteps

    • CausVid uses a Lora to process the video with just 8-10 steps, it is fast at a lower quality. It contains a Color Match option in postprocessing to cope with the increased saturation, the lora is introducing. Lora can be downloaded here: https://huggingface.co/Kijai/WanVideo_comfy/tree/main

      (Wan21_CausVid_14B_T2V_lora_rank32.safetensors)

    • Both have a version with FLorence or LTX Prompt Enhancer (LTXPE) for Caption, can use Loras and have Teacache included.


    V2.5: Wan 2.1. Image to Video with Lora Support and Skip Layer Guidance (improves motion)

    There are 2 version, Standard with Teacache, Florence caption, upscale, frame interp. etc. plus a version with LTX Prompt Enhancer as an additional captioning tool (see notes for more info, requires custom nodes: https://github.com/Lightricks/ComfyUI-LTXVideo).

    For Lora use, recommend to switch to own prompt with Lora trigger phrase, complex prompts might confuse some Loras.


    V2.0: Wan 2.1. Image to Video with Teacache support for GGUF model, speeds up generation by 30-40%

    It will render the first steps with normal speed, remaining steps with higher speed. There is a minor impact on quality with more complex motion. You can bypass the Teacache node with Strg-B

    Example clips with workflow in Metadata: https://civarchive.com/posts/13777557

    Info and help with Teacache: https://civarchive.com/models/1309065/wan-21-image-to-video-with-caption-and-postprocessing?dialog=commentThread&commentId=724665


    V1.0: WAN 2.1. Image to Video with Florence caption or own prompt plus upscale, frame interpolation and clip extend.

    Workflow is setup to use a GGUF model.

    When generating a Clip you can chose to apply upscaling and/or frame interpolation. Upscale factor depends on upscale model used (2x or 4x, see "load upscale model" node). Frame Interpolation is set to increase frame rate from 16fps (model standard) to 32fps. Result will be shown in "Video Combine Final" node on the right, while the left node shows the unprocessed clip.

    Recommend to "Toggle Link visibility" to hide the cables.


    Models can be downloaded here:

    Wan 2.1. I2V (480p): https://huggingface.co/city96/Wan2.1-I2V-14B-480P-gguf/tree/main

    Clip (fp8): https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/tree/main/split_files/text_encoders

    Clip Vision: https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/tree/main/split_files/clip_vision

    VAE: https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/tree/main/split_files/vae


    Wan 2.1. I2V (720p): https://huggingface.co/city96/Wan2.1-I2V-14B-720P-gguf/tree/main

    Wan2.1. Text to Video (works): https://huggingface.co/city96/Wan2.1-T2V-14B-gguf/tree/main


    location to save those files within your Comfyui folder:

    Wan GGUF Model -> models/unet

    Textencoder -> models/clip

    Clipvision -> models/clip_vision

    Vae -> models/vae


    Tips:

    • lower framerate in "Video combine Final" node from 30 to 24 to have a slow motion effect

    • You can use the Text to Video GGUF Model, it will work as well.

    • If video output shows strange artifacts on the very right side of a frame, try changing the parameter "divisible_by" in node "Define Width and Height" from 8 to 16, this might better latch on to the standard Wan resolution and avoid the artifacts.

    • see this thread if you face issues with LTX Prompt Enhancer: https://civarchive.com/models/1823416?dialog=commentThread&commentId=955337

    • Last Frame: If you face issues finding the pack for that node: https://github.com/DoctorDiffusion/ComfyUI-MediaMixer

    Full Video with Audio example:

    Description

    Wan Image2Video with CausVid Lora for fast processing

    Wan Image2Video with OSS for high quality

    (both come with a Florence or LTXPE Autocaption version, each as .json & .png file.)

    FAQ

    Comments (15)

    tremolo28
    Author
    May 29, 2025· 10 reactions
    CivitAI
    tremolo28
    Author
    May 31, 2025

    @ToxicBot thanks, this Lora is the best, but it is for Flux (not Wan). Here is the same lora on civit: https://civitai.com/models/1295758/

    ToxicBotMay 31, 2025· 1 reaction

    @tremolo28 Woops, it's been a long night, totally embarrassed now lol.

    LONGERSTARJun 5, 2025· 1 reaction
    CivitAI

    Thank you for the great work! However, when I use 3.0 OSS workflow, it causes noise glitch on every single video (even on the internet). Would you happen know why?

    tremolo28
    Author
    Jun 5, 2025

    Hi, not realy a clue, assumable your comfyui is up to date.

    Best guess is trying another sampler, i.e uni_pc and or play with the value in node "ModelSamplingSD3" (values between 2-10 might make sense). Ensure Model type =Wan in node "OptimalStepsScheduler"

    LONGERSTARJun 5, 2025

    @tremolo28 Thank you so much for the detailed response, I will try them for sure!

    tremolo28
    Author
    Jun 6, 2025

    @LONGERSTAR seems its working for you now, thanks for posting a clip

    LONGERSTARJun 6, 2025· 1 reaction

    @tremolo28 More testing is needed but right now, setting ModelSamplingSD3 to 2 did the job it seems. Thank you again for the solution!

    LONGERSTARJun 9, 2025

    @tremolo28 Hi! I'm sorry for asking another question but I just can't seem to figure this one out either. For "Florence CAUS" workflow, I would like to lower the CAUS LORA strength to enhance quality. But when I lower the value the videos come out with heavy amounts of afterimages or overlapping images (not sure about proper terms). I tried but I'm too noob to solve this I think, another advice would be appreciated, thank you!

    tremolo28
    Author
    Jun 9, 2025

    @LONGERSTAR Hi, did not use the CausVid Lora that much, I am more with the OSS version. Anyway, from testing , I remember to had only good result with a strength between 0.7-0.9. Another parameter to check is the sampler and scheduler. I recall to had same effect as you describe, like changing scenes, more like a T2V than I2V effect.

    LONGERSTARJun 9, 2025

    @tremolo28 So changing samplers and schedulers could possibly fix the issue. Thank you for responding and will try this one out as well!

    tremolo28
    Author
    Jun 17, 2025· 9 reactions
    CivitAI

    Here some tips regarding prompting, when using own prompts with I2V:

    In general, you do not need to describe what is in the picture, since ClipVision will see it anyway from the input picture. Instead focus on what shall happen, what the camera/view is supposed to do and end the prompt with a general mood or what is the source/type of the clip. 2-3 simple phrases will do it.

    Camera Control:

    "Camera pushes in on .." will let the view move forward towards something

    "Camera pulls back..." lets the view pull back.

    "camera pans..." lets the view move left or right.

    "camera remains static on..."

    A simple example prompt for a portrait of a woman in a park as input image:

    "The woman is about to sit on the floor and shows a shy smile. the crowd in the back are watching her with curiosity. The camera is pushing in on her face. A short video clip from a TV show."

    As usual it will not work in 100% of cases, but will increase the chances for the requested result.

    Here is a good YT video related to camera control:

    https://youtu.be/z1srayjh-OE?si=D9P1G4mPhjecS5b5

    GFrostJun 18, 2025· 1 reaction

    Wow. Thanks. Much help!

    RoboticeyeJul 20, 2025

    What about the global vs the sequence prompts? I get that sequence is action, but how does that play into the global? Is the global like the full list of sequenced things that will happen?

    tremolo28
    Author
    Jul 20, 2025

    Roboticeye Global section shows relevant parameters of all sequences, like individual prompt, lora, colormatch, etc. So you can manage them from one place. Within the sequence you can set specific parameter per sequence, like length, steps, seed, ect. Regarding Prompt: You can either use one global prompt for all sequences or switch per sequence to an individual prompt. I.e. keep sequence 1-2 global, but make an individual prompt for sequence 3.

    Workflows
    Wan Video 14B i2v 480p

    Looks like we don't have an active mirror for this file right now.

    CivArchive is a community-maintained index — we catalog mirrors that volunteers upload to HuggingFace, torrents, and other public hosts. Looks like no one has uploaded a copy of this file yet.

    Some files do get recovered over time through contributions. If you're looking for this one, feel free to ask in Discord, or help preserve it if you have a copy.

    Details

    Downloads
    2,217
    Platform
    CivitAI
    Platform Status
    Deleted
    Created
    5/29/2025
    Updated
    4/21/2026
    Deleted
    4/16/2026

    Files

    wan21IMAGEToVIDEOWith_v30.zip

    Mirrors