CivArchive
    Wan i2v with upscaling, interpolating, and smoothing - v1.2
    NSFW

    Now faster and easier to install

    This workflow uses a small baseline generation using the 14B image to video model, followed by upscaling, and then smoothing out the result using the 5B model.

    This lets you test prompts and iterate quicker on the base generation before upscaling to a final resolution.

    Links for all the required models and where to put them are now included in the workflow.

    FAQ

    Do I need both Wan 2.1 and 2.2 VAEs?
    Yes. The 2.2 VAE only works with the 5b model (confusing, I know). Make sure the main section loads the 2.1 VAE, and the upscale section loads the 2.2 VAE.

    Its frozen on VAE decode

    The second vae decode can take a long time. Just be patient.

    Description

    Includes Triton and non-triton version

    Add WanVideoBlockSwap to optionally reduce vram (tested working under 12GB!)

    Reduced default denoise in vid2vid to 0.1 for more consistent details

    FAQ

    Comments (131)

    SaintScrewMay 8, 2025· 3 reactions
    CivitAI

    Hazard is the undisputed GOAT of i2v.

    HazardAI
    Author
    May 8, 2025· 1 reaction

    I'm taking this complement with me to my grave. ❤️

    whiper6993May 8, 2025· 2 reactions
    CivitAI

    Cool scheme, I also improved it with auto-cleaning VRAM and also very useful auto-save of the last frame

    Fox009May 8, 2025· 2 reactions

    This sounds great. Can you explain how you did the auto-cleaning of the RAM and the auto-save for the last frame?

    rnvvmwMay 9, 2025· 1 reaction

    @Fox009 I used a clumsy but effective method.
    Add an INT node and connect it to the following two nodes:
    1. The location in the original workflow where the length is manually modified
    2. Add "🔧 Image From Batch"(fm comfyui_essentials).

    The input of the "🔧 Image From Batch" node is connected to "RIFE VFI (I recommend rife47 and rife49)", and its output is connected to "saving the image".

    whiper6993May 9, 2025· 1 reaction

    @Fox009 I just found custom node "CLean VRAM used" from "comfyui-easy-use" and add this after VAE
    And last frame like @rnvvmw said yeah

    rnvvmwMay 9, 2025
    CivitAI

    Can different images be batch processed? I tried a few batch nodes, but they didn't work.

    rnvvmwMay 12, 2025

    Replace "Load & Resize Image" as these step below
    1. Add "Load Image Batch(was-node-suite-comfyui)" node
    2. Add "Int(comfy-core)" node and connect to index from "Load Image Batch", set value=0; increment
    3. Add "Resize Image v2(comfyui-kjnodes)" connect to image from "Load Image Batch", set width=480 ;height=10000 ;upscale=landczos ;proportion=resize ;divisible=16

    edoz1986206May 9, 2025
    CivitAI

    small mistake in the wkf

    blobby99May 9, 2025· 1 reaction
    CivitAI

    The only 'interesting' part of this basic workflow is the second pass using Wan 1.3B. Unfortunately this process kills any unique details in the original video, and doesn't improve the apparent resolution. If you like a plastic look, this is for you.

    DRZ3000May 9, 2025

    yes, there is a slight loss of detail, I am very sensitive to that, but I far from agree with that statement, there is a slight loss, but not that significant compared to the artifacts that are present on the original video. Maybe you need to try a different upscaler or lower the denoise of the second pass

    DRZ3000May 9, 2025· 1 reaction

    Anyway, while there is no magic, there is a way to overcome some problems using available methods, this method is one of the best I have seen, if there are any others, please share with us.

    HazardAI
    Author
    May 11, 2025· 1 reaction

    @blobby99 Yep, that is the only "interesting" thing I have going on here, other than that, its quite basic. I wasn't happy with the output from directly upscaling 480p videos, so this is my best attempt to try and improve them. If you have a good enough GPU (and enough patience) to handle direct 720p generation, then that's likely a better option. Other upscaling processes like Topaz also probably do a better job. But this works well enough for me for now.

    Best of luck finding something you like better. As @DRZ3000 mentioned, if you find something that produces better results in comparable time, I'd love to know about it too.

    iysam1234567773May 9, 2025
    CivitAI

    Hello. I want to try myself in video generation, but I'm not sure that my RTX 3070 with 8 gigabytes of video memory can handle it. Tell me, is there a place to upgrade to 4060 TI with 16 gigabytes of VRAM? Or should I try to do something on 8 gigabytes? Or are there other nice video cards for neural networks that don't require a lot of money?

    kennysladefan293May 9, 2025· 2 reactions
    CivitAI

    My usual workflow is using wan2.1-i2v-14b-720p-Q8_0.gguf which is 16.8GB in size despite the fact that I only have an RTX 4070 Super with 12GB VRAM and I can generate 81 frame 720x720 videos just fine in around 39 minutes (20 steps). Not sure what I am doing right? Just saying to people don't give up hope that you can generate 720p videos directly using the 720p GGUF using 12GB VRAM, even the Q8 version...

    KristirinaMay 9, 2025

    I have a 5090 but I'm using 14b_480p_Q8.gguf and generating at 480x832 (something around it, can't remember the exact number in my head) for 98+ frame in around 5-6min. 39min is just too long for a few second of video. You are better off making small resolution videos, pick the one you like, then upscale it with RIFT or Topaz AI, total time is under 10min.

    whiper6993May 9, 2025

    Just try q4 ver, it takes on my 4070 ti about 10 min for 81 frames

    kennysladefan293May 10, 2025

    @whiper6993 Yes I might try the Q4, although quality is my main priority so I went for the Q8

    kennysladefan293May 10, 2025

    @Kristirina Yes it would be interesting to try that method and compare the quality, thanks. By the way, I normally queue around 12 videos and let ComfyUI run overnight, then check out the 12 videos next morning (which have taken about 40 minutes each) so if I was generating videos in 20 minutes I'd have 24 videos to sort through next morning which would be too many LOL

    whiper6993May 11, 2025

    @kennysladefan293 And here is such a topic that you simply will not be able to use the large version normally, not only because of the speed, but because the memory on the processor itself will be overloaded

    KristirinaMay 9, 2025· 1 reaction
    CivitAI

    Nice! Combined this with 2pass sampler and the result is awesome! Also, I'm surprised the interpolation in Comfyui is that quick like it can be done in 5 sec for 97 frames :"D

    Light6969May 9, 2025

    Sounds good, do you have a json or something for that 2 pass sampler? I'm not sure how to do it lol

    bhassMay 9, 2025

    Upload this please!

    CoronelMay 10, 2025

    Please more info about this. Thank you!

    shura490May 10, 2025
    CivitAI

    Great work and results. I wonder if using skyreel models can helps us smooth out a longer video

    HazardAI
    Author
    May 11, 2025

    I'm currently experimenting with them, but I don't have anything worth sharing yet. I'm especially excited for the release of the 5B models, since those seem like they could be ideal for the upscaling process. Fingers crossed on a release soon.

    shura490May 11, 2025· 1 reaction

    @HazardAI Awesome, look forward to your future work. Cheers

    edoz1986206May 10, 2025
    CivitAI

    If you use a painting model to batch process image details, is it possible?

    HazardAI
    Author
    May 11, 2025

    I bet you could. I don't know how well the t2v model would do smoothing out the result, but I'd love to see someone test that. Could be extremely useful.

    FLOW0308May 14, 2025· 1 reaction

    @HazardAI You can take a look at all the works in my files. Most of them were completed using your methods, which are great. I'm just still researching if there's a better way for smoothing. Have you currently researched any new methods?

    HazardAI
    Author
    May 14, 2025

    @FLOW0308 I've experimented with the Skyreels models a bit, but so far haven't gotten any results that I thought were better than Wan t2v. I'm hoping the upcoming 5B 720p Skyreels model will be a good fit for the smoothing process.

    bakaboy1234May 10, 2025
    CivitAI

    whats new in 1.2?

    HazardAI
    Author
    May 11, 2025· 2 reactions

    Not very much actually. Just added the BlockSwap node which you can use to manually offload some layers and save vram at the expense of speed (that's mostly going to be useful for users who weren't able to run at all before)

    I also decreased the denoise down to 0.1, since it seems like it still does a pretty good job, and addresses a fairly common comment that the upscaling changes things too much from the original.

    bakaboy1234May 11, 2025

    @HazardAI nice thanks !

    HMythAIMay 10, 2025
    CivitAI

    somehow there is very less movement.. I have tried using prompt enhancer as well but very less movements. otherwise a great fast workflow..
    any recommendations?

    HazardAI
    Author
    May 11, 2025· 1 reaction

    You can try to increase the flow shift, but WAN seems pretty sensitive to that, so you'd only want to increase it by small steps. Prompting for more movement usually gets the job well enough for me, but I've had some source images where I couldn't get it to do what I wanted with several attempts.

    Lots of the posts here from me and others include metadata so you can see what prompts are working for people.

    GenEric69May 12, 2025· 1 reaction
    CivitAI

    Great workflow but what are my options to speed up the initial KSAMPLER? Even putting .3 in TeaCache its so slow. I have a 4090.

    HazardAI
    Author
    May 12, 2025

    Using the triton/torch compile version helps a bit. Using a smaller resolution or video length helps. Making sure that the initial i2v model is fully loaded into vram by checking the console logs.

    binauralhealing100139May 13, 2025· 2 reactions
    CivitAI

    i get this error with the triton workflow:
    KSampler

    backend='inductor' raised: CalledProcessError: Command '['C:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\runtime\\tcc\\tcc.exe', 'C:\\Users\\User\\AppData\\Local\\Temp\\tmp3b78a60i\\cuda_utils.c', '-O3', '-shared', '-Wno-psabi', '-o', 'C:\\Users\\User\\AppData\\Local\\Temp\\tmp3b78a60i\\cuda_utils.cp312-win_amd64.pyd', '-fPIC', '-lcuda', '-lpython3', '-LC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '-LC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\lib\\x64', '-IC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '-IC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '-IC:\\Users\\User\\AppData\\Local\\Temp\\tmp3b78a60i', '-IC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Include']' returned non-zero exit status 1. Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information You can suppress this exception and fall back to eager by setting: import torch._dynamo torch._dynamo.config.suppress_errors = True

    binauralhealing100139May 13, 2025
    CivitAI

    How fast would a generation on a rtx4080 be? It says 1h 20min xd

    HazardAI
    Author
    May 13, 2025

    It depends on the settings you use, but generations typically take 10 minutes for me on a 3090. You should be able to get comparable times on a 4080 as long as Wan is fully loaded to vram.

    binauralhealing100139May 13, 2025· 1 reaction

    @HazardAI how do i know if wan is fully loaded to vram?

    binauralhealing100139May 13, 2025
    CivitAI

    Is it normal that the teacache doesnt load fully? I still get the video.
    This is the console text:
    Attempting to release mmap (18)

    0%| | 0/25 [00:00<?, ?it/s]

    TeaCache: Initialized

    8%|██████▋ | 2/25 [01:18<14:57, 39.04s/it][ComfyUI-Manager] The ComfyRegistry cache update is still in progress, so an outdated cache is being used.

    FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json [DONE]

    FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json [DONE]

    FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extras.json [DONE]

    FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json [DONE]

    96%|██████████████████████████████████████████████████████████████████████████████▋ | 24/25 [07:44<00:16, 16.35s/it]-----------------------------------

    TeaCache skipped:

    13 cond steps

    13 uncond step

    out of 25 steps

    -----------------------------------

    100%|██████████████████████████████████████████████████████████████████████████████████| 25/25 [07:44<00:00, 18.57s/it]

    HazardAI
    Author
    May 14, 2025· 1 reaction

    That looks normal. As far as I understand, Teacache is designed to basically skip steps when there's small amounts of movement, so the "skipped" part of the output is it working as intended.

    binauralhealing100139May 14, 2025· 1 reaction

    @HazardAI okay thank you ❤️

    DollarStoreAbrahamMay 14, 2025· 1 reaction
    CivitAI

    Works great on my 3080 10GB. 👍
    Only final VAE decode doesn't fit completely and ends up using shared vram.

    BMTZZMay 18, 2025

    how long does it take in total on your 3080 10gb to output a video

    DollarStoreAbrahamMay 18, 2025

    @bigmantingzz depends on the total size and settings ofc.

    but for something like this https://civitai.com/images/76097547 it's around 15-20 mins (for 5-6 sec)
    I usually test prompts without upscaling and interpolation, so it goes to just around 5-8 mins; and then turn them on once I'm with the result.

    BMTZZMay 18, 2025

    @DollarStoreAbraham thanks thats a great time for 5-6 seconds im gonna try it out !

    offartAIMay 14, 2025
    CivitAI

    Both workflows ask me for triton, why?

    HazardAI
    Author
    May 14, 2025

    If you run the Triton workflow first, you'll need to restart ComfyUI to use the non-triton version.

    kurtast88942May 14, 2025
    CivitAI

    I'm loving this workflow! I'm having one repeating issue though, every couple generations i get this error: "output with shape [1, 51975, 1536] doesn't match the broadcast shape [2, 51975, 1536]" during the ksampler in the 13b section. Any idea why?

    jasonafexMay 17, 2025

    I get the same error "output with shape [1, 45900, 1536] doesn't match the broadcast shape [2, 45900, 1536]" half the time

    DelavestraJun 19, 2025

    @jasonafex Any resolution here?

    jasonafexJun 20, 2025

    @Delavestra I believe I ended up switching to a different model/workflow

    Aurora_xoJun 23, 2025

    you might be getting that error if you're trying to generate with a landscaped image instead of a 1x1 or portrait image, to fix it just switch the Width and Height in the MAIN - Load & Resize Image node, and do the same for the PREP VIDEO - Resize Image V2 node.

    MagisterRayMay 14, 2025· 6 reactions
    CivitAI

    Thank you so much for adding that 'Models Needed' section. As someone just starting out, I have had so many headaches trying to get workflows to, well, work, because I have no idea what half of the errors mean, or where the various files are even supposed to go.

    So yeah, that has been super helpful, thank you!

    HazardAI
    Author
    May 14, 2025· 1 reaction

    If there's anything I've missed, the ComfyUI docs for Wan also has some great information. I refer to it very often.
    https://docs.comfy.org/tutorials/video/wan/wan-video

    ZojixMay 15, 2025
    CivitAI

    hi there, getting mat1 and mat2 shapes cannot be multiplied (77x768 and 4096x5120) , any solution ?

    nojojos2234May 15, 2025
    CivitAI

    This benefits from sageattention / triton, right? Btw, On 12GB vram, I had to use Q3_M quants with offload layers > 0 and reduce the resolution to 3xx ish x 5xx ish. The default settings for this workflow didn't fit in 12GB vram for me.

    HazardAI
    Author
    May 16, 2025

    Triton definitely. I'm less sure about Sage Attention. I don't know if the ComfyUI built in KSampler utilizes Sage Attention when running WAN. There is a node from Kijai which patches in Sage Attention, but I was unable to get it to work with the other performance optimizations. Of the combinations of performance optimizations I could get to work, the current set gave me the best times. The developers of ComfyUI, and Kijai move very quickly though, so additional tests to see if I can get times faster are definitely on my list.

    vgbestlyMay 16, 2025· 1 reaction
    CivitAI

    I was using your v1.0 no torch compile work flow for the past months. Just recently switched to v1.2. I am using a RTX 2070 Super 8GB to generate these pictures. I takes about, I wanna say, ~2 hours for the first pic and ~1.5 hours for any after. Yes, there is a timer in seconds after the job is finished but I checked once or twice and it was off by 30mins sometimes. And I mostly run a generation, sleep and post in the morning anyways so I'm not worried.

    I just wanna say THANK YOU. Very easy to use. Mostly used default settings that came with workflow. Just follow the instructions in the page to install. SOME hassle at the beginning trying to install due to broken or missing link like a month ago so I had to hunt down one of the thing missing manually but now its fixed.

    vgbestlyMay 16, 2025

    Additional notes for readers: I am using the Q3_K_M. Instead of the original Wan 2.1 link in page, I use ComfyORG i2v-480p GGUF. I used this since it links up with Civitai. Like when you try to "add a resource used in this gen." Civitai asks before posting.

    Additional notes for OP: Also, when you DL v1.2, the zip file gives you two files (one with triton, one with no triton). I'm a some what beginner to ComfyUI and I have no idea how to install this triton thing with pip or python or whatever code wizard thing it wants me to do...so I just use the wan2-upscale-no-triton-v1-2.json file for the workflow. Maybe add a "how to install" this triton thing for clueless people like me if you want people to use it.

    Edits: spelling and more notes.

    HazardAI
    Author
    May 16, 2025· 1 reaction

    @vgbestly Thank you so much for the kind words! It always makes me feel great hearing that people are finding this useful.
    Adding Triton install instructions is a great idea. Personally I use Linux, which thankfully makes it much much easier to install. But I could definitely include a link to installation instructions for Windows.

    sayntegeorgeMay 17, 2025· 1 reaction
    CivitAI

    Does anyone know how to enable the RfileXRope node?

    6028976May 21, 2025

    Maybe it is ComfyUI standard node ? I remember it wasn't working on a previous try (so I just disabled it) but since I updated ComfyUi and it seems to work now (but I still disabled it I think)

    slaad0May 22, 2025

    Right click it and then click Bypass, or CTRL+B

    iysam1234567773May 17, 2025· 1 reaction
    CivitAI

    this workflow is just great. it worked on rtx 3070 with 8gb but i had to wait for an hour. now i have 4060 ti with 16gb and i wait for 30 minutes for one render. please tell me why i am losing details from the face, it is distorted. maybe the problem is in q3 model? or i want something wrong?

    HazardAI
    Author
    May 17, 2025· 1 reaction

    If you're losing face details in the vid2vid portion, you can lower the denoise strength. If you're losing them in the initial 480p generation, then you'd need to use an image where the face is larger in frame, or use a slightly larger image.

    maverick7508May 18, 2025· 2 reactions
    CivitAI

    I am really new to this but I am pretty sure i got the required files downloaded and put into the right places, but the workflow keeps failing. is there a recommended walkthrough/tutorial for this type of workflow in ComfyUI?

    6028976May 21, 2025

    Make sure when you click on a model (inside the nodes) they allow you to select it (meaning the model indeed exist at the right place in your folders) sometimes it could just list the model name while you actually don't have it loaded or not at the right place

    Dimmed8145May 18, 2025
    CivitAI

    Hi everyone, I’m using this workflow with a 4090 with 24GB of VRAM. Using the same models and configuration. I’m not sure why it keeps encountering out of memory issues.

    I’ve redownloaded comfy a few times and deployed on a few different GPUs on vast ai but it still encounters allocation memory issues. Any advice is greatly appreciated!

    Woolyskink160May 18, 2025

    Hi, you can try using the flag: --lowvram, or even --medvram, worked for me to counter the mentioned issue. Hope it helps!

    cremedelacremeMay 21, 2025

    @Woolyskink160 hi, I'm having the same issue as well! by using the medvram (lowvram) will this increase the time it takes to load the entire WF?

    cremedelacremeMay 21, 2025

    @Woolyskink160 hi, is it possible to have these VRAM issues on a 5090? I'm using Runpod

    DollarStoreAbrahamMay 23, 2025

    I'm using the Q3 versions of the models and a 5sec clip (81 raw frame) fits in my 10GB 3080.
    Could give that a try, idk about the more VRAM hungry models tho

    jintrepit147May 18, 2025· 3 reactions
    CivitAI

    YOU ARE MY HEROOOOOOOOOOOOOOOOOOOOO!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!

    AliveAI_appMay 21, 2025· 1 reaction
    CivitAI

    This workflow is nice. Any chance to update it to use the speed improvement lora https://civitai.com/models/1585622

    I tried adding the lora to the current workflow but the effect is not good. not sure what the issue is

    tlzborowik830May 21, 2025
    CivitAI

    What kind of nodes do you use?

    tlzborowik830May 21, 2025
    CivitAI

    SOMEONE CAN HELP ME PLEAS????
    CLIPTextEncode

    mat1 and mat2 shapes cannot be multiplied (77x768 and 3072x768)

    slaad0May 22, 2025

    This is often an issue with the model you've chosen, are you using the recommended models (Wan, CLIP, VAE, etc?)

    slaad0May 22, 2025· 1 reaction
    CivitAI

    Wow, what an incredible workflow. After coming from the last one that produced a lot of garbage, this one is regularly producing good results.

    jespo49298May 23, 2025
    CivitAI

    Where can I install all the custom nodes? I used the node manager but I am still missing some

    6028976May 28, 2025

    it is likely your comfyUi is just outdated, many of the nodes seems native, just update to or reinstall last version

    lavasplit753May 29, 2025

    @fouchardmilcoupes311 Possibly. I'm running into the same issue where the rgthree nodes just refuse to install and comfy is up to date.

    fl4un8242May 23, 2025
    CivitAI

    I followed along but i get this error when trying to Run it:

    CLIPLoaderGGUF

    Mixing scaled FP8 with GGUF is not supported! Use regular CLIP loader or switch model(s) (/workspace/ComfyUI/models/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors)

    can someone help resolve this

    redsnailMay 24, 2025· 1 reaction

    need umt5_xxl_fp16.safetensors instead

    fl4un8242May 24, 2025

    @redsnail got yah. thanks!

    fronyaxMay 25, 2025· 1 reaction
    CivitAI

    Your workflow worked flawlessly with CauseVid 1.3b lora. I used it with 8 steps, and dont use tea cache and skip layer guidance node if using the lora. Check gallery for the result.

    AliveAI_appMay 25, 2025

    for me it does not work well at all? what are you ksampler settings? did you change the shift too?

    fronyaxMay 26, 2025

    @testay I just set the CFG to 1, steps to 8, Euler, simple, and removed the teacache and SLG node.

    I didn’t change the shift, I only set it to 5

    6028976May 26, 2025

    Interesting, so removing Teacache and SLG for causvid, that might explain why my result were retty bad when I tried, i'll give it a try later without teache and SLG

    gman_umschtMay 27, 2025

    @fouchardmilcoupes311 Basically Teacache and CausVid are punching each other in the face when they are used together in the workflow from what I gathered.

    HazardAI
    Author
    May 28, 2025

    @fronyax Thank you for the suggestion! I wasn't aware of CauseVid until your comment. So far in my testing I'm only getting a fairly small speed improvement using CauseVid for the first sampling. CauseVid with 15 steps is about 15% faster than TeaCache with 30 steps. I don't know if I can say that the quality is "worse" but definitely different. I see that KijAI recommends using two ksamplers with CauseVid, but I'm not entirely clear on how exactly that should be set up. (Kij is also using a kl_optimal for the scheduler, but I also couldn't get any coherent results out with that)
    More testing is definitely needed, but its an exciting advancement.

    fronyaxMay 29, 2025

    @HazardAI Yes, more testing is definitely needed. I haven’t had any success using your V2V setup using 2 samplers with causevid tho, it always outputs a black image, idk why. That said, for video generation using 2 samplers is generally recommended for improved motion.

    As for me, the sweet spot with causevid and your V2V is around 6–8 steps. I have very limited VRAM, so this setup lets me refine a video without waiting for hour. It only takes about 13–15 minutes to refine a 5 second, 1280x720 video.

    Maybe you can make another workflow version with causevid.

    gman_umschtMay 29, 2025· 1 reaction

    @HazardAI I've gotten decent results with CausVid as low as 5 steps @ 0.5. For higher quality I went to 8-10 steps with lower lora strength. If using LORAs I usually set the cfg to 2 instead of 1, but still tweaking and experimenting. I also got the suggestion for a 2-sampler approach over at @RalFinger Discord. In my first simple flow I use no CausVid with cfg 6 for the 1st sampler for prompt adherence and CausVid with cfg 1 for the 2nd sampler to refine the output. Even a 4+4 or 3+3 step approach was pretty good. But this is also still in the experiment phase.

    HazardAI
    Author
    May 29, 2025

    @fronyax Oh wow, I didn't realize you could go as low as 6 steps. (I pulled 15 from the Kijai example workflow). That does pretty drastically reduce the time. I just got a 61 frame, 8 steps, 480x688 render in 106 seconds vs 287 seconds for 24 steps with TeaCache. I'll definitely play with the multiple KSamplers some more. V1.3 will very likely include a CauseVid version, or at least instructions for what to bypass to use it.

    6028976May 29, 2025

    @HazardAI But do you get cartoon oversaturated (almost cell shading effect) with CausVid ? because that is what happen to me when I use it, sure it's super fast but too low quality a loss of realism, maybe I'm doing something wrong

    gman_umschtMay 29, 2025· 1 reaction

    @fouchardmilcoupes311 Noticed that too, although mostly when I feed the last frame into the next clip and stitch them together. After 3-4 clips it is rather noticeable. But this latest WAN workflow over at @tremolo28 sounds like a solution: "It contains a Color Match option in postprocessing to cope with the increased saturation, the lora is introducing."

    HazardAI
    Author
    May 29, 2025· 1 reaction

    @fouchardmilcoupes311 I definitely got that on my first few tests using 15 steps. With 8 steps, CFG 1, and Euler Simpler, I'm getting visually comparable results to 24 steps with TeaCache. At least in the few source images I've tried so far. Wan does seem somewhat picky about settings when switching between 2d and photo real.

    6028976May 27, 2025
    CivitAI

    I don't understand something, Wan2.1 14b I2V at about 768 resolution (for longuest side) at around 81 frames, on some GGUF variant take about 16 to 21 GB of Vram on my system, then when the T2V 1.3b model take over for the refine smooth pass it take up to 22+ GB Vram I don't get it, are the other model not removed from ram or something ? Is this normal and expected ? (it work in the end, but I don't understand the high memory of the second pass with a vastly lighter model

    HazardAI
    Author
    May 27, 2025

    You are correct that the initial model is not unloaded. ComfyUI seems to struggle with auto vram management with Wan models at the moment. I haven't yet found a good solution for automatically handling unloading the initial i2v model. (V1.0 of the workflow had a node to try and get it to unload, but it didn't seem to work.) If you want to use the least vram possible, you can use the group muter to only run the first step, then manually unload all models with ComfyUI Manager, and then enable and run the upscaling and refining. It's a bit tedious, but does work.

    6028976May 27, 2025

    @HazardAI Yeah that might be a good plan B

    6028976May 28, 2025

    @HazardAI I put an unload all model after each VAD decode (I mean those after Ksampler of course) seems to work, or at least doesn't seems to bother my PC, tested with Wan 2.1 14B t2v for the refiner too (in gguf format though, seems better quite long though)

    AiartsenpaiMay 28, 2025
    CivitAI

    For Unet Loader (GGUF) i am using

    wan2.1-i2v-14b-480p-Q8_0.gguf

    I am confused. because it wont work.

    Everything loads just fine but it stops at:

    My CLIPLoader (GGUF)

    it says: CLIPLoaderGGUF

    invalid tokenizer

    When i use umt5-xxl-encoder-Q8_0.gguf

    What text encoder am i supposed to use? Or is my unet loader much for my pc?

    I have the latest comfyui and using pinokio

    6028976May 28, 2025

    type = wan is selected on the clip loader node ? It work for me with Q4 to Q8 gguf version

    6028976May 28, 2025

    If all fail, maybe making sure you are using latest comfuUi might solve most of the problems (The number of time just upgrading/reinstalling comfyUI to latest solved problem for me is countless time)

    AiartsenpaiMay 28, 2025

    @fouchardmilcoupes311 i did the latest update for comfyui and it kinda crashed everything

    6028976May 28, 2025

    @Aiartsenpai you ran the *.bat in the update folder

    update_comfyui_and_python_dependencies.bat

    this is the real update I'm talking about, if nothing work well maybe make just another install and start fresh, it can't be that bad if it worked before.

    AiartsenpaiMay 31, 2025

    @fouchardmilcoupes311 oh i figured it but ty.

    fl4un8242May 30, 2025
    CivitAI

    some reason my last frame glitches / blurs / over saturates. I have set the frame length to 81. Before this wasn't an issue. Maybe i accidentally tinkered with a setting and forgot about. Does anyone know why this may be happening?

    DollarStoreAbrahamMay 30, 2025· 5 reactions
    CivitAI

    I would love to see a variant of this workflow with First-Last frame, so you can supply two images where it should start and end up

    AiartsenpaiMay 31, 2025
    CivitAI

    Help i have 64 ram but not sure how to use it all. Anyone know a good smooth 720 workflow or 480 that is at least somewhat fast? I don't want to have a workflow that takes 2-4 hours per video.

    LoraForgeJun 5, 2025

    64 ram or vram? What GPU do you have?

    AiartsenpaiMay 31, 2025
    CivitAI

    also where do i download RIFE VFI (recommend rife47 and rife49)? its an upscaler but not sure where to find the file to download it

    HazardAI
    Author
    May 31, 2025

    That one should be handled automatically by the node when you run the workflow.

    AiartsenpaiJun 2, 2025

    @HazardAI oh i figured it out. It was a bit buggy. updating fixed it

    AiartsenpaiJun 2, 2025· 1 reaction

    @HazardAI ty for helping out though

    jackolf2May 31, 2025
    CivitAI

    I get unetloadergguf and cliploadergguf not found, any idea what I'm doing wrong?

    6028976Jun 1, 2025

    you probably have an outdated 'obsolete' comfyUI version, by not found you mean the nodes ar enot there at all or circled in red like unknown ?

    oo4766229Jun 2, 2025
    CivitAI

    My GPU is 5070ti with 16GB VRAM

    I set the setting exactly like you suggested with triton, didn't alter other configuration. Yet the time to complete a 5 sec video is over 2 hours! Any idea why?

    6028976Jun 2, 2025

    check your memory usage, if you see 98-99M of your V-ram and 98-99% of your system ram used, most likely you are dealing with the program trying to deal with memory problem, it might still complete but you are just not having enough memory, so reduce settings, or make sure you don't run other program at the same time eating your memory, I doubt you will ever be able to make a video in 5 seconds, but 2 hours seems way to high

    6028976Jun 2, 2025

    and by the way, try 4 seconds video, it will significantly reduce any memory usage, maybe the problem is just 5 seconds lenght is a touch too much for your ram

    oo4766229Jun 3, 2025

    I finally fixed it! It's really the issue of V-ram problem. I tried a node which free up V-ram when loading model, now I can create a 5 sec vid in 5 minutes. Thanks!

    6028976Jun 3, 2025

    @oo4766229 cool, yeah it was really looking like that sort of problem

    oo4766229Jun 3, 2025

    @fouchardmilcoupes311 Wondering how that happened though. I've seen people with V-ram less than mine maked videos in a reasonable amount of time.

    davifpintoJun 4, 2025

    How did you fixed it? I have 12gb VRAm and 16gb RAM

    oo4766229Jun 4, 2025

    @davifpinto I used this node to load checkpoint https://github.com/pollockjj/ComfyUI-MultiGPU

    davifpintoJun 5, 2025

    @oo4766229 Thanks!!

    davifpintoJun 12, 2025

    @oo4766229 You figured out some way to use multigpu with i2v?

    MaddyBellJun 3, 2025· 1 reaction
    CivitAI

    So good!

    Workflows
    Wan Video 14B i2v 480p

    Details

    Downloads
    6,250
    Platform
    CivitAI
    Platform Status
    Available
    Created
    5/7/2025
    Updated
    5/12/2026
    Deleted
    -

    Files

    wanI2vWithUpscaling_v12.zip

    Mirrors

    HuggingFace (1 mirrors)
    CivitAI (1 mirrors)