Now faster and easier to install
This workflow uses a small baseline generation using the 14B image to video model, followed by upscaling, and then smoothing out the result using the 5B model.
This lets you test prompts and iterate quicker on the base generation before upscaling to a final resolution.
Links for all the required models and where to put them are now included in the workflow.
FAQ
Do I need both Wan 2.1 and 2.2 VAEs?
Yes. The 2.2 VAE only works with the 5b model (confusing, I know). Make sure the main section loads the 2.1 VAE, and the upscale section loads the 2.2 VAE.
Its frozen on VAE decode
The second vae decode can take a long time. Just be patient.
Description
Includes Triton and non-triton version
Add WanVideoBlockSwap to optionally reduce vram (tested working under 12GB!)
Reduced default denoise in vid2vid to 0.1 for more consistent details
FAQ
Comments (131)
Hazard is the undisputed GOAT of i2v.
I'm taking this complement with me to my grave. ❤️
Cool scheme, I also improved it with auto-cleaning VRAM and also very useful auto-save of the last frame
This sounds great. Can you explain how you did the auto-cleaning of the RAM and the auto-save for the last frame?
@Fox009 I used a clumsy but effective method.
Add an INT node and connect it to the following two nodes:
1. The location in the original workflow where the length is manually modified
2. Add "🔧 Image From Batch"(fm comfyui_essentials).
The input of the "🔧 Image From Batch" node is connected to "RIFE VFI (I recommend rife47 and rife49)", and its output is connected to "saving the image".
@Fox009 I just found custom node "CLean VRAM used" from "comfyui-easy-use" and add this after VAE
And last frame like @rnvvmw said yeah
Can different images be batch processed? I tried a few batch nodes, but they didn't work.
Replace "Load & Resize Image" as these step below
1. Add "Load Image Batch(was-node-suite-comfyui)" node
2. Add "Int(comfy-core)" node and connect to index from "Load Image Batch", set value=0; increment
3. Add "Resize Image v2(comfyui-kjnodes)" connect to image from "Load Image Batch", set width=480 ;height=10000 ;upscale=landczos ;proportion=resize ;divisible=16
small mistake in the wkf
The only 'interesting' part of this basic workflow is the second pass using Wan 1.3B. Unfortunately this process kills any unique details in the original video, and doesn't improve the apparent resolution. If you like a plastic look, this is for you.
yes, there is a slight loss of detail, I am very sensitive to that, but I far from agree with that statement, there is a slight loss, but not that significant compared to the artifacts that are present on the original video. Maybe you need to try a different upscaler or lower the denoise of the second pass
Anyway, while there is no magic, there is a way to overcome some problems using available methods, this method is one of the best I have seen, if there are any others, please share with us.
@blobby99 Yep, that is the only "interesting" thing I have going on here, other than that, its quite basic. I wasn't happy with the output from directly upscaling 480p videos, so this is my best attempt to try and improve them. If you have a good enough GPU (and enough patience) to handle direct 720p generation, then that's likely a better option. Other upscaling processes like Topaz also probably do a better job. But this works well enough for me for now.
Best of luck finding something you like better. As @DRZ3000 mentioned, if you find something that produces better results in comparable time, I'd love to know about it too.
Hello. I want to try myself in video generation, but I'm not sure that my RTX 3070 with 8 gigabytes of video memory can handle it. Tell me, is there a place to upgrade to 4060 TI with 16 gigabytes of VRAM? Or should I try to do something on 8 gigabytes? Or are there other nice video cards for neural networks that don't require a lot of money?
My usual workflow is using wan2.1-i2v-14b-720p-Q8_0.gguf which is 16.8GB in size despite the fact that I only have an RTX 4070 Super with 12GB VRAM and I can generate 81 frame 720x720 videos just fine in around 39 minutes (20 steps). Not sure what I am doing right? Just saying to people don't give up hope that you can generate 720p videos directly using the 720p GGUF using 12GB VRAM, even the Q8 version...
I have a 5090 but I'm using 14b_480p_Q8.gguf and generating at 480x832 (something around it, can't remember the exact number in my head) for 98+ frame in around 5-6min. 39min is just too long for a few second of video. You are better off making small resolution videos, pick the one you like, then upscale it with RIFT or Topaz AI, total time is under 10min.
Just try q4 ver, it takes on my 4070 ti about 10 min for 81 frames
@whiper6993 Yes I might try the Q4, although quality is my main priority so I went for the Q8
@Kristirina Yes it would be interesting to try that method and compare the quality, thanks. By the way, I normally queue around 12 videos and let ComfyUI run overnight, then check out the 12 videos next morning (which have taken about 40 minutes each) so if I was generating videos in 20 minutes I'd have 24 videos to sort through next morning which would be too many LOL
@kennysladefan293 And here is such a topic that you simply will not be able to use the large version normally, not only because of the speed, but because the memory on the processor itself will be overloaded
Nice! Combined this with 2pass sampler and the result is awesome! Also, I'm surprised the interpolation in Comfyui is that quick like it can be done in 5 sec for 97 frames :"D
Great work and results. I wonder if using skyreel models can helps us smooth out a longer video
I'm currently experimenting with them, but I don't have anything worth sharing yet. I'm especially excited for the release of the 5B models, since those seem like they could be ideal for the upscaling process. Fingers crossed on a release soon.
@HazardAI Awesome, look forward to your future work. Cheers
If you use a painting model to batch process image details, is it possible?
I bet you could. I don't know how well the t2v model would do smoothing out the result, but I'd love to see someone test that. Could be extremely useful.
@HazardAI You can take a look at all the works in my files. Most of them were completed using your methods, which are great. I'm just still researching if there's a better way for smoothing. Have you currently researched any new methods?
@FLOW0308 I've experimented with the Skyreels models a bit, but so far haven't gotten any results that I thought were better than Wan t2v. I'm hoping the upcoming 5B 720p Skyreels model will be a good fit for the smoothing process.
whats new in 1.2?
Not very much actually. Just added the BlockSwap node which you can use to manually offload some layers and save vram at the expense of speed (that's mostly going to be useful for users who weren't able to run at all before)
I also decreased the denoise down to 0.1, since it seems like it still does a pretty good job, and addresses a fairly common comment that the upscaling changes things too much from the original.
@HazardAI nice thanks !
somehow there is very less movement.. I have tried using prompt enhancer as well but very less movements. otherwise a great fast workflow..
any recommendations?
You can try to increase the flow shift, but WAN seems pretty sensitive to that, so you'd only want to increase it by small steps. Prompting for more movement usually gets the job well enough for me, but I've had some source images where I couldn't get it to do what I wanted with several attempts.
Lots of the posts here from me and others include metadata so you can see what prompts are working for people.
Great workflow but what are my options to speed up the initial KSAMPLER? Even putting .3 in TeaCache its so slow. I have a 4090.
Using the triton/torch compile version helps a bit. Using a smaller resolution or video length helps. Making sure that the initial i2v model is fully loaded into vram by checking the console logs.
i get this error with the triton workflow:
KSampler
backend='inductor' raised: CalledProcessError: Command '['C:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\runtime\\tcc\\tcc.exe', 'C:\\Users\\User\\AppData\\Local\\Temp\\tmp3b78a60i\\cuda_utils.c', '-O3', '-shared', '-Wno-psabi', '-o', 'C:\\Users\\User\\AppData\\Local\\Temp\\tmp3b78a60i\\cuda_utils.cp312-win_amd64.pyd', '-fPIC', '-lcuda', '-lpython3', '-LC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '-LC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\lib\\x64', '-IC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '-IC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '-IC:\\Users\\User\\AppData\\Local\\Temp\\tmp3b78a60i', '-IC:\\Users\\User\\Desktop\\Comfyui\\ComfyUI_windows_portable\\python_embeded\\Include']' returned non-zero exit status 1. Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information You can suppress this exception and fall back to eager by setting: import torch._dynamo torch._dynamo.config.suppress_errors = True
How fast would a generation on a rtx4080 be? It says 1h 20min xd
It depends on the settings you use, but generations typically take 10 minutes for me on a 3090. You should be able to get comparable times on a 4080 as long as Wan is fully loaded to vram.
@HazardAI how do i know if wan is fully loaded to vram?
Is it normal that the teacache doesnt load fully? I still get the video.
This is the console text:
Attempting to release mmap (18)
0%| | 0/25 [00:00<?, ?it/s]
TeaCache: Initialized
8%|██████▋ | 2/25 [01:18<14:57, 39.04s/it][ComfyUI-Manager] The ComfyRegistry cache update is still in progress, so an outdated cache is being used.
FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json [DONE]
FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json [DONE]
FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extras.json [DONE]
FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json [DONE]
96%|██████████████████████████████████████████████████████████████████████████████▋ | 24/25 [07:44<00:16, 16.35s/it]-----------------------------------
TeaCache skipped:
13 cond steps
13 uncond step
out of 25 steps
-----------------------------------
100%|██████████████████████████████████████████████████████████████████████████████████| 25/25 [07:44<00:00, 18.57s/it]
That looks normal. As far as I understand, Teacache is designed to basically skip steps when there's small amounts of movement, so the "skipped" part of the output is it working as intended.
@HazardAI okay thank you ❤️
Works great on my 3080 10GB. 👍
Only final VAE decode doesn't fit completely and ends up using shared vram.
how long does it take in total on your 3080 10gb to output a video
@bigmantingzz depends on the total size and settings ofc.
but for something like this https://civitai.com/images/76097547 it's around 15-20 mins (for 5-6 sec)
I usually test prompts without upscaling and interpolation, so it goes to just around 5-8 mins; and then turn them on once I'm with the result.
@DollarStoreAbraham thanks thats a great time for 5-6 seconds im gonna try it out !
Both workflows ask me for triton, why?
If you run the Triton workflow first, you'll need to restart ComfyUI to use the non-triton version.
I'm loving this workflow! I'm having one repeating issue though, every couple generations i get this error: "output with shape [1, 51975, 1536] doesn't match the broadcast shape [2, 51975, 1536]" during the ksampler in the 13b section. Any idea why?
I get the same error "output with shape [1, 45900, 1536] doesn't match the broadcast shape [2, 45900, 1536]" half the time
@jasonafex Any resolution here?
@Delavestra I believe I ended up switching to a different model/workflow
you might be getting that error if you're trying to generate with a landscaped image instead of a 1x1 or portrait image, to fix it just switch the Width and Height in the MAIN - Load & Resize Image node, and do the same for the PREP VIDEO - Resize Image V2 node.
Thank you so much for adding that 'Models Needed' section. As someone just starting out, I have had so many headaches trying to get workflows to, well, work, because I have no idea what half of the errors mean, or where the various files are even supposed to go.
So yeah, that has been super helpful, thank you!
If there's anything I've missed, the ComfyUI docs for Wan also has some great information. I refer to it very often.
https://docs.comfy.org/tutorials/video/wan/wan-video
hi there, getting mat1 and mat2 shapes cannot be multiplied (77x768 and 4096x5120) , any solution ?
This benefits from sageattention / triton, right? Btw, On 12GB vram, I had to use Q3_M quants with offload layers > 0 and reduce the resolution to 3xx ish x 5xx ish. The default settings for this workflow didn't fit in 12GB vram for me.
Triton definitely. I'm less sure about Sage Attention. I don't know if the ComfyUI built in KSampler utilizes Sage Attention when running WAN. There is a node from Kijai which patches in Sage Attention, but I was unable to get it to work with the other performance optimizations. Of the combinations of performance optimizations I could get to work, the current set gave me the best times. The developers of ComfyUI, and Kijai move very quickly though, so additional tests to see if I can get times faster are definitely on my list.
I was using your v1.0 no torch compile work flow for the past months. Just recently switched to v1.2. I am using a RTX 2070 Super 8GB to generate these pictures. I takes about, I wanna say, ~2 hours for the first pic and ~1.5 hours for any after. Yes, there is a timer in seconds after the job is finished but I checked once or twice and it was off by 30mins sometimes. And I mostly run a generation, sleep and post in the morning anyways so I'm not worried.
I just wanna say THANK YOU. Very easy to use. Mostly used default settings that came with workflow. Just follow the instructions in the page to install. SOME hassle at the beginning trying to install due to broken or missing link like a month ago so I had to hunt down one of the thing missing manually but now its fixed.
Additional notes for readers: I am using the Q3_K_M. Instead of the original Wan 2.1 link in page, I use ComfyORG i2v-480p GGUF. I used this since it links up with Civitai. Like when you try to "add a resource used in this gen." Civitai asks before posting.
Additional notes for OP: Also, when you DL v1.2, the zip file gives you two files (one with triton, one with no triton). I'm a some what beginner to ComfyUI and I have no idea how to install this triton thing with pip or python or whatever code wizard thing it wants me to do...so I just use the wan2-upscale-no-triton-v1-2.json file for the workflow. Maybe add a "how to install" this triton thing for clueless people like me if you want people to use it.
Edits: spelling and more notes.
@vgbestly Thank you so much for the kind words! It always makes me feel great hearing that people are finding this useful.
Adding Triton install instructions is a great idea. Personally I use Linux, which thankfully makes it much much easier to install. But I could definitely include a link to installation instructions for Windows.
Does anyone know how to enable the RfileXRope node?
Maybe it is ComfyUI standard node ? I remember it wasn't working on a previous try (so I just disabled it) but since I updated ComfyUi and it seems to work now (but I still disabled it I think)
Right click it and then click Bypass, or CTRL+B
this workflow is just great. it worked on rtx 3070 with 8gb but i had to wait for an hour. now i have 4060 ti with 16gb and i wait for 30 minutes for one render. please tell me why i am losing details from the face, it is distorted. maybe the problem is in q3 model? or i want something wrong?
If you're losing face details in the vid2vid portion, you can lower the denoise strength. If you're losing them in the initial 480p generation, then you'd need to use an image where the face is larger in frame, or use a slightly larger image.
I am really new to this but I am pretty sure i got the required files downloaded and put into the right places, but the workflow keeps failing. is there a recommended walkthrough/tutorial for this type of workflow in ComfyUI?
Make sure when you click on a model (inside the nodes) they allow you to select it (meaning the model indeed exist at the right place in your folders) sometimes it could just list the model name while you actually don't have it loaded or not at the right place
Hi everyone, I’m using this workflow with a 4090 with 24GB of VRAM. Using the same models and configuration. I’m not sure why it keeps encountering out of memory issues.
I’ve redownloaded comfy a few times and deployed on a few different GPUs on vast ai but it still encounters allocation memory issues. Any advice is greatly appreciated!
Hi, you can try using the flag: --lowvram, or even --medvram, worked for me to counter the mentioned issue. Hope it helps!
@Woolyskink160 hi, I'm having the same issue as well! by using the medvram (lowvram) will this increase the time it takes to load the entire WF?
@Woolyskink160 hi, is it possible to have these VRAM issues on a 5090? I'm using Runpod
I'm using the Q3 versions of the models and a 5sec clip (81 raw frame) fits in my 10GB 3080.
Could give that a try, idk about the more VRAM hungry models tho
YOU ARE MY HEROOOOOOOOOOOOOOOOOOOOO!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
This workflow is nice. Any chance to update it to use the speed improvement lora https://civitai.com/models/1585622
I tried adding the lora to the current workflow but the effect is not good. not sure what the issue is
What kind of nodes do you use?
SOMEONE CAN HELP ME PLEAS????
CLIPTextEncode
mat1 and mat2 shapes cannot be multiplied (77x768 and 3072x768)
This is often an issue with the model you've chosen, are you using the recommended models (Wan, CLIP, VAE, etc?)
Wow, what an incredible workflow. After coming from the last one that produced a lot of garbage, this one is regularly producing good results.
Where can I install all the custom nodes? I used the node manager but I am still missing some
it is likely your comfyUi is just outdated, many of the nodes seems native, just update to or reinstall last version
@fouchardmilcoupes311 Possibly. I'm running into the same issue where the rgthree nodes just refuse to install and comfy is up to date.
I followed along but i get this error when trying to Run it:
CLIPLoaderGGUF
Mixing scaled FP8 with GGUF is not supported! Use regular CLIP loader or switch model(s) (/workspace/ComfyUI/models/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors)
can someone help resolve this
Your workflow worked flawlessly with CauseVid 1.3b lora. I used it with 8 steps, and dont use tea cache and skip layer guidance node if using the lora. Check gallery for the result.
for me it does not work well at all? what are you ksampler settings? did you change the shift too?
@testay I just set the CFG to 1, steps to 8, Euler, simple, and removed the teacache and SLG node.
I didn’t change the shift, I only set it to 5
Interesting, so removing Teacache and SLG for causvid, that might explain why my result were retty bad when I tried, i'll give it a try later without teache and SLG
@fouchardmilcoupes311 Basically Teacache and CausVid are punching each other in the face when they are used together in the workflow from what I gathered.
@fronyax Thank you for the suggestion! I wasn't aware of CauseVid until your comment. So far in my testing I'm only getting a fairly small speed improvement using CauseVid for the first sampling. CauseVid with 15 steps is about 15% faster than TeaCache with 30 steps. I don't know if I can say that the quality is "worse" but definitely different. I see that KijAI recommends using two ksamplers with CauseVid, but I'm not entirely clear on how exactly that should be set up. (Kij is also using a kl_optimal for the scheduler, but I also couldn't get any coherent results out with that)
More testing is definitely needed, but its an exciting advancement.
@HazardAI Yes, more testing is definitely needed. I haven’t had any success using your V2V setup using 2 samplers with causevid tho, it always outputs a black image, idk why. That said, for video generation using 2 samplers is generally recommended for improved motion.
As for me, the sweet spot with causevid and your V2V is around 6–8 steps. I have very limited VRAM, so this setup lets me refine a video without waiting for hour. It only takes about 13–15 minutes to refine a 5 second, 1280x720 video.
Maybe you can make another workflow version with causevid.
@HazardAI I've gotten decent results with CausVid as low as 5 steps @ 0.5. For higher quality I went to 8-10 steps with lower lora strength. If using LORAs I usually set the cfg to 2 instead of 1, but still tweaking and experimenting. I also got the suggestion for a 2-sampler approach over at @RalFinger Discord. In my first simple flow I use no CausVid with cfg 6 for the 1st sampler for prompt adherence and CausVid with cfg 1 for the 2nd sampler to refine the output. Even a 4+4 or 3+3 step approach was pretty good. But this is also still in the experiment phase.
@fronyax Oh wow, I didn't realize you could go as low as 6 steps. (I pulled 15 from the Kijai example workflow). That does pretty drastically reduce the time. I just got a 61 frame, 8 steps, 480x688 render in 106 seconds vs 287 seconds for 24 steps with TeaCache. I'll definitely play with the multiple KSamplers some more. V1.3 will very likely include a CauseVid version, or at least instructions for what to bypass to use it.
@HazardAI But do you get cartoon oversaturated (almost cell shading effect) with CausVid ? because that is what happen to me when I use it, sure it's super fast but too low quality a loss of realism, maybe I'm doing something wrong
@fouchardmilcoupes311 Noticed that too, although mostly when I feed the last frame into the next clip and stitch them together. After 3-4 clips it is rather noticeable. But this latest WAN workflow over at @tremolo28 sounds like a solution: "It contains a Color Match option in postprocessing to cope with the increased saturation, the lora is introducing."
@fouchardmilcoupes311 I definitely got that on my first few tests using 15 steps. With 8 steps, CFG 1, and Euler Simpler, I'm getting visually comparable results to 24 steps with TeaCache. At least in the few source images I've tried so far. Wan does seem somewhat picky about settings when switching between 2d and photo real.
I don't understand something, Wan2.1 14b I2V at about 768 resolution (for longuest side) at around 81 frames, on some GGUF variant take about 16 to 21 GB of Vram on my system, then when the T2V 1.3b model take over for the refine smooth pass it take up to 22+ GB Vram I don't get it, are the other model not removed from ram or something ? Is this normal and expected ? (it work in the end, but I don't understand the high memory of the second pass with a vastly lighter model
You are correct that the initial model is not unloaded. ComfyUI seems to struggle with auto vram management with Wan models at the moment. I haven't yet found a good solution for automatically handling unloading the initial i2v model. (V1.0 of the workflow had a node to try and get it to unload, but it didn't seem to work.) If you want to use the least vram possible, you can use the group muter to only run the first step, then manually unload all models with ComfyUI Manager, and then enable and run the upscaling and refining. It's a bit tedious, but does work.
@HazardAI Yeah that might be a good plan B
@HazardAI I put an unload all model after each VAD decode (I mean those after Ksampler of course) seems to work, or at least doesn't seems to bother my PC, tested with Wan 2.1 14B t2v for the refiner too (in gguf format though, seems better quite long though)
For Unet Loader (GGUF) i am using
wan2.1-i2v-14b-480p-Q8_0.gguf
I am confused. because it wont work.
Everything loads just fine but it stops at:
My CLIPLoader (GGUF)
it says: CLIPLoaderGGUF
invalid tokenizer
When i use umt5-xxl-encoder-Q8_0.gguf
What text encoder am i supposed to use? Or is my unet loader much for my pc?
I have the latest comfyui and using pinokio
type = wan is selected on the clip loader node ? It work for me with Q4 to Q8 gguf version
If all fail, maybe making sure you are using latest comfuUi might solve most of the problems (The number of time just upgrading/reinstalling comfyUI to latest solved problem for me is countless time)
@fouchardmilcoupes311 i did the latest update for comfyui and it kinda crashed everything
@Aiartsenpai you ran the *.bat in the update folder
update_comfyui_and_python_dependencies.bat
this is the real update I'm talking about, if nothing work well maybe make just another install and start fresh, it can't be that bad if it worked before.
@fouchardmilcoupes311 oh i figured it but ty.
some reason my last frame glitches / blurs / over saturates. I have set the frame length to 81. Before this wasn't an issue. Maybe i accidentally tinkered with a setting and forgot about. Does anyone know why this may be happening?
I would love to see a variant of this workflow with First-Last frame, so you can supply two images where it should start and end up
Help i have 64 ram but not sure how to use it all. Anyone know a good smooth 720 workflow or 480 that is at least somewhat fast? I don't want to have a workflow that takes 2-4 hours per video.
64 ram or vram? What GPU do you have?
also where do i download RIFE VFI (recommend rife47 and rife49)? its an upscaler but not sure where to find the file to download it
That one should be handled automatically by the node when you run the workflow.
@HazardAI oh i figured it out. It was a bit buggy. updating fixed it
@HazardAI ty for helping out though
I get unetloadergguf and cliploadergguf not found, any idea what I'm doing wrong?
you probably have an outdated 'obsolete' comfyUI version, by not found you mean the nodes ar enot there at all or circled in red like unknown ?
My GPU is 5070ti with 16GB VRAM
I set the setting exactly like you suggested with triton, didn't alter other configuration. Yet the time to complete a 5 sec video is over 2 hours! Any idea why?
check your memory usage, if you see 98-99M of your V-ram and 98-99% of your system ram used, most likely you are dealing with the program trying to deal with memory problem, it might still complete but you are just not having enough memory, so reduce settings, or make sure you don't run other program at the same time eating your memory, I doubt you will ever be able to make a video in 5 seconds, but 2 hours seems way to high
and by the way, try 4 seconds video, it will significantly reduce any memory usage, maybe the problem is just 5 seconds lenght is a touch too much for your ram
I finally fixed it! It's really the issue of V-ram problem. I tried a node which free up V-ram when loading model, now I can create a 5 sec vid in 5 minutes. Thanks!
@oo4766229 cool, yeah it was really looking like that sort of problem
@fouchardmilcoupes311 Wondering how that happened though. I've seen people with V-ram less than mine maked videos in a reasonable amount of time.
How did you fixed it? I have 12gb VRAm and 16gb RAM
@davifpinto I used this node to load checkpoint https://github.com/pollockjj/ComfyUI-MultiGPU
@oo4766229 Thanks!!
@oo4766229 You figured out some way to use multigpu with i2v?
So good!