CivArchive

    All in One, Wan for All

    We are excited to introduce our latest model to our talented community creators:

    Wan2.1-VACE, All-in-One Video Creation and Editing model.

    Model size: 1.3B, 14B License: Apache-2.0

    If we are in Wan Day, what will it be like? 如果我们在万相世界,会是什么样子?

    模型支持两种文本到视频模型(1.3B 和 14B)和两种分辨率(480P 和 720P)。

    WAN-VACE is not a T2V model per se, but rather R(reference)2V, Can be understood as Video ControlNet for WAN , so there is no way to provide a T2V workflow. The CausVid accelerator is a distillation accelerator technology that can be used on WAN-VACE to provide 4-8 steps of accelerated generation.

    WAN-VACE本身不是T2V模型,而是R(参考)2V,可以理解为WAN的视频CN,因此无法提供T2V工作流程。CausVid加速器是一种蒸馏加速技术,可用于WAN-VACE,提供4-8步加速生成。

    Introduction

    VACE is an all-in-one model designed for video creation and editing. It encompasses various tasks, including reference-to-video generation (R2V), video-to-video editing (V2V), and masked video-to-video editing (MV2V), allowing users to compose these tasks freely. This functionality enables users to explore diverse possibilities and streamlines their workflows effectively, offering a range of capabilities, such as Move-Anything, Swap-Anything, Reference-Anything, Expand-Anything, Animate-Anything, and more.

    VACE是一款专为视频创建和编辑而设计的一体化模型。它包括各种任务,包括视频生成(R2V)、视频到视频编辑(V2V)和屏蔽视频到视频剪辑(MV2V),允许用户自由组合这些任务。此功能使用户能够探索各种可能性,并有效地简化他们的工作流程,提供一系列功能,如移动任何内容、交换任何内容、引用任何内容、扩展任何内容、为任何内容设置动画等。


    About CausVid-Wan2-1:

    5-16 The PERFECT solution to CausVid from Kijai (Best practices)

    Wan21_CausVid_14B_T2V_lora_rank32.safetensors · Kijai/WanVideo_comfy

    Through weight extraction and block separation,

    KJ give us a universal CausVid LoRA in rank32 for Any 14B WAN model,

    EVEN including FT models and I2V model!

    Although this may not have been CausVid's initial intention, by flexibly adjusting the LoRA parameters (0.3~0.5), we have achieved unprecedented availability on home grade graphics cards!

    KJ-Godlike also provides a 1.3B bidirectional inference version of LoRA export file

    Wan21_CausVid_bidirect2_T2V_1_3B_lora_rank32.safetensors

    same time, we also noticed that xunhuang1995 uploaded the Warp-4Step_cfg2 autoregressive version 1.3B CausVid model from: tianweiy/CausVid

    与为壹,全部在

    Best Adaptation for WAN-VACE full Models

    5/15 REDCausVid-Wan2-1-14B-DMD2-FP8 Uploaded 8-15 steps CFG 1

    本页面右侧下载列表,Safetensors 格式,workflow 在 Trainning data 压缩包内

    The download list on the right side of this page is in Safetensors format, and the workflow is included in the Training data compressed file. The example images and videos also include workflows (yes, you can directly throw the original video files into ComfyUI and try to capture the workflow)

    5/15 Aiwood WAN-ACE Fully functional workflow Uploaded

    5/15 ComfyUI KJ-WanVideoWrapper have been updated

    5/14 autoregressive_checkpoint.pt 1.3b Uploaded , PT UNET Loader

    5/14 bidirectional_checkpoint2.pt 1.3b Uploaded , PT UNET Loader

    NEW Sampler Flowmatch_causvid in KJ-WanVideoWrapper

    Releases from:

    kijai/ComfyUI-WanVideoWrapper

    ⭐ leave a star⭐

    [ The adaptability test results of WAN1.2 LoRAs for VACE show that about 75% of I2V/T2V LoRA weights can take effect, but the sensitivity is reduced ( try to increase the LoRA weight ,more than 100% Sometimes it can be helpful ) ]

    Fullview of Aiwood WAN-ACE Fully functional workflow:

    source: https://www.bilibili.com/video/BV1FGE6zGEDK ⭐ leave a star⭐

    CausVid 加速器项目页 https://causvid.github.io/


    WAN-VACE 模型的参数和配置如下:

    📌 Wan2.1-VACE provides solutions for various tasks, including reference-to-video generation (R2V), video-to-video editing (V2V), and masked video-to-video editing (MV2V), allowing creators to freely combine these capabilities to achieve complex tasks.

    👉 Multimodal inputs enhancing the controllability of video generation.

    👉 Unified single model for consistent solutions across tasks.

    👉 Free combination of capabilities unlocking deeper creative

    📌 Wan2.1-VACE为各种任务提供解决方案,包括参考视频生成(R2V)、视频到视频编辑(V2V)和屏蔽视频到视频剪辑(MV2V),允许创作者自由组合这些功能来实现复杂的任务。

    👉 多模态输入增强了视频生成的可控性。

    👉 统一的单一模型,实现跨任务的一致解决方案。

    👉 自由组合功能,释放更深层次的创造力


    WAN实时生成来了Hybrid AI model crafts smooth, high-quality videos in seconds

    The CausVid generative AI tool uses a diffusion model to teach an autoregressive (frame-by-frame) system to rapidly produce stable, high-resolution videos.

    Wan2.1based 混合AI模型在几秒钟内(9帧/秒)制作出流畅、高质量的视频

    CausVid生成AI工具使用扩散模型来指导自回归(逐帧)系统快速生成稳定的高分辨率视频。

    Hybrid AI model crafts smooth, high-quality videos in seconds | MIT News | Massachusetts Institute of Technology

    From Slow Bidirectional to
    Fast Autoregressive Video Diffusion Models

    CausVid https://causvid.github.io/

    tianweiy (Tianwei Yin)

    RedCaus/REDCausVid-Wan2-1-14B-DMD2-FP8 Uploaded / WAN-VACE14B 最佳适配

    CausVid/autoregressive_checkpoint uploaded / 自回归模型基于 WAN1.3B 已收录

    CausVid/bidirectional_checkpoint2 uploaded / 双向推导模型基于 WAN1.3B 已收录

    Kijai/Wan2_1-T2V-14B_CausVid_fp8_e4m3fn.safetensors / HF仓库 WanVideo_comfy

    ⭐ leave a star⭐

    Brief computer-generated animation of a character in an old deep-sea diving suit walking on a leaf

    licensed by Creative Commons Attribution Non Commercial 4.0

    Thank you for this friend's additional comment. I was too excited last night and didn't sleep, so I stopped updating before finishing:

    We’ll need to use the official Python-based inference codes

    1) Clone https://github.com/tianweiy/CausVid and follow instructions to install requirements

    2) Clone https://huggingface.co/Wan-AI/Wan2.1-T2V-1.3B into wan_models/Wan2.1-T2V-1.3B

    3) Put the pt file inside checkpoint_folder/model.pt

    4) Run inference code, python minimal_inference/autoregressive_inference.py --config_path configs/wan_causal_dmd.yaml --checkpoint_folder XXX --output_folder XXX --prompt_file_path XXX

    Reddit posts about CausVid: https://www.reddit.com/r/StableDiffusion/comments/1khjy4o/causvid_generate_videos_in_seconds_not_minutes/

    https://www.reddit.com/r/StableDiffusion/comments/1k0gxer/causvid_from_slow_bidirectional_to_fast/

    We have tested the CausVid based on Wan1.3b version, which has incredible speed, and are currently testing the 14B version produced by lightx2v.

    LightX2V: Light Video Generation Inference Framework

    Supported Model List

    HunyuanVideo-T2V

    HunyuanVideo-I2V

    Wan2.1-T2V

    Wan2.1-I2V

    Wan2.1-T2V-CausVid

    SkyReels-V2-DF

    How to Run

    Please refer to the documentation in lightx2v.

    ⭐ leave a star⭐


    通义实验室 WAN 2.1 Model Zoo

    Institute for Intelligent Computing专注于各领域大模型技术研发与创新应用。实验室研究方向涵盖自然语言处理、多模态、视觉AIGC、语音等多个领域。我们并积极推进研究成果的产业化落地。实验室同时积极参与开源社区建设,全方位拥抱开源社区,共同探索AI模型的开源开放。

    Developer / Models Name / Kijai`s ComfyUI Model


    RedCaus/REDCausVid-Wan2-1-14B-DMD2-FP8 Uploaded / WAN-VACE14B 最佳适配

    CausVid/autoregressive_checkpoint included / 自回归模型基于 WAN1.3B 已收录

    CausVid/bidirectional_checkpoint2 included / 双向推导模型基于 WAN1.3B 已收录

    CausVid/wan_causal_ode_checkpoint_model testing / 自回归因果推导 测试中

    CausVid/wan_i2v_causal_ode_checkpoint_model testing / 文生图模型 测试中

    lightx2v/Wan2.1-T2V-14B-CausVid unqualify / 自回归模型14B AiWood实测不达标

    lightx2v/Wan2.1-T2V-14B-CausVid quant unqualify / 自回归模型14B量化版 实测不达标


    Wan Team/1.3B text-to-video included / 文生视频1.3B 已收录

    Wan Team/14B text-to-video included / 文生视频14B 已收录

    Wan Team/14B image-to-video 480P included / 图生视频14B 已收录

    Wan Team/14B image-to-video 720P included / 图生视频14B 已收录

    Wan Team/14B first-last-frame-to-video 720P included / 视频首尾帧 已收录

    Wan Team/Wan2_1_VAE included / KiJai‘s WAN视频VAE 已收录

    ComfyORG/Wan2.1_VAE included / Comfy‘s WAN视频VAE 已收录

    google/umt5-xxl umt5-xxl-enc safetensors included / TE编码器 已收录

    mlf/open-clip-xlm-roberta-large-vit-huge-14 safetensors included / CLIP编码器 已收录


    DiffSynth-Studio Team/1.3B aesthetics LoRA 美学蒸馏-通义万相2.1-1.3B-LoRA-v1

    DiffSynth-Studio Team/1.3B Highres-fix LoRA 高分辨率修复-通义万相2.1-1.3B-LoRA-v1

    DiffSynth-Studio Team/1.3B ExVideo LoRA 长度扩展-通义万相2.1-1.3B-LoRA-v1

    DiffSynth-Studio Team/1.3B Speed Control adapter 速度控制-通义万相2.1-1.3B-适配器-v1


    PAI Team/ WAN2.1 Fun 1.3B InP 支持首尾帧 / Kijai/WanVideo_comfy

    PAI Team/ WAN2.1 Fun 14B InP 支持首尾帧 / Kijai/WanVideo_comfy

    PAI Team/ WAN2.1 Fun 1.3B Control 控制器 / Kijai/WanVideo_comfy

    PAI Team/ WAN2.1 Fun 14B Control 控制器 / Kijai/WanVideo_comfy

    PAI Team/ WAN2.1 Fun 14B Control 控制器 / Kijai/WanVideo_comfy

    PAI Team/ WAN2.1-Fun-V1_1-14B-Control-Camera / Kijai/WanVideo_comfy

    IIC Team/ VACE-通义万相2.1-1.3B-Preview / Kijai/WanVideo_comfy


    IC ( In-Context ) Controler 多模态控制器 :

    ali-vilab/ VACE: All-in-One Video Creation and Editing / Kijai/WanVideo_comfy

    Phantom-video/Phantom Subject-Consistent via Cross-Modal Alignment

    KwaiVGI/ ReCamMaster Camera-Controlled 镜头多角度 / Kijai/WanVideo_comfy


    Digital Character 数字人 via Wan2.1 :

    ali-vilab/ UniAnimate-DiT 长序列骨骼角色视频 / Kijai/WanVideo_comfy

    Fantasy-AMAP/ 音频驱动数字人 FantasyTalking / Kijai/WanVideo_comfy

    Fantasy-AMAP/ 角色一致性身份保留 FantasyID / Fantasy-AMAP/fantasy-id


    Uncensored NSFW 解锁版本:

    REDCraft AIGC / WAN2.1 720P NSFW Unlocked / forPrivate use【非公开】

    CubeyAI / WAN General NSFW model (FIXED) / The Best Universal LoRA


    昆仑万维发布 SkyReels based on Wan2.1

    Skywork / SkyReels-V2-I2V-14B-720P / Image-to-Video / Kijai/WanVideo_comfy

    Skywork / SkyReels-V2-I2V-14B-540P / Image-to-Video / Kijai/WanVideo_comfy

    Skywork / SkyReels-V2-T2V-14B-540P / Text-to-Video / Kijai/WanVideo_comfy

    Skywork / SkyReels-V2-T2V-14B-720P /Text-to-Video / Kijai/WanVideo_comfy

    Skywork / SkyReels-V2-I2V-1.3B-540P / Image-to-Video / Kijai/WanVideo_comfy


    AutoRegressive Diffusion-Forcing 无限长度生成架构

    Skywork / SkyReels-V2-DF-14B-720P / Text-to-Video / Kijai/WanVideo_comfy

    Skywork / SkyReels-V2-DF-14B-540P / Text-to-Video / Kijai/WanVideo_comfy

    Skywork / SkyReels-V2-DF-1.3B-540P / Text-to-Video / Kijai/WanVideo_comfy


    昆仑万维发布 SkyReels 视频标注模型:

    Skywork / SkyCaptioner-V1 Skywork (Skywork) / Skywork/SkyCaptioner-V1


    Tiny AutoEncoder / taew2_1 safetensors / Kijai/WanVideo_comfy

    A tiny distilled VAE model for encoding images into latents and decoding latent representations into images


    WAN Comfy-Org/Wan_2.1_ComfyUI_repackaged

    【例图页面蓝色Nodes或下载webp文件-可复现视频工作流】

    Gallery sample images/videos (WEBP format) including the ComfyUI native workflow

    This is a concise and clear GGUF model loading and tiled sampling workflow:

    Wan 2.1 Low vram Comfy UI Workflow (GGUF) 4gb Vram - v1.1 | Wan Video Workflows | Civitai

    节点:(或使用 comfyui manager 安装自定义节点)

    https://github.com/city96/ComfyUI-GGUF

    https://github.com/kijai/ComfyUI-WanVideoWrapper

    https://github.com/BlenderNeko/ComfyUI_TiledKSampler

    * 注意需要更新到最新版本的 comfyui-KJNodes GitHub - kijai/ComfyUI-KJNodes: Various custom nodes for ComfyUI update to the latest version of Comfyui KJNodes


    Kijai ComfyUI wrapper nodes for WanVideo

    WORK IN PROGRESS

    @kijaidesign 's works

    Huggingface - Kijai/WanVideo_comfy

    GitHub - kijai/ComfyUI-WanVideoWrapper

    主图视频来自 AiWood

    https://www.bilibili.com/video/BV1TKP3eVEue

    Text encoders to ComfyUI/models/text_encoders

    Transformer to ComfyUI/models/diffusion_models

    Vae to ComfyUI/models/vae

    Right now I have only ran the I2V model succesfully.

    Can't get frame counts under 81 to work, this was 512x512x81

    ~16GB used with 20/40 blocks offloaded


    DiffSynth-Studio Inference GUI

    Wan-Video LoRA & Finetune training.

    DiffSynth-Studio/examples/wanvideo at main · modelscope/DiffSynth-Studio · GitHub


    💜 Wan    |    🖥️ GitHub    |   🤗 Hugging Face   |   🤖 ModelScope   |    📑 Paper (Coming soon)    |    📑 Blog    |   💬 WeChat Group   |    📖 Discord  


    Wan: Open and Advanced Large-Scale Video Generative Models

    通义万相Wan2.1视频模型开源!视频生成模型新标杆,支持中文字效+高质量视频生成

    In this repository, we present Wan2.1, a comprehensive and open suite of video foundation models that pushes the boundaries of video generation. Wan2.1 offers these key features:

    • 👍 SOTA Performance: Wan2.1 consistently outperforms existing open-source models and state-of-the-art commercial solutions across multiple benchmarks.

    • 👍 Supports Consumer-grade GPUs: The T2V-1.3B model requires only 8.19 GB VRAM, making it compatible with almost all consumer-grade GPUs. It can generate a 5-second 480P video on an RTX 4090 in about 4 minutes (without optimization techniques like quantization). Its performance is even comparable to some closed-source models.

    • 👍 Multiple Tasks: Wan2.1 excels in Text-to-Video, Image-to-Video, Video Editing, Text-to-Image, and Video-to-Audio, advancing the field of video generation.

    • 👍 Visual Text Generation: Wan2.1 is the first video model capable of generating both Chinese and English text, featuring robust text generation that enhances its practical applications.

    • 👍 Powerful Video VAE: Wan-VAE delivers exceptional efficiency and performance, encoding and decoding 1080P videos of any length while preserving temporal information, making it an ideal foundation for video and image generation.

    This repository features our T2V-14B model, which establishes a new SOTA performance benchmark among both open-source and closed-source models. It demonstrates exceptional capabilities in generating high-quality visuals with significant motion dynamics. It is also the only video model capable of producing both Chinese and English text and supports video generation at both 480P and 720P resolutions.

    Description

    Comfy-Org/Wan_2.1_ComfyUI_repackaged

    Wan 2.1 repackaged for ComfyUI use. For examples see: https://comfyanonymous.github.io/ComfyUI_examples/wan

    ---

    Wan 2.1 Models

    Wan 2.1 is a family of video models.

    ---

    Files to Download

    You will first need:

    Text encoder and VAE:

    umt5_xxl_fp8_e4m3fn_scaled.safetensors goes in: ComfyUI/models/text_encoders/

    wan_2.1_vae.safetensors goes in: ComfyUI/models/vae/

    ---

    Video Models

    files go in: ComfyUI/models/diffusion_models/

    These examples use the 16 bit files but you can use the fp8 ones instead if you don’t have enough memory.

    ---

    Workflows

    ---

    Text to Video

    Workflow in Json format

    files (put it in: ComfyUI/models/diffusion_models/). You can also use it with the 14B model.

    ---

    Image to Video

    Workflow in Json format

    Note this example only generates 33 frames at 512x512 because I wanted it to be accessible, the model can do more than that. The 720p model is pretty good if you have the hardware/patience to run it.

    FAQ

    Comments (60)

    genuralMar 5, 2025· 2 reactions
    CivitAI

    Is it possible for you to create GGUF versions?

    AiMetatron
    Author
    Mar 5, 2025· 2 reactions

    The GGUF version of the model produced by ComfyORG is currently being uploaded

    Please refer to the list page

    AiMetatron
    Author
    Mar 5, 2025· 1 reaction

    UNET loader (GGUF) doesn't work, so I switched to this simple and straightforward workflow: Wan 2.1 Low vram Comfy UI Workflow (GGUF) 4gb Vram - v1.1 | Wan Video Workflows | Civitai

    AiMetatron
    Author
    Mar 5, 2025· 2 reactions

    我更新了下comfyui-GGUF组件,直接加载就开始干活了!

    Updated comfyui GGUF , now able to load Wan's GGUF model directly

    ohddugi0516816Mar 5, 2025· 3 reactions
    CivitAI

    Is there a workflow for the GGUF version that went live today?

    AiMetatron
    Author
    Mar 5, 2025

    Replace UNET loader with UNET loader (GGUF)?

    AiMetatron
    Author
    Mar 5, 2025· 2 reactions

    UNET loader (GGUF) doesn't work, so I switched to this simple and straightforward workflow: Wan 2.1 Low vram Comfy UI Workflow (GGUF) 4gb Vram - v1.1 | Wan Video Workflows | Civitai

    AiMetatron
    Author
    Mar 5, 2025

    我更新了下comfyui-GGUF组件,直接加载就开始干活了!

    Updated comfyui GGUF , now able to load Wan's GGUF model directly

    rockiecxhMar 7, 2025

    @METAFILM_Ai I tried this workflow, it didn't work well.

    OzzyOsmanMar 5, 2025· 2 reactions
    CivitAI

    Please add some workflow for the Comfyorg GGUF model. Thanks

    AiMetatron
    Author
    Mar 5, 2025

    Replace UNET loader with UNET loader (GGUF)?

    AiMetatron
    Author
    Mar 5, 2025· 3 reactions

    UNET loader (GGUF) doesn't work, so I switched to this simple and straightforward workflow: Wan 2.1 Low vram Comfy UI Workflow (GGUF) 4gb Vram - v1.1 | Wan Video Workflows | Civitai

    AiMetatron
    Author
    Mar 5, 2025· 2 reactions

    我更新了下comfyui-GGUF组件,直接加载就开始干活了!

    Updated comfyui GGUF , now able to load Wan's GGUF model directly

    SantaonholidaysMar 6, 2025· 2 reactions
    CivitAI

    1 Question.

    Do you guys really wait 30 minutes or do you use runpod? I require the link please if you do c:

    AiMetatron
    Author
    Mar 7, 2025· 2 reactions

    I hope this is not an advertisement. I will run a 720P model locally to generate 33 frames of images, which will only take 3-5 minutes (4090D). However, the generation size will be set lower, and then super-resolution will be performed

    tzdivMar 12, 2025

    @METAFILM_Ai Hello, thx for model!!! how you upscale video?

    itwasentme83Mar 6, 2025· 5 reactions
    CivitAI

    about as uncensored as my granma and she been dead for over 35 years

    AiMetatron
    Author
    Mar 7, 2025· 2 reactions

    Fortunately, we have just been resurrected. Let monsoons flood the censors' dictionaries—since when did seedlings need permission to crack concrete with their roots?

    TartrixroarMar 6, 2025· 1 reaction
    CivitAI

    Can't open worfklow it says missing custom node: wanimage2video

    AiMetatron
    Author
    Mar 7, 2025· 1 reaction

    So you can use ComfyUI manager to install the missing nodes, also You can manually download it here: https://github.com/kijai/ComfyUI-WanVideoWrapper

    KurielMar 8, 2025· 1 reaction

    update ComfyUI to latest

    NtggMar 9, 2025· 1 reaction

    I couldn't get the missing node until I updated by opening the bat files: "update_comfyui", and "update_comfyui_and_python_dependencies" from the update folder. Updating and installing manually didn't work for me. I'm also using the portable version btw.

    AiMetatron
    Author
    Mar 9, 2025· 1 reaction

    @Ntgg Right~After ComfyUI officially supports Wan2.1, we need to update the Comfy version

    7408956Mar 7, 2025· 3 reactions
    CivitAI

    So it looks like Wan has vastly superior I2V than Hunyuan, as long as it doesn't need to generate private areas?

    Why can't we just get a model that's good at everything? We're running out of multi-billion dollar Chinese tech giants to gift us free stuff.

    HunYuan team: Please put out a HunYuan T2V without any censorship please. I know you trained it on all kinds of things (concepts are still partially there or in reduced quality/accuracy) and apparently walked back some of that ability for release.

    AiMetatron
    Author
    Mar 7, 2025· 1 reaction

    Perhaps these tech giants do not belong to any localization, and similarly, open source models still have to rely on us to figure out everything ourselves. Why did you delete your own account?

    7447100Mar 9, 2025

    @METAFILM_Ai  Yeah, I'm personally irritated by the need to keep an account just to download some stuff, so I delete and make new ones all the time. I don't engage in an abusive way or spam, so it shouldn't be an issue.

    luongtuananh2017761Mar 8, 2025· 2 reactions
    CivitAI

    can't use increase resolution(Upscaler and frame interpolation), it terminates the program immediately, what happened?

    AiMetatron
    Author
    Mar 8, 2025· 1 reaction

    The most likely cause is CPU memory overflow. ComfyUI has good VRAM management, but excessive memory requests for components can cause program interruptions

    luongtuananh2017761Mar 8, 2025· 1 reaction

    @METAFILM_Ai In my case with Google Colab Pro+, everything is stable and good when I skip the upscaler and frame interpolation process. However, unfortunately, there is no error message, it simply terminates the task.

    AiMetatron
    Author
    Mar 8, 2025· 1 reaction

    @luongtuananh2017761 Can fluctuations in memory resources be detected? Especially at the moment when a large number of super-resolution slices are loaded into the node program. Recommend using devices with 64GB-128GB memory for super-resolution and frame insertion

    gausssidorov928Mar 9, 2025

    Файл подкачки на SSD установи 64-Gig,

    uggiugi681Mar 10, 2025

    @METAFILM_Ai Personally I recommend 128gb for upscaling

    dillion1920Mar 8, 2025· 2 reactions
    CivitAI

    This goes in text encoders?

    wanAIWan21VideoModelSafetensors_kijaiOpenCliplVitWan

    AiMetatron
    Author
    Mar 9, 2025· 1 reaction

    Yes, this is the encoder required for text generated videos.

    Eliz103Mar 10, 2025

    @METAFILM_Ai Hi! where I can find workflows for Kijai? im very new in this, please help!

    EshinioMar 9, 2025· 1 reaction
    CivitAI

    How is this model in regards to NSFW and censorship? Are you able to generate NSFW stuff without any Loras or do you have to make clever prompting to get around it?

    AiMetatron
    Author
    Mar 9, 2025· 3 reactions

    While the official model probably went through rigorous compliance checks, Seems like the model's creators had to jump through tons of compliance hoops, but hey – that's why we've got genius community wizards doing mind-blowing hacks! With their i2v models using reference images as prompts? Seriously, the creative possibilities are endless.

    alan0nMar 9, 2025· 2 reactions
    CivitAI

    This is really solid. I'm honestly impressed how well my dinky little 4080 ran this. Kudos for putting everything needed to do it all in one convenient place.

    AiMetatron
    Author
    Mar 9, 2025· 2 reactions

    Yeah, don't set the Excessive resolution (I only used 480 * 720 during testing) and 33 frames are enough for testing

    2legsRises357Mar 9, 2025· 3 reactions
    CivitAI

    looks great but doesnt work - ksampler says triton not isntalled."torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised:"

    what even is that?

    alan0nMar 9, 2025· 2 reactions

    Assuming you tried 1.5? It doesn't work because you didn't install triton and sageassist. Those are external dependencies that comfymanager will not just auto download.

    AiMetatron
    Author
    Mar 9, 2025· 1 reaction

    in Win10 environment, the inference accelerator can be set to SDPA

    AiMetatron
    Author
    Mar 9, 2025· 1 reaction

    This way, there is no need to tinker with the compilation environment to install Triton and Sageassist, as they are both equally slow

    dollgeneratorMar 10, 2025

    Ask chatgpt how to install triton and sageattention and follow the steps. Sageattention and triton are used to speed up video generation by 50%. If you really cannot install it, you can remove those nodes.

    cuddlekittyboutique679Mar 9, 2025· 2 reactions
    CivitAI

    Total novice question: This only works on comfy UI? Not on other UIs?

    AiMetatron
    Author
    Mar 9, 2025· 3 reactions

    Give Diffsynth Studio a shot! You can use it for inferences—it even handles Basemodel and LoRA training, GitHub - modelscope/DiffSynth-Studio: Enjoy the magic of Diffusion models!

    However, the most commonly used inference UI is still ComfyUI =)

    NoneofYourBusiness2Mar 10, 2025· 10 reactions
    CivitAI

    There's a lot of stuff on this post. Can someone summarize few key points.

    -What is the point of comfyorg models?
    -What are Kijai versions of the models?

    Are these just rehosted models for OPs own purposes, or do these for example alter the model somehow to enable NSFW?

    omar92Mar 10, 2025· 1 reaction

    +1

    ratamahaeva757Mar 14, 2025

    +1

    losveratos102486Mar 15, 2025

    +100

    Kung_fu_PronMar 10, 2025· 1 reaction
    CivitAI

    Any reason why the comfyui native worflow give a way better nsfw results using the 480p GGUF instead of the 720p_fp8 at 512x780? the image is way more cleaner and way more natural "it's a little too slow motion" but natural motion.


    combotMar 11, 2025· 2 reactions
    CivitAI

    Is there a difference between using the 720P and 480P models to generate 832x480 video?

    AiMetatron
    Author
    Mar 14, 2025

    Yes, there is a difference. The 720P model may exhibit excessively high material density when generating small sizes.

    azeliMar 15, 2025· 1 reaction

    @METAFILM_Ai What does this mean?

    engineerMar 12, 2025· 1 reaction
    CivitAI

    Is there a GGUF of the 14B 720p I2V model?

    AiMetatron
    Author
    Mar 14, 2025· 1 reaction

    soon

    tzdivMar 12, 2025· 3 reactions
    CivitAI

    Hello, how you upscale video?

    AiMetatron
    Author
    Mar 14, 2025· 2 reactions

    Webp format videofiles contain workflows, which can be dragged to ComfyUI to view

    StDismasMar 14, 2025· 1 reaction

    @METAFILM_Ai can you explain to me how to get .webp with workflow from videos on civitai? Before this I only dragged images or opened workflow from my own videos, dragging images that are additionally saved by videohelpersuite node. thx

    AiMetatron
    Author
    Mar 15, 2025· 1 reaction

    @StDismas Find a video file in WEBP format generated by ComfyUI, download it to local, and drag it into the ComfyUI web interface. (If the WEBP file saves the workflow) you can see it in ComfyUI, just like complete files with workflows such as PNG or JPG.

    Checkpoint
    Wan Video

    Details

    Downloads
    5,447
    Platform
    CivitAI
    Platform Status
    Available
    Created
    3/4/2025
    Updated
    4/30/2026
    Deleted
    -

    Files

    onTHEFLYWanAIWan21VideoModel_comfyorgWan21Vae.safetensors

    Mirrors

    Huggingface (115 mirrors)