CivArchive
    Stable Video Diffusion (SDV) Workflow ComfyUI - txt2img
    Preview 3898985Preview 3898994

    Stable Video Diffusion (Original Stability Release)

    As of writing this there are two image to video checkpoints. Here are the official checkpoints for the one tuned to generate 14 frame videos and the one for 25 frame videos. Put them in the ComfyUI/models/checkpoints folder.

    The most basic way of using the image to video model is by giving it an init image like in the folowing workflow that uses the 14 frame model. You can download this webp animated image and load it or drag it on ComfyUI to get the workflow.



    Some explanations for the parameters:

    video_frames: The number of video frames to generate.

    motion_bucket_id: The higher the number the more motion will be in the video.

    fps: The higher the fps the less choppy the video will be.

    augmentation level: The amount of noise added to the init image, the higher it is the less the video will look like the init image. Increase it for more motion.

    VideoLinearCFGGuidance: This node improves sampling for these video models a bit, what it does is linearly scale the cfg across the different frames. In the above example the first frame will be cfg 1.0 (the min_cfg in the node) the middle frame 1.75 and the last frame 2.5. (the cfg set in the sampler). This way frames further away from the init frame get a gradually higher cfg.



    Repost, available from:
    https://huggingface.co/stabilityai/stable-video-diffusion-img2vid-xt/tree/main

    Description

    Workflows
    Other

    Details

    Downloads
    1,561
    Platform
    CivitAI
    Platform Status
    Available
    Created
    11/25/2023
    Updated
    9/28/2025
    Deleted
    -

    Files

    stableVideoDiffusionSDV_txt2img.zip

    Mirrors