CivArchive
    Live Portrait Workflow [XL, PONY, SD1.5] - v1.0
    NSFW

    This workflow is a streamlined Video2Image2Video to create live portraits. You can cartoonify yourself or use videos to animate your favorite characters! I'll include a .json file of the workflow and an example copyright-free video for you to follow.

    Here’s a breakdown of each step:

    Step 1: Load the Video and Get the First Frame

    To start, you'll need to grab the first frame from your video, which will be the image to which the nodes map all the facial points for the animation.

    • Bypass all other groups in the workflow

    • Load the video and queue up your prompt to start.

    • Copy the first frame to your clipspace once it's loaded.

    • If your video requires audio, connect the audio output to the corresponding node.

    • After copying the frame, bypass this node and move to step 2.

    Step 2: Generate the Character or Face Using Image-to-Image (I2I)

    Now, it's time to create the character or face you want to animate.

    • Bypass all other groups except for this one.

    • Paste the first frame from your clipboard into the load image node. (you can stretch a pipe with the image to the upscale image node if you prefer.)

    • Choose a ControlNet model that ensures the face's position stays consistent throughout the video (OpenPose and LineArt work well here). Adjust the ControlNet strength and end_percent and the scheduler denoise to achieve the image you like. The basic guidance is turning the denoise down, and the ControlNet strength and end_percent up makes the input image influence the final result more.

    • Run image-to-image multiple times to adjust and fine-tune the facial features and animation style until you’re happy with it. I like using the style selector from the Easy-use custom nodes pack to capture styles easily. (I highly recommend this checkpoint by 7whitefire7 for cartoon styles. It is amazing! RealCartoon-XL - V7 | Stable Diffusion XL Checkpoint | Civitai)

    • If needed, you can copy the output image, paste it back into the load image node, and run it again to refine details further while keeping the size and framing intact.

    • Once you're satisfied, copy the I2I output to clip space and move on to step 3. (again, you can connect the pipe from the VAE decode node to the step 3 upscale image node if you prefer.)

    Pro Tip: A 1:1 ratio works best with live portrait nodes, so keep that in mind while generating.

    Step 3: Match the First Frame for Best Face Mapping

    This step ensures that your animation starts in sync with the first frame of the video.

    • Bypass groups 2 and 4 .

    • Paste the I2I output image from step 2 into the load image node.

    • Queue the prompt to match the first frame's expression and head position. This is where the tracking points will be mapped. In this example, I did not get a good map. I will turn down the sample_ratio input so that sample frame 1 has less influence.

    • Adjust expressions if needed (e.g., increase "aaa" for an open mouth or "smile" if the character is smiling). In this example, I used rotate_pitch to turn her head down a little, adjusted me and woo to change her mouth shape and turned the sample_ratio down.

    • Ensure auto-queuing is enabled or queue up the prompt manually to see the expression changes in the expression editor node.

    • Adjust the crop factor input to zoom in 2.0 is no crop, and 1.0 is a max crop. Match this setting in the advanced live portrait node in step 4.

    • Once the face mapping is set, move on to the final step.

    Step 4: Generate the Final Video, Interpolate, and Save

    This is where it all comes together!

    • Bypass step 2, and make sure steps 1, 3, and 4 are active.

    • Check if you're including the audio, ensuring it's connected properly.

    • Queue up your prompt, and the workflow will handle your face mapping and interpolation.

    • Once everything looks good, save your final video!

    Wrapping Up

    That’s it! This simple four-step workflow makes using a video to guide an animated portrait easily.

    If you run into any issues or have questions, comment, and I’ll be happy to help!

    Workflow Author: OpalSky
    Resources:

    Enjoy creating your live portraits, and I can't wait to see what you come up with!

    Description

    FAQ

    Workflows
    Pony

    Looks like we don't have an active mirror for this file right now.

    CivArchive is a community-maintained index — we catalog mirrors that volunteers upload to HuggingFace, torrents, and other public hosts. Looks like no one has uploaded a copy of this file yet.

    Some files do get recovered over time through contributions. If you're looking for this one, feel free to ask in Discord, or help preserve it if you have a copy.

    Details

    Downloads
    675
    Platform
    CivitAI
    Platform Status
    Deleted
    Created
    10/23/2024
    Updated
    4/27/2026
    Deleted
    4/17/2026

    Files

    livePortraitWorkflowXL_v10.zip

    Mirrors