CivArchive
    Wan 2.2 Animate & InfiniteTalk+UniAnimate - InstantID & Supir Upscale
    NSFW
    Preview 104949575
    Preview 104950310
    Preview 104950509
    Preview 104949623
    Preview 104949685
    Preview 104949787
    Preview 104949877
    Preview 104950069
    Preview 104950391

    Wanvideo 2.2 Animate

    I uploaded a version of ComfyUI without models (4.38G), it has all the nodes and workflows to run and avoid errors, if you have your own ComfyUI, you can ignore this part

    https://drive.google.com/file/d/1lcK9QX6FmLO6rQNP6200rbLiqnHpt0uK/view?usp=drive_link

    1.Model

    ComfyUI\models\diffusion_models

    Wan22Animate-Kijai

    ComfyUI\models\vae

    Wan2_1_VAE_fp32.safetensors

    ComfyUI\models\text_encoders

    umt5-xxl-enc-fp8_e4m3fn.safetensors

    umt5-xxl-enc-bf16.safetensors

    ComfyUI\models\clip_vision

    clip_vision_h.safetensors

    ComfyUI\models\Loras

    WanAnimate_relight_lora_fp16

    lightx2v_I2V_14B_480p_cfg_step_distill_rank256_bf16.safetensors

    Kijai Lightx2v

    ComfyUI\models\detection

    vitpose-l-wholebody

    yolov10m

    ComfyUI\models\sams\

    Sec-4B

    ---

    ComfyUI\models\custom_nodes

    auto_wan2.2animate_freamtowindow_server

    ComfyUI-WanAnimatePreprocess

    Download the ZIP file and extract to custom_nodes.

    ---

    Wan2.2 Kijai I2V

    I uploaded a version of ComfyUI without models (4.38G), it has all the nodes and workflows to run and avoid errors, if you have your own ComfyUI, you can ignore this part

    https://drive.google.com/file/d/1lcK9QX6FmLO6rQNP6200rbLiqnHpt0uK/view?usp=drive_link

    1.Model

    ComfyUI\models\diffusion_models

    Kijai bf16

    Wan2_2-I2V-A14B-HIGH_bf16.safetensors

    Wan2_2-I2V-A14B-LOW_bf16.safetensors

    Kijai fp8

    WanVideo_comfy_fp8_scaled

    lightx2v Distill

    lightx2v/Wan2.2-Distill-Models

    (After testing, it can greatly improve video dynamics)

    ---

    ComfyUI\models\vae

    Wan2_1_VAE_fp32.safetensors

    ComfyUI\models\text_encoders

    umt5-xxl-enc-fp8_e4m3fn.safetensors

    umt5-xxl-enc-bf16.safetensors

    ComfyUI\models\clip_vision

    clip_vision_h.safetensors

    ComfyUI\models\Loras

    lightx2v_I2V_14B_480p_cfg_step_distill_rank256_bf16.safetensors

    Wan2.2-Fun-A14B-InP-high-noise-MPS.safetensors

    Wan2.2-Fun-A14B-InP-low-noise-HPS2.1.safetensors

    Kijai Lightx2v

    ---

    InfiniteTalk + UniAnimate

    I uploaded a version of ComfyUI without models (4.38G), it has all the nodes and workflows to run and avoid errors, if you have your own ComfyUI, you can ignore this part

    https://drive.google.com/file/d/1lcK9QX6FmLO6rQNP6200rbLiqnHpt0uK/view?usp=drive_link

    1.Model

    ComfyUI\models\diffusion_models

    Wan2.1-I2V-14B-480P

    Wan2_1-InfiniTetalk-Single_fp16

    ComfyUI\models\unet

    InfiniteTalk_GGUF

    Wan2.1-I2V-14B-480P-gguf

    ComfyUI\models\vae

    Wan2_1_VAE_fp32

    ComfyUI\models\text_encoders

    umt5-xxl-enc-fp8_e4m3fn.safetensors

    umt5-xxl-enc-bf16.safetensors

    ComfyUI\models\clip_vision

    clip_vision_h.safetensors

    ComfyUI\models\Loras

    UniAnimate-Wan2.1-14B-Lora-12000-fp16

    lightx2v_I2V_14B_480p_cfg_step_distill_rank256_bf16.safetensors

    Kijai Lightx2v

    ---

    InstantID

    InstantID does not support the latest version of ComfyUI. Please download the old version. I deleted the model and kept the nodes to ensure that there will be no errors when running (5.62g)

    https://drive.google.com/file/d/1_HPGG2iMAyovS3jO8ubhti4YIcOhOLUj/view?usp=drive_link

    1.Model

    ComfyUI\models\checkpoints

    leosamsHelloworldXL_helloworldXL50GPT4V

    lustifySDXLNSFWSFW_v20

    majicmixRealistic_v7

    SUPIR-v0Q

    ComfyUI\models\loras

    POVMissionary

    pov-squatting-cowgirl-lora-1-mb

    PornMaster-Amateur&DPO

    Hand v3 SD1.5

    ComfyUI\models\unet

    FLUX.1 [dev]

    FLUX.1 [dev] fp8

    ComfyUI\models\clip\

    t5xxl_fp16.safetensors

    t5xxl_fp8_e4m3fn.safetensors

    clip_l.safetensors

    CLIP-ViT-bigG-14-laion2B-39B-b160k

    Rename 「open_clip_model.safetensors」 to 「CLIP-ViT-bigG-14-laion2B-39B-b160k」

    clip-vit-large-patch14

    Rename 「model.safetensors」 to 「clip-vit-large-patch14」

    ComfyUI\models\vae

    ae.safetensors

    vae-ft-mse-840000-ema-pruned.ckpt

    ComfyUI\models\controlnet

    ttplanetSDXLControlnet_v20Fp16.safetensors

    diffusion_pytorch_model

    ComfyUI\models\instantid

    ip-adapter.bin

    Description

    https://drive.google.com/file/d/1_HPGG2iMAyovS3jO8ubhti4YIcOhOLUj/view?usp=drive_link

    InstantID does not support the latest version of ComfyUI. Please download the old version. I deleted the model and kept the nodes to ensure that there will be no errors when running (5.62g)

    FAQ

    Comments (55)

    2882001j227Oct 17, 2025
    CivitAI

    the drive is in view mode
    couldn't able to download the custom nodes

    2882001j227Oct 21, 2025

    @gsk80276 thanks i download it

    2882001j227Oct 21, 2025

    Can I upload a video and do a character video from it , beacause i dont see any video upload option here

    gsk80276
    Author
    Oct 21, 2025

    @2882001j227 No, the workflow is I2V

    kjj432Oct 22, 2025
    CivitAI

    Can somebody list which models should i get for 3090 and nsfw?

    dontforg3t2137Oct 22, 2025
    CivitAI

    i keep getting the background of the video reference instead of the background from my image. node i need to disable?

    gsk80276
    Author
    Oct 22, 2025

    Yes, you need to turn off the mask and background

    markopolo791Oct 23, 2025

    @gsk80276 what is the point of that background thing then when it stops generating after the background process is finished?

    dontforg3t2137Oct 24, 2025

    @gsk80276 i seem to be doing something wrong. possible to send me screenshot?

    gsk80276
    Author
    Oct 25, 2025

    "AdaptiveWanVideoAnimateEmbeds" cannot support the latest "WanVideoSampler". There is a problem with it. It cannot be imported into Dwpose when generating videos in segments. The characters in the video will not move in sync with the reference video. I have replaced it.

    dontforg3t2137Oct 28, 2025

    @gsk80276 thank you for your help

    kudon44Oct 26, 2025
    CivitAI

    Testing this and so far it is really good compared to other I2V. The fact I can do 10+ seconds with this is especially nice. I haven't tried 20s or longer yet though, but plan to. This is a huge improvement to many of your other workflows which require nodes and stuff we can't easily get. This one was easy to install and you provided easy links to all but one file which was no longer available (fortunately I have).

    I do have one issue I'd like to see fixed or if you can tell me how to fix it as Google couldn't help me finding the answer...

    How to fix the nodes that have Japanese language connections to English? A portion of the nodes have the connecting dots in Japanese still so I can't modify this as easily since I don't know what they say.

    Also, is there a way to use some of the few odd loras that have an A/B part on a single phase like A & B high, then just a single low? I've seen a few like the cowgirl that does this and I don't see how to apply it in your workflow.

    gsk80276
    Author
    Oct 27, 2025

    「How to fix the nodes that have Japanese language connections to English? A portion of the nodes have the connecting dots in Japanese still so I can't modify this as easily since I don't know what they say.」

    What is the workflow you use?

    kudon44Oct 27, 2025

    @gsk80276 Wan 2.2 Kijai I2V. I also ran into another two issues. I can generate 10s video on my system with your default 480x832 resolution... However, I cannot generate beyond that on my RTX 4090 and 64 GB RAM. Beyond that it freezes always at WanVideo Sampler (#2 - 0%). Even worse, when I tried to do 10s at 1280x720 (as it has no upscaling as far as I can tell) it crashed multiple drivers on my PC including GPU driver during the processing.

    I thought context windows was supposed to let you do 81s every group, indefinitely, for however many frames you put but it would just gradually degrade right? If so why doesn't above 10s work and why does it fail at 1280x720 when others work for 81 frames at that resolution?


    I feel like I'm misunderstanding something here...?

    gsk80276
    Author
    Oct 27, 2025

    @kudon44 https://drive.google.com/file/d/1dAmPWfKPE9uThYIrHe2hHgDYpBvkpXBn/view?usp=sharing
    There aren't many Chinese characters in my workflow.

    gsk80276
    Author
    Oct 27, 2025

    You can use tiled_vae and WanVideoBlockSwap to reduce system consumption.

    gsk80276
    Author
    Oct 27, 2025

    I have a 4090 48GB GPU, tested 1280*720 161 frames per second, WanVideoImageToVideoEncode used 26GB of VRAM, WanVideoSampler used 30GB, and the model was KJ FP8.

    gsk80276
    Author
    Oct 27, 2025

    Using tiled_vae, block swap 30, I kept VRAM usage to 20GB

    v1vecOct 28, 2025
    CivitAI

    With the default Wan2.2-Animate.json as the workflow, running it would only 1. load the video - 2. perform dwpose - push to video combine - 3. perform sec segmentation - push to video combine - 4. perform blockify mask - push to video combine. It seems that the image is never loading, nor does the WanModelLoader and thus no generation at all, but getting the pose / masks / blocked masks videos all the time (seems that only preprocessing workflow has been running). Was I missing anything? Thank you.

    v1vecOct 28, 2025

    just checked the logs, I was missing the detection .onnx files. Adding those would resolve this issue.

    gsk80276
    Author
    Oct 28, 2025
    TheBodyHunterOct 29, 2025
    CivitAI

    Good Job brother !
    It will be more amazing if you can make it render faster and have Upscaler :3 .
    Thank you very much for your amazing work ! <3

    gsk80276
    Author
    Oct 29, 2025· 1 reaction

    Considering that some users' computer systems are not so good, I decided not to use it. Regarding speed, since using Sageattn is not easy for beginners, I will keep the default settings and only use Lightx2v.

    TheBodyHunterOct 29, 2025

    @gsk80276 Thanks again bro !

    BTW , the result is very good , i can keep my original style of my images and make it work like magic . Please add an Upscaler brother !

    maaadeNov 1, 2025
    CivitAI

    Cannot execute because a node is missing the class_type property.: Node ID '#483'

    I am getting this error for wan animate. What to do?

    gsk80276
    Author
    Nov 2, 2025

    Download the ZIP file and extract to custom_nodes.

    ComfyUI\models\custom_nodes

    auto_wan2.2animate_freamtowindow_server

    https://drive.google.com/file/d/1wy8syijO7aKPCEIiWl2S-gIWJ9wxWrNM/view?usp=drive_link

    maaadeNov 2, 2025

    @gsk80276 I already did that.

    gsk80276
    Author
    Nov 2, 2025· 1 reaction

    After updating the nodes using ComfyUI Manager, I encountered no errors. You can download another version of ComfyUI that I prepared for testing (the model has been deleted, but usable nodes are retained).

    https://drive.google.com/file/d/18geDYU9h3_F8bkVOvyri78i751_BzVOJ/view?usp=sharing

    nurdyoga878Nov 4, 2025
    CivitAI

    I keep getting the error:

    WanVideoAnimateEmbeds

    'Tensor' object has no attribute 'get'

    Is there any workaround for this?

    gsk80276
    Author
    Nov 4, 2025

    After updating the nodes using ComfyUI Manager, I encountered no errors. You can download another version of ComfyUI that I prepared for testing (the model has been deleted, but usable nodes are retained).

    https://drive.google.com/file/d/18geDYU9h3_F8bkVOvyri78i751_BzVOJ/view?usp=sharing

    KaidokuNov 9, 2025
    CivitAI

    I'm using the wan 2.2 animate flow, I loaded all the models, uploaded a video and uploaded an image, ran it and I see it pulls the mask and allows me to adjust the green points, but how do I now make it replace my image with the person in the video? It doesn't seem to complete, just shows the masking spots...I'm guessing I need to connect something but I can't figure out what.

    gsk80276
    Author
    Nov 9, 2025

    https://drive.google.com/file/d/1MlxsUvZdKRg_5GCPEYW-e6j4PO-QvO5Y/view?usp=drive_link

    After selecting the green dots, adjust the values ​​of the BlockifyMask node, depending on the desired size of the black square for the mask. Then, rerun COMFYUI to generate the video.

    RL1775Dec 27, 2025

    I think you had (or are still having) the same issue as me. There's a node hidden behind the one for SeC Video Segmentation called ONNX Detection Model. It needs to be populated with the .onnx files you downloaded (they go in ComfyUI\models\detection, if you don't have that folder, create it). Then the workflow should run correctly.

    IliarAINov 11, 2025
    CivitAI

    Thank you very much for sharing, the animate workflow is working very nicely. One thing I cant get to work though are the hair, I always get the original hair instead of the ones from the image input. Any way to fix that?

    gsk80276
    Author
    Nov 11, 2025

    Adjusting the size of the black block through nodes

    FreddinNov 14, 2025
    CivitAI

    sorry for the dumb question but i have to ask: This is not likely for a 3060 12gbvram I am guessing. Nevertheless amazing, wish i could run this but i am still learning and this might be a bit much :D .. one day.

    gsk80276
    Author
    Nov 14, 2025

    If only 81 frames are generated, using the block swap node in the Kijai workflow can save VRAM usage. Enabling tiled VAE during encoding and decoding should also be achievable. Note that at least 64GB of memory is recommended.

    FreddinNov 14, 2025

    @gsk80276 Good to know. I do have 65Gb of ram and recently have been able to get good results with this card and 12gbvram. I will definitely be giving it a try.

    creatorjulie743Feb 6, 2026· 1 reaction

    I was able to run the Infinite Talk workflow on my RTX-3060 12GB. 480 x 832 at 4 steps (takes about 7 minutes) and 512 x 896 at 5 steps (takes about 11 minutes). Each was 5 second long. Set blocks_to_swop to 40 and ran. The GPU VRAM was used at about 45-50% and the RAM at 50-60. I do have 128GB of RAM, though, and use the 12GB GPU for AI work only, the monitor is hooked up to another GPU.
    Pretty much satisfied with the results. Next step, will try to make longer clips.

    markharper80266Nov 16, 2025
    CivitAI

    Getting the error

    WanVideoUniAnimateDWPoseDetector

    cannot convert float NaN to integer

    gsk80276
    Author
    Nov 16, 2025· 2 reactions

    Try adjusting the score_threshold of the WanVideoUniAnimateDWPoseDetector node to 0.

    PanagosNov 22, 2025· 2 reactions
    CivitAI

    Keep getting this error:
    WanVideoAnimateEmbeds
    shape '[1, 8, 4, 104, 60]' is invalid for input of size 205920

    i use the latest version of comfy in mimic pc

    markharper80266Dec 1, 2025

    Same issue trying to run the animate workflows. I run into the issue using the authors provided comfy package as well as my own.

    bgyhnddr716Feb 24, 2026

    same issue with my own comfy

    ArtPencilDec 2, 2025
    CivitAI

    After reading the comments, it looks like it's not even worth trying. LOL. I'll pass.

    xiongvang12851285Mar 28, 2026

    Works fine for me.

    ceelogreen855Dec 9, 2025
    CivitAI

    I am trying to put my image's face onto the video but the face that generates is different. I have changed values for pose_strength and face_strength (up and down for each) but nothing works. Does higher face_strength mean the picture is weighted more heavily and will show up in results more? Does video resolution have an impact? If I am only trying to change the face should I disable detect_hand and detect_body and enable detect_face?

    AsianEngineer_877Dec 27, 2025
    CivitAI

    im using the wan animate template and im stuck on propagate video. Any fix?

    RL1775Dec 27, 2025

    Is it creating the animation and pose videos just fine but not actually generating an output video?

    AsianEngineer_877Dec 28, 2025

    @RL1775 it gets stuck at around 20% during sam segmentation

    AsianEngineer_877Dec 28, 2025

    @RL1775 I've got 32gb ram that i suspect has something to do with it. When doing the segmentation ram spikes to 75% and stops there.

    RL1775Dec 28, 2025

    @AsianEngineer_877 yeah that's probably the reason. 32Gigs is a significant choke point for anything running Wan 2.2 unless you're using something like a 5090 32gb edition. I've run out of RAM using this workflow with 64gb on a 5080.

    obsidiancloudFeb 18, 2026
    CivitAI

    AdaptiveWanVideoAnimateEmbeds cannot find this node to install anywhere

    Workflows
    SDXL 1.0

    Details

    Downloads
    728
    Platform
    CivitAI
    Platform Status
    Available
    Created
    10/8/2025
    Updated
    4/28/2026
    Deleted
    -

    Files

    wan22Animate_instantidSupir.zip

    Mirrors

    wan22Animate_instantidSupir.zip

    Mirrors

    Huggingface (1 mirrors)