This is the SVI 2.0 PRO version.
v3.5 12-section fixed - An error was discovered where the prompt that should have been entered in "Section 8" was incorrectly duplicated with the 6th prompt. This issue has been fixed, so please reinstall the workflow.
Thank you so much to @MarcanOlsson for discovering the issue!
v3.5 12-Sec - This version is based on v3.5 and enables video merging through 12 generation steps.
When actually generating the video, you will notice color shifts compared to 6 generation steps. While 6 steps generally maintain better quality in practice, the ability to specify 12 steps offers advantages depending on the application, so we decided to add this version.
The 12-Section version enlarges the workflow, so unless you need more than 7 generation steps, I recommend using the standard v3.5.

v3.5 - The subgraph specification in the model input area has been deprecated and reverted to the v2 specification. Additionally, it is now possible to generate videos for only the first section.
We received multiple reports from the community that models such as CLIP and VAE were not functioning correctly due to the subgraph, and we also received feedback that the model placement was unclear. Therefore, we decided to revert to the v2 specification.
However, the subgraph of the generation section, which includes the sampler, remains unchanged. We believe that performing the generation process within the subgraph serves to prevent a decrease in generation quality. While the model area issue is simply a layout issue, the subgraph cannot be removed because it affects the quality of the generation section. If there are issues with the subgraph itself, please avoid using this workflow.
Regarding the video generation for only the first section, given the nature of SVI, we initially omitted it, believing that a single generation was unnecessary. However, we received feedback from the community requesting that a video be generated for each section, and that videos be added gradually while reviewing the generated videos. This was a very logical approach, so we added the "first video" and modified the workflow to allow videos to be accumulated while keeping the seed value fixed.
v3.4 - Layout adjustments.
v3.3 - Changed the seed node from "CR Seed" to "Seed (rgthree)". This change was made to align with commonly used custom nodes in this workflow, following reports of implementation issues with CR Seed.
v3.2 - Modified the layout to make it easier to disable Lightx2v Lora.
v3.1 - Modified the layout to make it easier to disable the Sage Attention node.
v3.0 released.
Video length can now be changed in each of the six generation sections, providing more flexible control over video content.
The frame rate (fps) was previously fixed at 16fps, but can now be changed arbitrarily. Accordingly, the RIFE-VFI node's scaling factor can now be changed in the input area.
GGUF model loader is now included as standard.

Version 2.0 changed the number of generation sections to six.
The layout has also been updated, allowing Seed node input to be processed in one place. Furthermore, the layout has been significantly redesigned to unify the user experience with Painter I2V versions, reducing the input burden. With this change, the wildcard prompt input method has been discontinued.
Please note that the explanations in this workflow are solely my personal opinions. I do not have expertise in AI generation, so some information may be inaccurate.
The main goal of this workflow is to achieve compact operation when performing repeated generation. It minimizes screen scrolling during operations such as prompt input, input image selection, specifying time, number of steps, resolution, and, most importantly, LORA selection. To further enhance compactness, all nodes are fixed to prevent accidental operation.
Links to the Models and LORAs and nodes used in this workflow
SVI LORA :
Wan Advanced I2V (Ultimate) :
https://github.com/wallen0322/ComfyUI-Wan22FMLF
This node was updated on January 27th, but the version available for installation from ComfyUI Manager may be an older version. While the older version will still work, you won't be able to set "SVI Motion Strength," and you'll likely experience more color misalignment. Therefore, if you can Git clone, we recommend installing the latest version.
Links to the Basic models of the Wan2.2
CLIP:
https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/tree/main/split_files/text_encoders
VAE:
https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/tree/main/split_files/vae
CLIP Vision :
https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/tree/main/split_files/clip_vision
You can generate up to six generations, and each generation is assigned a unique seed value. Normally, click "Randomize Every Time" to display "-1". Generation will be random. In this case, the seed value for each generation will be displayed at the bottom of the screen. If you want to fix the seed value, click the seed value field or enter the seed value directly. For example, you can fix the first and second generations and regenerate the third generation and beyond randomly. However, regenerating a section before the generation section you want to fix will change the final frame, so you cannot fix subsequent sections. As a general rule, regenerate after the section you want to fix.
By combining the six generated videos, you can create six different types of movement. For example, by generating and combining six videos of different durations, you can create a long video containing six complex movements. This is one of SVI's strengths, enabling complex processing that is impossible with a single generation.
However, SVI V2.0 PRO also has its drawbacks. Because SVI uses the first image as a reference point, the AI tries to restrict movements that deviate significantly from the reference point. As a result, the movement becomes sluggish and unnatural. Furthermore, this constraint imposed by the reference point also reduces the responsiveness to prompts.
In short, the use of excellent LORA is essential in SVI. In my experience, movements without LORA are very unnatural, lack impact, and resemble something out of a horror movie. Fortunately, there are many excellent adult-oriented motion LORAs available. However, if you want to create completely original movements, expect it to be difficult with the current version of SVI.
I hope this workflow helps make video production with SVI more enjoyable.
Description
v3.5 12-Section:
This version is based on v3.5 and enables video merging through 12 generations.
FAQ
Comments (132)
Hi kenpechi, thanks for sharing this for free. I am no expert, so I am having some problems: I use a RTX 4090 on Windows, and initially I couldn't run your workflow because I didn't have the module 'triton'. However, talking with GPT, it said that it is a module available for Linux and not Windows; it suggested me to bypass the nodes Sage_Attention, Patch_Torch_Settings and the NG ones. I did it and indeed I was able to generate: however, the character changes a lot during the video, making it too much different from the original picture. Is it because I bypassed those nodes? Is there a way to fix this on Windows? Thanks.
I'm not sure if what I'm about to tell you is correct, but I'll share what I've learned from my experience.
First, things like Sage Attention affect the processing speed of the video generation process, not directly affect the quality of the video.
Incidentally, I'm a Windows user, but I can install Sage Attention, so the idea that it's only usable on Linux is outdated information.
Video quality varies greatly depending on the type of model, the custom nodes used, prompts, LORA, etc., so it's impossible to generalize, but at least bypassing Sage Attention only increases the time it takes to complete the generation; it doesn't significantly affect the quality. In fact, these techniques can be said to sacrifice quality for speed.
Therefore, the creation of videos that seem completely unintended, as you mentioned, is due to other factors. However, there are so many possibilities that it's extremely difficult to pinpoint one here.
Model WanVAE prepared for dynamic VRAM loading. 242MB Staged. 0 patches attached. Force pre-loaded 52 weights: 28 KB. Killed
CPU: Intel Core i5-14600K
GPU: NVIDIA GeForce RTX 3090
RAM: 32GB DDR5
When asking someone a question, be polite.
Well, even if you ask me politely, I still don't know what kind of error it is.
I was just trying to keep it brief and technical.
I like your work, by the way.
When I run your workflow, I get this error:
"Model WanVAE ... Killed"
Specs:
RTX 3090, 32GB RAM
Do you know what might be causing it? It seems like this workflow might not be suitable for RTX 3090 — would you agree?
@e7lew It seems the generated data isn't fitting into the RAM. My understanding is that the RTX 3090's 24GB of VRAM and 32GB of RAM should be sufficient, but with the 30 series, there might be issues with data offloading or other factors.
Depending on the video length and resolution, replacing "VAE Decode" with "VAE Decode (Tiled)" might allow for splitting and avoid the "Killed" message.
Try going into the subgraphs of each section and replacing "VAE Decode" with "VAE Decode (Tiled)".
Hi,
thank you for this workflow — it’s really well made and easy to use. I appreciate the effort you put into it, and it’s been working great overall.
I noticed something during testing:
When I use all 6 scenes, the output is correctly exported as MP4 (H.264).
However, when I use fewer scenes (for example 2 or 3), the output changes to MKV using FFV1 instead of MP4.
I wanted to ask:
Is there a way to force the workflow to always output MP4 (H.264), regardless of the number of scenes used?
Thanks again for your work — really appreciate it.
@e7lew This workflow uses the "~st video" node, which corresponds to the last number of sections being generated, and enables it in the fast bypass node. All other Video nodes are disabled. This should output everything in mp4 format.
If it doesn't, modify all Video sections to H264/mp4.
after 5m I get below error,16GB VRAM 50 series, 96GB RAM
DLL load failed while importing _fused: The specified procedure could not be found.
This could be a troublesome error.
It might be related to version inconsistencies in PyTorch or other components.
A DLL is a high-speed processing file used by Python, and it seems to be saying that it couldn't find a component called "_fused".
However, ComfyUI installs a large number of programs, so it's difficult to pinpoint exactly where the incompatibility is occurring.
Ideally, you should put the log file of the entire generation where the ComfyUI error occurred into Gemini or chatGPT and ask for solutions, but in my experience, this only solves the problem about half the time.
If similar errors occur in other workflows, a complete clean install of ComfyUI, including PyTorch, might be necessary.
However, I'm sorry, but this error is too difficult for me to understand, so I might be saying something incorrect.
Thank you for sharing this wonderful workflow to public. Just wondering, what model do you use for generating the initial image? Just curious cuz that penis looks really alsome compairing to some normal t2i models.
Lately, I've been using the "Wai Realism IL" model from Illustrious. I've posted some examples in the "Image" section of my profile.
Thanks for your workflow. Great work!
I like your PainterI2V workflow very much and how does it compare to this? I mean I need to generate a long video like this workflow. But I do like the PainterI2V more.
Which one would you recommend? Or the PainterI2V can be extended by tweaking the parameter?
Sorry for asking question across different workflows.
Like you, I prefer the Painter I2V version. However, it is possible to combine videos in a workflow similar to the SVI version. In fact, I've tried generating it three times. But even with the current two-times method, the transitions between videos are problematic; doing it three times results in a video that's frankly incomprehensible.
In short, to create long videos, the SVI version is currently the only option. The great thing about SVI is that it seamlessly and naturally connects video segments while maintaining character consistency. However, as you've also noticed, the motion is very poor.
Therefore, many people who use the generator combine the cut videos using other video editing software. However, in this case, you need to ensure that the same character images are used throughout.
In other words, achieving excellent motion in a long, single-shot video is currently difficult.
Ultimately, it's a choice between length or motion.
@kenpechi Well, I agree to all points. I've got tone of storage space for archiving but I am getting lazy haha, so I will try the SVI first. Thanks anyway!
Thank you. I was able to get up an running in a couple of minutes with the instructions in the workflow. It is logically organized and easy to figure out. Works great.
I'm shocked, this is the best wf I ever found, solid result, no error, easy use, thank you !!!
NOOB question - where are the videos output?
If no changes are made, I think it will go into ComfyUI/output/video.
@kenpechi hhmm... it's not there and it's not in the usual folder...
@Mr_mr1 It seems the "save_output" setting in the "Video Combine" node isn't set to "true". Check the "~video" subgraph for each number and make sure the "save_output" setting in the "Video Combine" node is set to "true". If it's "false", the video won't be saved at all.
I was gonna say, 'your mother's facebook account'
@kenpechi Yeah, that worked... but now randomly nodes are missing...
@Mr_mr1 Something unusual is happening to you. However, in most cases, it's because your settings are fundamentally wrong.
Please try various things, including other workflows.
If the problem persists, please provide more specific details, including the situation.
You could really clean this up by switching LoRAs as strings. LoRA Manager has nodes specifically to handle the string/combo data. I hid all mine in a subgraph and let the prompt call the model. God what an awful mess WAN is. LTXV has just completely blown it out of the water. Months of fiddling with SVI were completely wasted. Ah well. High noise is still useful for pulling some motion guides in, but otherwise...
You don't need to waste time here. Do what you want to do, the way you want to do it.
@kenpechi I wasn't insulting your workflow. It's very pretty. I'm really proud of you.
it works, but...
It ocassionally just exits. How, why & can I prevent?
i.e.
Patching torch settings: torch.backends.cuda.matmul.allow_fp16_accumulation = True
100%|████████████████████████████████████████████████████████████████████████████████████| 3/3 [03:15<00:00, 65.21s/it]
Patching torch settings: torch.backends.cuda.matmul.allow_fp16_accumulation = False
Requested to load WAN21
loaded partially; 4366.60 MB usable, 3894.05 MB loaded, 9737.39 MB offloaded, 472.55 MB buffer reserved, lowvram patches: 845
Patching torch settings: torch.backends.cuda.matmul.allow_fp16_accumulation = True
100%|████████████████████████████████████████████████████████████████████████████████████| 4/4 [04:46<00:00, 71.59s/it]
Patching torch settings: torch.backends.cuda.matmul.allow_fp16_accumulation = False
Requested to load WanVAE
loaded partially; 175.01 MB usable, 152.23 MB loaded, 89.78 MB offloaded, 22.78 MB buffer reserved, lowvram patches: 0
(venv) H:\ComfyUI>
no warnings or errors.
This is what's known as OOM (Out of Memory). The OS is forcibly "Killing" (terminating) the process without displaying an error log because it doesn't have enough system memory (RAM).
The simplest solution is to replace the "VAE Decode" node with "VAE Decode (Tiled)".
Go into the subgraphs of each section and try replacing the nodes. "VAE Decode (Tiled)" splits the VAE processing, making it less likely to terminate prematurely. Please try it.
@kenpechi Managed to make it work, thx. Turned out it was eventually not so much the VAE decoder issue. Maybe also. But eventually increasing my pagefile.sys did the trick. Although, only for 1 run, any subsequent run still fails but I guess that's a(nother) cache clearing issue.
This is the best workflow so far.
this workflow has more input controls than houston command center
That's true.
I need help, I am missing a lot of models and Lora, I need the download links for these files
The main models and LORA are listed on this page and in the annotations of my videos, right?
@kenpechi Hello, I know what is causing the problem. There is no issue with my LoRA and model; however, the connections are missing, and I don't know how to connect them. I need help.
@Greedy_Pineapple9399 I don't really understand what you're saying, and I don't understand the situation, so I can't help you.
Hi Kenpechi, are the base models, Lightx2v and SVI models replacable? I tried wan2.2_i2v_high_noise_14B_fp16, wan2.2_t2v_lightx2v_4steps_lora_v1.1_high_noise and SVI_Wan2.2-I2V-A14B_high_noise_lora_v2.0 but all videos generated became a mess like sea surface.
Replacing Lightx2v LORA is possible, but finding the appropriate strength will be extremely difficult.
I think the only LORA that works well with a strength of "1" is probably the one I'm using.
There are countless suggested methods for configuring Lightx2v LORA. A typical example is the 3-Sampler setup, which also includes methods that bypass Lightx2v LORA and use a separate Sampler.
Therefore, I think there are LORAs that are incompatible with my 2-Sampler method. So, I think it will be a difficult process.
I don't know much about SVI LORA.
@kenpechi Does the base model have any impact on this? You're using fp8_scaled in this workflow, but I replaced them with fp16.
@kenpechi Now I see. The Lightx2v and SVI MUST NOT be replaced, but replacing the base model doesn't matter.
@Nothingtosatisfy I think various base models should work. For fp16, you should be able to fix it by changing the text encoder.
The noise that looks like sea surface is probably from Lightx2v Lora.
Hey there, thanks for sharing the workflow!
Would you mind sharing how long it takes for you to generate 6 sections with and without SageAttention? I've read that SageAttention2 is supposed to give a 30-40% boost but I barely notice a difference, so I'm not sure if there's a problem on my end (i.e. running out of memory) or if the performance difference is not that big for that workflow. I'm running on a 4070 Ti Super 16 GB VRAM/32 GB RAM/torch2.8/cu128/Win11
I haven't specifically measured the generation time with or without sage attention, and it also varies depending on the resolution and video length, so I can't give you a definitive answer. However, it's unlikely that sage attention is ineffective in this particular workflow.
In fact, it seems that sage attention isn't functioning, but I don't know why.
seriously really awesome workflow. Very happy with it. 1 Q though, in order for my pc to run 6 steps on your base resolution, I needed to increase my pagefile.sys to 128gb. That's on my main ssd though, how harmful is that for daily use? GPT said I'm wearing pretty swiftly through it that way (<10months or so), that true?
I'm sorry, but I don't even know much about pagefile.sys, so I have absolutely no idea.
Can 3060 12G VRAM run smoothly
You may run out of memory.
Please implement general memory shortage countermeasures. For example, you could add a node to refresh VRAM after the VAE Decode node, or replace the VAE Decode node with a (Tiled) version. There are other methods, but you'll have to figure them out yourself. I won't teach you how.
@kenpechi Okay, thank you.
i have exactly this and with sage and fp16 acc. it runs fine even at the predefined size, i find that upscaling and interpolation are usually the culprints of ooo so if i really want it to generate without the risk of ooo i only do interpolation and upscaling in the end if and if i really like the result.
Great workflow, i just started to use it and wondering why my videos are not getting merged. Is it intentional ? (before i start opening all the sub graphs) :)
Thank you again, great job, very easy and nice layout @kenpechi
Thank you so much for the amazing workflow! I am using it really well and truly appreciate your great work!
I just have one simple question: When the character blinks or changes expressions, their face slightly changes and becomes different from the original face. How can I prevent this and keep the face consistent?
Thank you again for your masterpiece!
I think the problem lies in the timing of the video transitions, so the best approach is to generate it multiple times until you get it right.
I almost never get it to work on the first try.
By the way, I don't know what base model you're using, but consistency varies depending on the model. I won't mention it here, but there are some models that change frequently, so I wanted to let you know.
@kenpechi Thank you so much for your honest and incredibly helpful reply!
It brings me a lot of relief to know that generating multiple times is part of the normal process, rather than a fundamental flaw in my setup. I will definitely experiment more with batch generations and try out different base models like you suggested!
I truly appreciate you taking the time to share your insights. Your workflow is still an absolute masterpiece to me! Thank you again!
@strong4leaf330 Honestly, if you possibly have a character LoRa for the figure being animated (even if it's T2V), I found that more reliably helps to keep the face from deviating much between the different clips/scenes.
My videos aren't as smooth and coherent as yours. Should I just use the workflow prompts directly?
The prompts should be adjusted precisely depending on the reference image.
I previously watched a video from someone who reported that it wasn't working for them, but the composition of their reference image was quite different from mine. It's no wonder it didn't work with the same prompts as mine. However, SVIs are inherently difficult to implement. I've only posted the ones that worked after many failures, so repeated attempts are probably important.
I forgot to mention, LORA is also important. Some versions work well with SVI, while others don't. Try different ones.
@kenpechi thx
This workflow is great, one of the few that is easy to install all needed nodes. Thank you!
Question: should the prompts be modified for each video step? In your workflow the prompt is the same in all 6 steps. But should it more like 1) A woman sits down 2) A woman takes her shirt off 3) A woman touches her breasts. Etc. ?
I'm not sure which video you're referring to, but try setting all six prompts to different values, specifying different LORA values, and specifying different timestamps. This should allow for more flexible and complex movements.
Hi Kenpechi, I tried searching for these LoRAs on Hugging Face and Civitai, but I couldn’t find them. Could you let me know where to download them? Thanks!
1st step-lora high and low:
NSFW-22-H-e8.safetensors/NSFW-22-L-e8.safetensors2st step-lora high and low:
Wan2_Undressing_-_V1.safetensors3st step-lora high and low:
BreastRubv2_HighNoise.safetensors/BreastRubv2_LowNoise.safetensors4st step-lora high and low:
same as 1st never mind
5st step-lora high and low:
Body-Cumshot-Pullout-HIGH-v1.safetensors/Body-Cumshot-Pullout-LOW-v1.safetensors6st step-lora high and low:
same as 4st never mind
I've found most of the LoRAs, but there are still a few left that I couldn't find. Could you please help me? Thank you so much!
the list:
Body-Cumshot-Pullout-HIGH-v1.safetensors
Body-Cumshot-Pullout-LOW-v1.safetensors
@kenpechi Thank you very much!
@MachuPichuMan thank you!
I can't find BreastRubV2 loras
@qqianyu
https://civitai.red/models/1913617/wan22-breast-play-i2v?modelVersionId=2448070
Didn't you notice that I specifically displayed LORA in the video? You should look at things more carefully. I'm not exactly beginner-friendly, because I learned through trial and error, drawing on the many helpful resources from more experienced users. You should put in more effort. Don't expect me to answer beginner questions like this again.
This is a solid implementation of SVI. Interesting technique, using ultra-short segments and leveraging the strengths of the LoRA to limit degradation while maintaining the motion. I do think LoRA Manager is indispensable here. I'd advise anyone who has it to switch the LoRA loaders and leave them empty, using the stack inputs and setting up pre-set combinations from which to choose. If you've only got three or four combos alternating throughout a dozen stages it's much easier to set all instances at once, wherever they may be in the flow. The other huge benefit being the auto-complete and popup previews you get from the manager nodes. Same goes for prompts - planning out presets for switching is annoying but goes a long way.
The workflow itself is super clean, certainly as clean as a SVI setup with so many stages can be. I had no problems at all testing it out with the defaults. The usual rules apply. The better your model and input image, the better the result. If you need ggufs to fit the models in memory the swapping will probably take forever. 6 stage 16 second clip at 720x1280 took the XMP ~9 minutes. Same input takes 2 minutes for LTXV, and that has sound and sync, but the WAN motion is much more dynamic. This is where WAN excels, and it's what SVI is so good at preserving when doing extensions. This workflow be good at that.
interesting experiment - took 6 step prompts from missionary sex video and 6 step prompts from blow job video. put them in a 12 step prompt workflow. extended stages to 5 secs. got ~45sec video of pretty good stability.
do the upscale works at the section where the video is output? or it had be select every section even no video is output?
Please enable this only for the sections where the video will be played.
There are multiple options so that they can be easily bypassed.
any suggested settings for 'patch sage attention' and 'model patch torch settings'? i have rtx4090 with 24gb vram. running 800x1024x16fps only uses ~50% vram, spills ~27gb to shared dram... any settings to keep more in vram? love the workflow. thx.
I apologize, but I'm not familiar with the details of your point.
However, as far as I understand, ComfyUI currently processes data in roughly the same amount of time whether using VRAM or DRAM.
@kenpechi thx. just wondering if you had settings for those nodes that worked well for you.
@tedbiv I have only tried it with sage attention set to "auto" and patch torch setting set to "true".
@kenpechi ok, thx
This is so solid Worfklow, I use rtx 4070 super with this crazy Optimization i got 5 sec with Model WAn Remix 2.2 V3, with uspcaling and 64 FPS movie in 300 sec!!!!! 480p to 1440p Crazy Broo. :D 7 steps use SVI without Lighting lora. Like sage is on and torch. Really Thankfull for such a MAsterpiece. <3
Great process. Everything works, but for some reason the video output is noisy. Look at the photo in the link. I clonong(WAN Advanced I2V (Ultimate):) What do you think could be causing this noise? https://ibb.co/JwrjpJN9
I suspect the settings (weights, high/low) of your LightX2V Lora model are incorrect. Or perhaps you're adding another LightX2V Lora model to a base model that already has one merged into it.
What do you think?
@kenpechi im getting some pinkish texture, and i think i didn't install the power lora at all (but it didnt show any failure)
For the DR34ML4Y, i cannot find your exact match, should i get the t2v or i2v for your sample here?
https://civitai.red/images/126523989
@homuragremory594 I've shortened and changed the name of LORA. DR34ML4Y is the V2 of the most popular one.
By the way, that pinkish noise is almost always caused by inconsistencies in the model settings. It often happens to me when I first try a new workflow. A quick check reveals a setting error, which is easily fixed.
For example, the wrong type of SVI LORA, incorrect weight values, or wrong SD3 shift values.
@kenpechi Thanks, i did not change any values, i took the workflow and parameters directly by downloading the video and adding it to comfy ui. But it looks like there is still a slight pinkish hint and the videos are not lining up together. im not sure if it is due to seed change, but i did turn off sage attention, which is the only changes i made, and i took a screenshot of the first frame of the photo as the starting image. i will try saving the first frame of the video and check again
@kenpechi Guys, I figured it out. First, I increased the Power Lora Loader h/l values to 1.0 — all the artifacts disappeared, but a green tint was still there. Then I carefully read the comment from the respected @homuragremory594 about the “incorrect SVI LoRA type.” I was using the regular version, and after I downloaded the Pro version from the description, everything started working properly.
Conclusion: you need to carefully read the description instead of trying to rush things.
Sorry for causing confusion. @homuragremory594 is just amazing — creating such an efficient workflow takes real genius!
@solder912219 That's great! Have fun, by kenpechi. lol
@kenpechi @solder912219 i got it too! i downloaded the SVI from the github page instead of the link provided here, which likely caused some generation issue.
Thanks for sharing this workflow kenpechi!
Мне выдает ошибку и пишет отсутсвует ffmpeg , что бы сохранить видео , что делать ?
Please translate this into English.
@kenpechi It gives me an error and says ffmpeg is missing to save the video, what should I do?
@edward18881973 Look inside all the "~video" subgraphs. For some reason, the "format" of the "video combine" node is set to "ffmpeg-gif". Please change this to the more common "h264-mp4".
@kenpechi ProcessLookupError: ffmpeg is required for video outputs and could not be found. In order to use video outputs, you must .. I have already changed it to what you wrote, but it still gives me this error. It's like there's no video output.
@edward18881973 Have you checked all the video combine nodes? There are many of them in this workflow.
If that's still not the problem, then I don't think the error is related to this workflow, since I don't have ffmpeg installed. I can't think of any other possible causes besides the ones mentioned above.
When i try to download this workflow, i see this error
"This XML file does not appear to have any style information associated with it. The document tree is shown below.
<Error>
<Code>NoSuchKey</Code>
<Message>The specified key does not exist.</Message>
</Error>"
I can't figure out how to customize video length? Any help?
In this workflow, the length is specified in seconds, not frames.
Enter "5" seconds in the "Second" field in the "Input" area. The number of frames is automatically calculated by multiplying this number of seconds by the "fps" value.
@kenpechi amazing!! One more question. In the workflow, it says nsfw22 lora. Which one is that? Is that the general nsfw one? The one with the most likes and downloads?
@TheSemenDemon That's right, it's the most popular "(WAN 2.2 experimental) WAN General NSFW model".
hello, great workflow. i have a question. my first and second video generates fine, but when the first video is transitioning to the second video, there is this jerking ghosting effect. tried playing around with the overlap nodes and the overlapping type (cut, linear, ease), but i cant seem to find a way to smoothen the transition between the 2 videos
resolved, i wasnt using the pro SVI low lora lol. too many models to keep track
do you have any plan to add audio later? ;)
It's not zero, but I'm not particularly interested at the moment.
Hello, i'm getting the error:
# ComfyUI Error Report
## Error Details
- Node ID: 1046:1429
- Node Type: KSamplerAdvanced
- Exception Type: torch.AcceleratorError
- Exception Message: CUDA error: invalid argument
Search for `cudaErrorInvalidValue' in https://docs.nvidia.com/cuda/cuda-runtime-api/group__CUDART__TYPES.html for more information.
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1
Compile with TORCH_USE_CUDA_DSA to enable device-side assertions.
Some times at firts section, sometimes at second section
Any advice?
Have you implemented Sage Attention? While it's highly likely the error is related to the Sage Attention node, I can't say for sure. Have you tried disabling it?
@kenpechi yes i have disabled Sage Attention, i tryed disabling several nodes, and seems like disabling CFGZeroStar nodes finally make it work, thank you!
Btw, pro-tip, leverage wildcards on the prompting part as well. So first you just build the 6/12 prompts, replace them all by wildcards mentioning different expressions, makes a huge difference in creating a bit more variety.
Do you have any Loras for the faces of the people or are they random? They look pretty but I don't wanna generate videos, I only want to generate images with them.
I don't use FaceLORA. However, I don't publicly share my face generation method. I think it's best to experiment and find what works for you.
Is it possible to include a node where I can attach the final output images?💡
Of course, it's possible.
I think I asked the wrong question. I actually meant a 'last frame image' node, not the 'final output' node... My apologies.😑
If possible, I would appreciate it if you could consider adding this to the next version of the workflow.🙏🏻
@GentlePudding I wouldn't post here specifically for that purpose, but for example, my painterI2V workflow also has a node for extracting the final image, so you just connect that after the VAE Decode node and then connect it to the Save Image node.
Many other experienced users have incorporated the same technique into their workflows, so I highly recommend trying to build one yourself.
Still a newbie at local ComfyUI, but I'm studying hard! I'll look into your suggestions. Big thanks to the Kenpechi Workflow for making the start so much easier!👍🏻
Which Python version is most compatible with the nodes? I’m getting dozens of LoRA key errors in the logs, and the video feels like it’s stuttering or freezing. I’ve reinstalled ComfyUI dozens of times, but I still can’t find the cause of the problem.
Which Python version is most compatible with the nodes? I’m getting dozens of LoRA key errors in the logs, and the video feels like it’s stuttering or freezing. I’ve reinstalled ComfyUI dozens of times, but I still can’t find the cause of the problem.
Unlike A1111, ComfyUI is tolerant of Python versions and works fine with versions 3.10 to 3.13. Incidentally, I'm using 3.12.10. The LORA key error occurs because you're using Lightx2V Lora for WAN2.1, but this is normal. Don't worry about it.
In short, the cause of your problem is something else, but it's too vague to explain. Please describe the situation in more detail.
I keep getting this error and idk how to fix it:
Cannot read properties of undefined (reading 'output')
I was previously getting a similar error with 'workflow' instead of 'output',
The error with ComfyUI could be caused by using an outdated version of a custom node.
All I can say is to update all custom nodes to the latest version. I don't know anything else.
There is some weird connection going on between 6th_clip and 8th_clip. If I load a Lora with a prompt in 6th it happens again in 8th, even if I don't have the lora or prompt. Similarly, if I load a Lora with a prompt in 8th it will happen in 6th, even tho it have a totaly different prompt and loras.
I understand that this workflow is prone to the kind of problems you've pointed out.
However, I re-checked for the problems you mentioned, but I couldn't find any factors causing problems with the 6th and 8th steps.
@kenpechi Thanks for checking. I mainly noticed it when I have Combo HJ/BJ in 6th and Deepthroat in 8th, without it being in the prompt she move her hand up and start jerking.
@MarcanOlsson The node connection of the workflow may have changed from its default state. Does the same problem occur if you download this workflow again?
@kenpechi I downloaded it 3 days ago, same problem. Only difference from the original workflow is that I changed some loras around and clip length. 6th and 8th are both 3 seconds long. Don't know if that affect anything.
@MarcanOlsson Thank you! I figured out the problem! The 8th prompt wasn't being reflected, and it was loading the 6th prompt instead. I'll post the corrected version right away.
太强了,密密麻麻的节点与连线,惊呆了
流程太强了,如果你也感兴趣,真的应该试试,在尝试的过程中学习,非常推荐
I beg you to share your image generation workflow for your starter images. So high quality and consistent.
I don't use any special workflows. I use very common workflows.
However, I create images using two workflows. The images posted on my profile are generated using one of these workflows.
After that, I generate them again using another workflow. In other words, it's image to image generation. However, I don't disclose how I do that.
There are two main reasons. First, the image generation process is a unique part of my videos, so disclosing it wouldn't be interesting. Second, disclosing it would require explaining the process, which is quite troublesome. It's not difficult, but it's a tedious process.
I disclose over 90% of my generation process, but wouldn't it be more appealing if there was a little bit of mystery?
It's a really powerful workflow, spend a lot of time on it and haven't ever had anything better!
Out of curiosity though: It feels the switching between clips can still be a bit 'paused' if you will. Clip1>pause>clip2>pause>clip3>pause>... Was wondering if you'd have any ideas/suggestions on how to reduce that.
In the case of I2V, because it starts from a reference image, it's almost always impossible to start from a moving state; it has to start from a stationary state. Even if you use a 5-frame overlap with SVI, if it takes more than 5 frames to start from a stationary state, it will appear as if it's stopped.
I think this is a fundamental mechanism of wan2.2 I2V, so it's difficult to solve.
In short, if the action starting from the reference image is kept within 5 frames, it shouldn't be noticeable.
In my generated videos, the blow job exceeds 5 frames, so there's a brief pause where the penis is removed from the mouth, but the sex scene is kept within 5 frames, so it's not very noticeable.
I think it's necessary to compose video scenes taking these realities into account.