Turbo LoRA trained on preview3. Use CFG 1 and 8-12 steps. Works well on other Anima checkpoints. You can decrease the LoRA strength a bit below 1 for more variety.
Work in progress. Increases stability, anatomy, and style consistency. Details are a bit worse than base.
Description
FAQ
Comments (41)
works well. thx for posting.
Have you ever thought about collaborating with @reakaakasky regarding turbo LORAs as their CFG distilled and DMD2 RDBT - Anima Checkpoints and LORAs are great. And have been great for all the preview versions. Especially for stability, consistency and overall quality.
This lora is amazing. One standout for me is how well it preserves the influence of artist tags. With other Anima acceleration loras I’ve tried, the effect of artist tags became noticeably weaker after applying them, but this one doesn’t seem to suffer from that issue that much.
Some comparisons between 3 dmd2 models. link. cfg 1, steps 12, euler, seed 1,
turbo v0.1, definitely the winner of "style consistency" and has best colors, imo. However, it seems to prefer generating images with a more orange-ish tint.
rdbt, more green or gray-ish tint, it's also a finetuned model so I can't tell the style. But it has best details.
cosmos r64, seems to have the best prompt adherence and very stable,
from experience I can tell you that the cosmos-dmd LoRA is good for mixing with the others
so the real User Preferred Speed Distillation setup is some bizarre mix of them all :' )
I love RDBT, but I hate how bad it is with styles. If the prompt is somewhat short like 30 tokens, it basically ignores the artist tag completely, and it seems even more prone to do so if you prompt something that is NSFW. But even with longer prompts it's still a major issue, aswell as things just looking very washed out.
Maybe it would look a lot more faithful to style if I went crazy like 1000 tokens or something, but I shouldn't have to just for the sake of getting the style I want. I have no idea why RDBT does this when style is basically the most important factor when genning.
It's still quite fun to use if you don't care about style and want something that looks more visually pleasing (other than washed out colors) faster with what appears to be a bit better quality.
I might use it every now and then for casual fun SFW memes or cute stuff. But the instability is just insane. I had a relatively simple prompt where it had "photorealistic" tag and it worked as expected, then I add in a couple NSFW tags and boom, SD1.5 style lol.
Now if only anima could do some basic furry concepts. I really hope they decide to include E621 into the training data, or at least make another version with it in.
There's so much damn potential with Anima, and to think we're only at a preview version, not a finished model or a finetune of the finished model.
I prayed for this day
我为你推出这个lora而喜悦
It's great, but it looks like it accelerates the early steps a lot and the model doesn't do much at later steps losing a lot of the prompted artstyle.
Have you considered training two layered loras? First one just for late steps (say 0.6->1) and then a second one (training 0->1) on top of the base+late_lora?
Which sampler and scheduler should I use? Is it the same as when I'm not using a Turbo LoRA?
Almost the same
Works quite nicely with trained Loras and other checkpoints. Getting good results at 1cfg 10steps
Problem - CFG 1 disables the negative prompts, and you might NEED those to improve the output. Care to make a Lightning lora instead? That sped up outputs on SDXL while still keeping CFG low and low enough for negative prompts.
CFG 1.1 will probably work fine and allow negative prompts. You could also try NAG or one of the other no cfg negative methods.
There's Normalized Attention Guidance, NegPiP (in comfyui-ppm), or making your own silly setup by running the turbo LoRA at a lower strength and compensating with https://civitai.com/models/2364703?modelVersionId=2684678
some portion of "anima_preview_rdbt_v0.6_cfg_distilled_only" on top
@VeerGeer Tried NegPip, 'Anima' object has no attribute 'text_processing_engine
@VeerGeer "Normalized Attention Guidance" don't work for Anima
@chaosleges sounds like the text encoder isn't being loaded in as expected
https://civitai.com/images/128368192
a workflow with it working looks like that, at least with the nodes from this repository;
https://github.com/pamparamm/ComfyUI-ppm
This seems to be the only distill that preserves artstyles well, very impressive.
Works surprisingly well with 6 steps CFG1:
3 steps with Euler/simple lora@1
3 steps with gradient estimation/beta [email protected]
Best LoRa.
No greenish-swamp tint like in RDBT.
No artifacts in detailed objects or eyes or hair or leafs like other Turbo LoRas.
No washed out white low contrast output like Cosmos Predict LoRa.
Minimum alteration in artist style.
Dramatically drop down fantacy of Anima.
Is there a good anima workflow or is the base one in comfui for anima usable with this?
I'm using this one: https://civitai.red/models/2426853/anima-preview-workflow
If you play with shift below 1.0 and add some natlang descriptions to your prompt that describe the visual features of the artist style you use, then you can actually get output similar to non-distilled Anima. You can furtther finetune the texture with eta and s_noise on a sampler node to find the right balance between those and shift and your scheduler. Very impressive.
Could you share your workflow ?
Unbelievable Turbo solution. Anime enjoyers are eating good right now...
遵循提示词的效果比较差,或者说需要更详细的描述才能达到同样的画面,细节确实少了不少。但是将时间缩短了起码五倍,这很牛逼
the results definitely get worse, details are lost and color is washed but what surprised me is that the result is still consistent and faithful.
Same here, less details, less color, sometimes limbs are not complete but significantly faster gen speeds. Hands on hip, but no forearm just floating wrist and palm on hip. CFG1, 12Steps.
Had i started with turbo i would have been satisfied but after experiencing what Anima is capable of, I'm just too spoiled to be satisfied with turbo. Perhaps I'm doing something wrong, perhaps its just what it is.
提示词遵从度和画面细节确实打了折扣,但是速度确实快了不少
As far as I can tell so far Preview 3's complex natural language prompt adherence is actually SIGNIFICANTLY better when using this lora than when not using it.
Three things I now know(?) after trying this out:
1. I don't understand AI at all.
2. Magic is real and this is proof.
3. It Just Works o_o
Okay, jokes aside, I'm actually flabbergasted how effective this lora is. The only words I can use to describe the way it fixes anatomy and composition is... aggressively stable.
And, like, I stress-tested it.
Hands and fingers refuse to break even when confronted with challenging concepts (like holding chopsticks or multiple hands interacting) and being partially occluded. I can count on the hands of one finger the number of instances of sixth digit or fused fingers I've managed to spot so far.
Perspective and composition also stay coherent at most of the camera angles and multiple views variations I tried. I have seen a few anomalous generations, but the overall consistency rate is still impressive.
Even painterly style loras and artist tags (which are notorious for destabilising anatomy) work surprisingly well without getting overwhelmed by the style of this lora.
If this is what preview 3 is capable of with the right tuning, I can't wait to see how far Anima still can go beyond this point >o<
The hands looking good is a result of high accuracy on anima, this basically gives anima a mini version that is trying to replicate anima in the lightest way possible, instead of using all the data in the big model. This makes it less complex, which means it takes less steps to be more accurate. This comes at the cost of the model needing to essentially listen to itself more than it listens to your prompt, and thus being less capable at creating more complex, abstract concepts that aren't common in the data the model was trained on.
@CocoLarge Ohhh... I... still don't know if I get it. So the smushing of the Turbo lora into fewer steps results in better consistency, at the cost of constrained creativity? So it's a choice between more complex generations that allow for errors, or more rudimentary ones that don't do anything that might result in errors?
@Derpression Not exactly. Because the turbo lora makes the model want to reach a final image quickly by simplifying data, rushing towards the training data and then asking itself if the final image looks like the training data. This method of asking itself if the image looks like the training data conflicts with cfg and can cause problems. Which is why the instructions for turbo/lightning/dmd2 ask for you to use cfg 1 (cfg 1 is equal to not having cfg turned on). The problem is, cfg helps the model follow your prompt more closely. Errors aren't really a part of this equation. Errors can either be a thing that just happens randomly, or the result of not having enough steps (aka low accuracy).
The simplifying of the data results in the model knowing less.
so if cfg is at 1, that means negative prompt will be ignored? please let me know.
Yes unless you use NAG (Normalized Attention Guidance): https://github.com/BigStationW/ComfyUI-NAG-Extended/tree/main
@AnimaXx i guess the workflow that im currently using might be a challenge for me to edit the nodes, rip. . .
You can set it to 1.5, and it will still work and look largely the same (although, from what I understand, long negative prompts might not work reliably?). If some parts of your negatives still get ignored, pushing it up to 2.0 might help. Beyond that, you might start noticing some fried contrast.
My images look like shit when I follow the instructions. Please help


