CivArchive
    LTX-2 NSFW Text Encoder - Gemma 3 12b Abliterated - v1.0
    NSFW

    Gemma 3 12b Abliterated LTX 2 compatible text encoder

    just put into the text_encoder folder and use

    hello Reddit! appreciate the love https://www.reddit.com/r/StableDiffusion/comments/1q8yxcv/comment/nysz6ea/

    Description

    initial upload

    FAQ

    Comments (150)

    Ragamuffin20Jan 9, 2026
    CivitAI

    so the lack of examples is a red flag this probobly dosnt work but ima give you the benefit of the doubt and test it out.

    mastercstesterJan 9, 2026

    Testing it right now.

    azra1lJan 9, 2026· 1 reaction

    this is just the gemma 3 12b model with removed guard rails for censoring. but the image model is trained with the censorship in place so it doesn't know anything about this uncensored stuff, so best case nothing will change, worst case the text encoder output will be a pile of gibberish to the image model. but i've never tried using nsfw clip models myself, so maybe i am wrong..

    ultimo_intento
    Author
    Jan 9, 2026

    lol, i appreciate the faith. Right after I posted we lost power here, so I'll upload more examples soon. Also, Civit has to review all nsfw photos before posting in case you aren't seeing any at all.

    It does make a difference in my testing, but I wouldn't call it dramatic. Biggest boost is in i2v gens. Otherwise, t2v, it does nipples better and in general I like the nude bodies better though it makes writing less reliable. Nonetheless, I'm uploading it now because people on reddit requested specifically a safetensors abliterated version of the text encoder.

    mastercstesterJan 9, 2026

    @ultimo_intento I'm also seeing this. I'll have videos posted soon too.

    brendacaroJan 9, 2026· 1 reaction

    Either way it doesnt seem to hurt much. I'm getting quality results

    Degenerator123Jan 9, 2026· 5 reactions
    CivitAI

    This is only useful if you are using the LLM as a prompt refiner right? As a text encoder creating embeddings it doesn't "refuse" anything, abliteration will not suddenly teach the model nsfw concepts.
    Not discouraging people from testing, but I'd be surprised if this would really do something.

    Ragamuffin20Jan 9, 2026· 6 reactions
    CivitAI

    well i can confirm this dose jack shit for text to video

    edit: some people are getting results from text to video so it might just be me not prompting right or something which is good to know but ima wait till we get an NSFW checkpoint before i start using it tbh

    PrefoJan 9, 2026
    CivitAI

    Is this based on this version?

    https://huggingface.co/mlabonne/gemma-3-12b-it-abliterated/tree/main

    I tried that one out already and didn't really notice it making nsfw better in LTX2. But too early to tell.

    zengrathJan 9, 2026

    his version didn't work for me. but this one did. downloaded it with git clone same way as other models. Like others say it doesn't do a whole lot since underlying model won't do NSFW.

    ultimo_intento
    Author
    Jan 10, 2026

    @zengrath correct the model layers have to be edited to make it compatible

    and actually it can do nsfw, I have a full time job so I don't have all the free time for goon R&D though look on reddit

    zengrathJan 9, 2026
    CivitAI

    Doesn't work. Says, "TXVGemmaCLIPModelLoader

    embed_dim must be divisible by num_heads (got embed_dim: 1280 and num_heads: 12)."

    I have no errors using the recommended ones. are you sure you uploaded it right? I'd rather get it from HF but not sure this is compatible model

    ultimo_intento
    Author
    Jan 9, 2026· 1 reaction

    Use the "LTXV Audio Text Encoder Loader" available in the ComfyUI workflow.

    brendacaroJan 9, 2026· 2 reactions
    CivitAI

    Everyone is asking if this will make any difference. It might, not sure about that, but one thing using this is a lot more trouble free than the original gemma 3. Just use this and you will avoid alot of technical issues that others are dealing with.

    BrAInB0tJan 9, 2026· 11 reactions
    CivitAI

    Will only make a difference right now for training. LTX-2 is incapable of generating NSFW because the video component of the model was never trained to do so. Video from uploader is fake.

    Lora_AddictJan 9, 2026· 1 reaction

    lol it's not fake, it's just i2v

    ultimo_intento
    Author
    Jan 10, 2026

    absolutely incorrect

    AivanjoJan 9, 2026· 12 reactions
    CivitAI

    Cool, thanks!

    I don't understand what everyone's complaining about. Guys, don't expect miracles, the model is completely censored even at the training stage. But this is the first step toward getting rid of the censorship. Hopefully we'll see some NSFW loras soon.

    In the meantime, we can experiment ourselves. Renting the necessary equipment from Runpod isn't that expensive. If there are guides for training Lora on LTX 2 that are understandable to most people, then I believe that with our persistence we can break through this wall of censorship!

    WhatTheGuyJan 9, 2026· 2 reactions

    wouldn't call it 'completely censored', just undertrained. You can prompt for boobs, you are getting boobs. They just aren't great, but the model knows what they are. Of course WAN is absolutely superior in that department. But I think the base it good enough to get good results with finetuning it ;)

    AivanjoJan 9, 2026

    @WhatTheGuy 
    If the base model can learn new (unknown) concepts well, then this is easily overcome, and NSFW loras can fix it. And it looks like that's exactly what will happen.

    g1263495582Jan 9, 2026· 1 reaction

    @Aivanjo For me, LTX-2 is just for messing around. As for NSFW, we have to wait and see how good the upcoming LoRAs are. Not to mention the i2v quality is straight-up horror material, plus the model is super sensitive to prompts.

    kkkdsadasJan 9, 2026· 2 reactions
    CivitAI

    I've tested it. It makes kissing between two people smoother

    Ronny0815Jan 9, 2026· 14 reactions
    CivitAI

    Hope there is an FP8 version.

    desm0ntJan 9, 2026· 19 reactions
    CivitAI

    Pls do a fp8 variant.

    undertheblacksunJan 10, 2026

    Does it not work with fp8?

    fox23vang226Jan 9, 2026· 2 reactions
    CivitAI

    I cant use this, it breaks the audio on my generations. Ive heard that the model is very picky on an encoder that is trained directly on it or not. Kinda like Z-Image turbo, all the merged/trained models are not compatible with loras trained on Z-image turbo base.

    ychristian008402Jan 9, 2026· 14 reactions
    CivitAI

    please do Q4 GGUF. thanks!

    ychristian008402Jan 10, 2026

    @razzz thanks!

    ychristian008402Jan 10, 2026

    @atomtheunbreakable thanks!

    Actually, in the short term an fp8 variant would be useful. Eventually the GGUF nodes will be able to handle gemma3 but atm the stock fp8 on Distorch is all I can run.

    ultimo_intento
    Author
    Jan 10, 2026

    they aren't directly compatible with ltx nodes@razzz 

    https://huggingface.co/DreamFast/gemma-3-12b-it-heretic/tree/main/comfyui
    This is an abliterated fp8 encoder. Haven't tried it. LTX-2 seems horrible on character consistency so I'm going back to Wan for now.

    raidou88Jan 9, 2026
    CivitAI

    i have a 5060ti 16gb vram and 32 ram, and can't make it work, i tried a lot of encoders and workflows and all give me different problems. i hope they start working soon on quantizied models. If anyone knows a good workflow let me know please

    g1263495582Jan 9, 2026

    What problem?

    CelebCreatorJan 9, 2026

    increase the swap file in windows to 70GB

    PartisanoJan 10, 2026

    it's too messy right now (remember Wan when it came out?) Just wait a week or two, it'll all be settled by then

    raidou88Jan 10, 2026

    @Partisano yeah you are right, one tends to want to test the new thing as soon as possible. but sometimes we must wait haha

    raidou88Jan 10, 2026

    @CelebCreator i tried doing it the pagefile up to 64 and had some problems too it said something like werfault.exe error . but i decided to just wait

    MalekthaJan 10, 2026

    @raidou88 download pinokio and find wangp there, it has ltx2 running. Ima doing 720p 15 seconds with only 12gb ram

    aureliusJan 10, 2026

    I got mine to work with --reserve-vram 4 --lowvram

    xaocchaos423Jan 10, 2026

    Try to change "run_nvidia_gpu_fast_fp16_accumulation.bat" file: .\python_embeded\python.exe -s ComfyUI\main.py --windows-standalone-build --fast fp16_accumulation --lowvram --cache-none --reserve-vram 8

    LacidarionJan 11, 2026

    I am using the stock comfy workflow and have pushed out 512 frames no problem on a 3090. One thing I have learned... NEVER use all of these "best ever" workflows you find on YT. The stock comfy workflows are always the best ones.

    NFTGamer666Jan 12, 2026

    Yeah the very FIRST step with LTX2 is memory config otherwise you are starting off on a bad foot. The startup ComfyUI bat file mentioned by xaocchaos423 is what you need. For the --reserve-vram x section, replace x with your total GPU VRAM minus 4 e.g. 16GB will be 12

    Edit: Regarding the --reserve-vram flag, take aurelius' advice in the post below.

    aureliusJan 12, 2026

    @NFTGamer666 That is a very dumb idea. --reserve-vram 12 means you're giving up 12GB of VRAM, as in you're not using 12GB. With a 16GB GPU, it means you're only using like 2.5GB of VRAM for the generation. What you want is reserve as little as possible without running into OOM, which is probably like 1-4 GB. So the smart thing is to put --reserve-vram 1-4, or however little you can do without crashing, starting from like 0.5 and then working up from there.

    Reserving 12GB of VRAM means your speed is suffering badly because that VRAM is not used. Your config turns your GPU into something generations weaker. Open Task Manager while generating, and you'll see that only 2-ish GB of VRAM will be active with your config.

    NFTGamer666Jan 14, 2026

    @aurelius It makes sense. It doesnt make a difference for me though. I can still generate a video in about a minute.

    aureliusJan 15, 2026

    @NFTGamer666 Yes, it does make a difference for you. I can guarantee that. You're not immune to the effects of VRAM.

    NFTGamer666Jan 15, 2026

    @aurelius Ok

    spammer666666497Jan 9, 2026
    CivitAI

    I see someone using 10 step wan 2.2 model with this. How the hell did you manage to do that? That specific wan model is thr best there is for wan when it comes to nsfw.

    friendlyfriend4000673Jan 10, 2026· 1 reaction

    how about you read the post before posting a retarrded comment

    spammer666666497Jan 11, 2026

    @friendlyfriend4000673 retarded has only 2 r's. Retard.

    vsaanJan 11, 2026

    @friendlyfriend4000673 username does not check out.

    Baka_OppaiJan 11, 2026

    no one is using wan with ltx its just a video attached to the input node. usually loaded by "video node (Upload) VHS".

    g1263495582Jan 9, 2026· 10 reactions
    CivitAI
    ArtfulGenie69Jan 10, 2026· 1 reaction

    Could use the huggingface gguf space to gguf it too I bet.

    ultimo_intento
    Author
    Jan 10, 2026· 1 reaction

    mlabonne, but it's not a direct conversion. the layers have to be edited to make it work for ltx

    milbyJan 10, 2026· 1 reaction
    CivitAI

    are you doing v2v to add audio?

    ultimo_intento
    Author
    Jan 10, 2026· 1 reaction

    native LTX gen audio

    TeqnoBlasterJan 11, 2026· 18 reactions
    CivitAI

    LTX2 is neat but im gonna get nightmares like 1 of out of 100 are good and well lets say i just watched a girl rub her hand off and it stuck fingers out on her asshole, got sucked into it then her whole body snapped. but at least she said it felt good lol

    MalekthaJan 11, 2026· 4 reactions

    i need to wash my monitor now xD

    sexgod1979Jan 13, 2026· 1 reaction

    I had one of those too, truly terrible to see

    ss9999Jan 15, 2026· 2 reactions

    New method of censorship: Give people enough body horror and they'll stop asking.

    BananaUnitedJan 11, 2026· 1 reaction
    CivitAI

    The restrictions on voice output have been reduced, but it seems like ambient sounds aren't being output properly.

    GrimmsterJan 11, 2026· 16 reactions
    CivitAI

    The example video isn't very impressive. What is better, exactly?

    Triple_Headed_MonkeyJan 11, 2026· 2 reactions

    I'm sorry did you even look at those nipples xD

    spammer666666497Jan 11, 2026· 1 reaction

    I think it stops or bypasses Gemma from censoring, or it tells Gemma to go ahead with inappropriate prompts. Gemma is strict.

    g1263495582Jan 12, 2026· 4 reactions

    @spammer666666497 In fact, Gemma doesn't censor when used as a text encoder. If it did, it would cut words or change the prompt. Take Qwen 2.5 VL (used in Qwen Image Edit) as an example: it enforces censorship, so if you prompt it to 'remove clothes,' it won't generate nudity. In Gemma's case, the result actually matches the prompt. But regardless, for good NSFW quality, you still need to rely on a LoRA anyway.

    stduhpf893Jan 12, 2026· 1 reaction

    Abliteration is a process that makes the LLM unable to respond something like "I'm sorry, I can't help you with that". When it is used as a text encoder, I'm not sure it achieves anything.

    g1263495582Jan 13, 2026· 1 reaction

    @stduhpf893 If it's running inference, it would respond like that. But if it's functioning as a text encoder, it won't.

    ylvlylvuyv374212Jan 13, 2026· 1 reaction

    @g1263495582 okay I meant in terms of nsfw prompts, it gives gemma the green light but ltx-2 either it heavily refuses the do certain motions when when gemma gives a prompt or its just not trained on nsfw motions.

    ss9999Jan 15, 2026· 1 reaction

    @stduhpf893 I don't know either but I have to say it's very irritating when models do things like put people into shorts (despite very clear prompting about them being nude) as part of the baked-in censorship.

    dragon509127Jan 12, 2026· 1 reaction
    CivitAI

    This may end up being unrelated to the text encoder, so I'm sorry if it's off topic, but anyone have any thoughts on why the audio might play but the video doesn't speak? As in it's not generating with the person in the video talking, they just stand there. The first test I did worked fine, after that all I did was lower the number of frames to take it from a 10 second video down to a 5 second one and I started having issues

    Edit: Ok it stopped working cause I had a bit of language in the second prompt that must have gotten it flagged as NSFW, so the outcome of that is it still generates the audio but doesn't sync the video to it. Not sure how people are managing to get actual NSFW audio with it

    FemBroJan 16, 2026· 1 reaction

    The second I saw the shit was using Gemma by Google I knew it was going to be bullshit.

    dragon509127Jan 16, 2026· 1 reaction

    @FemBro Actually in retrospect I'm not sure what caused the initial issue but since then I've had no problem getting nsfw audio out of it, despite gemma supposedly being locked down. 🤷‍♂️

    solshicllyfvhexgdf977Jan 12, 2026· 3 reactions
    CivitAI

    Don't waste time, it will not render nsfw

    aureliusJan 12, 2026· 1 reaction

    Are you telling me those football-ass abomination breasts it makes are SFW? 🤣

    sexgod1979Jan 13, 2026· 1 reaction

    I beg to differ :)

    mobdik17378Jan 13, 2026· 21 reactions
    CivitAI

    LTX looks like a MASSIVE step backwards from Wan, can someone convince me otherwise?

    dragon509127Jan 13, 2026· 4 reactions

    I think the ability to have spoken audio in the same workflow for locally generated stuff is appealing to people, from what I've seen to get that you generally have to pay to use models like kling to do the same otherwise. I'm probably ignorant of other options though

    g1263495582Jan 13, 2026· 4 reactions

    That's not very surprising. LTX-2 has 14B+5B parameters—the 14B is for video and the 5B is for audio. As for Wan2.2, it has 27B parameters but uses a Mixture of Experts (MoE) architecture with 14B active parameters (per expert), switching roles between motion generation (High noise) and refining (Low noise).

    darkshang0Jan 13, 2026· 7 reactions

    @dragon509127 In comfy, I can do custom audio/lipsync in WAN(Wan+humo//infinitetalk). Because of this, I'm not seeing value in LTX2...yet. But if I didn't have comfy, I'd definitely see this as an upgrade over Ovi.

    NFTGamer666Jan 14, 2026· 9 reactions

    Ill take a drop in video quality anytime for combined video and audio. LTX2 FLF workflow plus Qwen Image Edit with the multi angle lora is unmatched atm. The biggest hurdle now is your imagination and creativity...oh and your prompt mastery :D

    MrReclusive666Jan 14, 2026· 5 reactions

    ltx is a pretty big step back from any other video model out now, only perk, it does audio, but it reaaaly kind of sucks video quality wise.

    MaximilianPsJan 14, 2026· 4 reactions

    Quality is worst but the speed and the audio sync are amazing also, it runs on just 10 gigs of VRAM!

    MrReclusive666Jan 14, 2026· 2 reactions

    @MaximilianPs thats a plus, but i think one of the worst things about this model, is even if you have 24gb vram, it doesn't get better.

    MrReclusive666Jan 14, 2026· 2 reactions

    @MaximilianPs also, it does have length going for it, wan struggles at 7 seconds, i can push this to 20, it looks like crap, but it will do 20 seconds, lol

    ss9999Jan 15, 2026· 2 reactions

    @MrReclusive666 You can try SVI 2.2 with Wan

    wqn999Jan 15, 2026· 10 reactions

    Those who like it will naturally like it. This is the world's first open-source audio-video synchronization multimodal model. There is only this one. This title was originally supposed to belong to wan2.5. Unfortunately, what happened later, as everyone knows, although it still has many problems now, we all believe that the power of the community and the ltx2 team can solve them. We look forward to the next generation of lxtv! Thank you very much to ltx for its open source, and to all the developers in the community for your contributions. Thank you!

    MrReclusive666Jan 15, 2026· 2 reactions

    @wange999 yeah, this is huge, it has its issues visually, but the addition of audio and audio sync is MASSIVE! with the community being behind it so much, we should get some good stabilization in video soon, i think its biggest hangup is that its designed for such HIGH resolution, which is fine if you have hardware for it, but, even on 24gb vram that resolution is a bottleneck so you want to run lower res and then upscale, but right now, it falls apart at lower res, "720p" is its lower limit right now, and even it has issues, but pushing something like 1600x960 or higher looks great, for those 3 seconds you get.

    give me a slightly smaller model, with the audio, but more geared toward lower res so it can resolve the lipsync at lower res better. or do what wan did, give me 2 models, one for motion, one for detail.

    but to each there own, i see huge potential in this because of the addition of audio, but if it wasn't for audio, wouldn't even think twice, and its not even its ability to generate audio that I want, its its sync with external audio that is huge for me.

    MaximilianPsJan 15, 2026· 1 reaction

    @MrReclusive666 12 seconds "Prompt executed in 395.01 seconds" with 10 VRam gigs

    MrReclusive666Jan 15, 2026· 1 reaction

    @MaximilianPs its not bad, I know that, and if I had my shit wired differently, I probably wouldn't mind things as much, or if I wasn't so impatient, lol.
    my last gen, some stuff im testing, 10 seconds, 1280x640, finished in 162.15 seconds, but my problem is thats 23.5gb ram offset with distorch ontop of the 24gb vram I already have, so 47.5gb for that generation, and yeah, not that big of a deal, but the 4090 is on a 1x pcie riser so huge performance hit there for data transfer.
    this is a 3d rendering rig, 4080 in the case, 2 4090's external. i blame nvidia for removing nvlink, other wise the 4090's would be nvlinked and Id have 48gb vram on "one card"

    ss9999Jan 16, 2026· 2 reactions

    @MrReclusive666 The way Wan (2.2 and later) splits the model (mixture of experts I think is the terminology) is a must, in my opinion, since Nvidia, AMD, and Intel refuse to provide enough VRAM in consumer high-end GPUs.

    MrReclusive666Jan 16, 2026· 4 reactions

    @ss9999 yeah, its called moe, and yes, I agree, needs to become the standard now for opensource models if we are going to be able to use them.

    g1263495582Jan 16, 2026· 1 reaction

    @MrReclusive666 It's about different mindsets. Western models focus on being user-friendly so anyone can use them, but quality takes a hit. Chinese models go for MoE and flexibility, but they're complex and harder to train. In the end, it really depends on the user's skill level and how much they know what they're doing.

    chrisbraeuer41172035Jan 16, 2026· 3 reactions

    wasted 3 days already with trying to get something useful out of ltx2.

    MaximilianPsJan 16, 2026· 1 reaction

    @MrReclusive666 Oh I work with much smaller resolution like 400x1200 or similar. Event 400x800 

    MrReclusive666Jan 16, 2026· 2 reactions

    @MaximilianPs yeah, i tried, I wanted to, but couldn't get the model to perform decently at all at low res, only high res seems to work (for its lipsync and all that), I wanted to do in low res first then upscale, but can't get ltx to operate at low res.

    IthilgondJan 18, 2026· 5 reactions

    So far... It's better and faster than anything else I've tried. I'm running on Pinokio's Wan2GP, with a RTX4090 - 64GB Ram. 90 seconds to create a 16 second video at 720p. https://civitai.com/images/117507883

    agentgerbilJan 18, 2026· 3 reactions

    LTX is more heavily dependent on your prompts. You need to be a prompt wizard to make the most of it.

    MrReclusive666Jan 18, 2026· 1 reaction

    @agentgerbil its a lot like hyv1.5 prompting, it needs a lot of guidance, but you can make it do things it wasn't trained on, I did a sex scene in a car with no lora's. i think thats why ltx recommends that slow ass llm prompt enhancer.

    MaximilianPsJan 19, 2026· 1 reaction

    Is this only intended for nude and porn content? I'm testing a workflow using this text‑encoding model for NTX2 in ComfyUI, where a wolf is supposed to consume a bird. The expected behavior is to trigger a bite action and make the bird asset disappear with a small 'puff' of feathers, but the event isn’t firing correctly. Even the NTX2 demo prompt—where the frogs are expected to ingest the fly during the meditation sequence—is not functioning as intended, and the frog refuses to eat the fly.

    agentgerbilJan 19, 2026· 2 reactions

    @MaximilianPs looking at the submissions below, seems it works for more than just pr0ns

    ss9999Jan 20, 2026· 2 reactions

    @g1263495582 No, in the end it comes down to Nvidia's monopoly on prosumer GPUs and AMD's complicity. We should have 128 GB GPUs for reasonable prices but instead we're still being sold 32 GB at a premium. Since it doesn't look like Intel has any interest in doing anything about the monopoly situation I don't expect things to improve, particularly since RAM is also in a shortage situation.

    MrReclusive666Jan 20, 2026· 7 reactions

    @ss9999 while i agree mostly, 128gb is excessive for anything consumer, even prosumer. we are in a weird transitional state in computer hardware. the XX90's are essentially titan's, those aren't intended as everyday use gaming cards, but are being used as such, effectively killing the prosumer market, titan's/XX90's are prosumer cards, the "a" series are pro cards, H's are server/workhorse cards, then of course when you switch from RTX to A, you leave the cheaper GDDR7 behind and go to HBM, still not worth the cost, but they charge that cost because they can, for those cards, they don't give a shit about us because business's will pay the price.
    is the pricing astronomical, yes, it is, that needs to change, in the 5090, the gddr7 is roughly $10 a gb, thats 320 for vram. so yeah, pricing is fucked, but there's the other side of it, the gpu's are expensive and problematic to make, in a lot of cases, the "lower model" gpu's are made from failed chips from the higher models, they just deactivate the dead parts and mark down prices. not trying to defend them, because thats still no excuse they can't bring back nvlink or just through more vram on the card at cost, thats just greed. we know nvidia is greedy, and amd doesn't have much to say about that right now, nvidia rules the AI space.

    ss9999Jan 21, 2026· 2 reactions

    @MrReclusive666 128 GB is only excessive if you're Nvidia. For AI work, it's the minimum reasonable amount for 2026 products costing $2500 and up. No one is talking about HBM.

    blonkybonkJan 21, 2026· 2 reactions

    If you have low VRAM then it's incredible that it can do what it does in the speed it does. If you can already run better models then this probably has no use for you.

    MrReclusive666Jan 21, 2026· 1 reaction

    @ss9999 the ai work we are trying to do, is not prosumer though, this is professional grade requirements, its amazing we can even do it at all. and for ai, we are talking hbm, ai workflow works so much better with hbm. we are in a transitional state in hardware requirements for the things we do at home. a lot of us have been though this before, at one point 500mhz single core cpu with 64mb ram was normal, then things shifted, next thing you know we are surpassing 3ghz with 4gb ram, things shift, we are in one of those, the shift is always expensive for early adopters, most people who do this, arnt doing it on consumer hardware, consumer hardware isn't made for this, its highly inefficient for it, its why most pro cards have lower tdp then consumer cards yet have higher specs, they are built for different things, but gaming also sucks on pro cards.

    but again, its not to say nvidia can't make high vram low cost cards, it would literally cost them nothing to put 64gb on a 5090, yes mostly greed, but part of it is marketability, yeah, we want it, we are in 1% of users though, the average user wants 4060 specs, that's the most used card right now, followed by the 3060, volume vs cost. my biggest complaint was them dropping nvlink from consumer cards.

    ylvlylvuyv374212Jan 13, 2026· 4 reactions
    CivitAI

    For those pulling their hair out when both characters say the same word or the wrong person says something what helped is when I switch CFG on model 1 from 1.0 to 3.5

    sexgod1979Jan 13, 2026· 1 reaction

    thats a good tip, thanks

    dreammoto12Jan 14, 2026· 1 reaction
    CivitAI

    Is it possible to use it with Wan2GP?

    mouselabber332Jan 14, 2026· 1 reaction

    I'm trying now, but I'm pretty sure that you can take the original encoder, backup it somewhere, put this one and name it the same.

    dreammoto12Jan 14, 2026· 1 reaction

    @mouselabber332 I tried to replace the encoder. But when I replace it, I get incompatibility errors.

    TophnessFeb 11, 2026· 1 reaction

    We would need an equivalent quant for it. Raise an issue to deepbeepmeep. There are also tutorials to make quants yourself, just match it with the existing quant type WAN2GP uses (q4_0-unquantized_quanto_bf16_int8)

    MonzoniuzJan 14, 2026· 1 reaction
    CivitAI

    doesnt work at all for me. "text encoder not loaded"

    R240Jan 16, 2026· 1 reaction
    CivitAI

    when you prompt your character to pull down their pants or underwear it doesn't work

    yawox22499972Jan 16, 2026· 13 reactions
    CivitAI

    this model is garbage for nsfw

    Agent_SmthJan 28, 2026· 1 reaction

    Then the day comes when the monkey, who previously thought the banana was rotten, takes a closer look on some results around.
    And so the monkey, together with all its other know-it-all friends who liked its comment
    oo-oo-aa-aa the banana is garbage”
    discovers something it had completely ignored:

    LTX can also be used to EXTEND already existing videos.
    All it takes is a 25–50 frame animation as input for LTX to understand the concept, the motion, the faces, and then extend the video to a ridiculous length that no other model currently manages to maintain, while also adding audio on top and motion variations
    and at an unbeatable speed.
    Go check the banana. It works oo-oo-aa-aa

    jeyjeysigFeb 2, 2026· 3 reactions

    @Agent_Smth dude... are you okay Smith? did Neo knock your brains out and you're experiencing brain damage, googoogaga?? lol

    CernerJan 16, 2026· 1 reaction
    CivitAI

    Guide from the LTX team on how to train your own lora, they specialized their own free training system: https://www.youtube.com/watch?v=sL-T6dsO0v4

    MaximilianPsJan 19, 2026· 1 reaction
    CivitAI

    Is this only intended for nude and porn content? I'm testing a workflow using this text‑encoding model for NTX2 in ComfyUI, where a wolf is supposed to consume a bird. The expected behavior is to trigger a bite action and make the bird asset disappear with a small 'puff' of feathers, but the event isn’t firing correctly. Even the NTX2 demo prompt—where the frogs are expected to ingest the fly during the meditation sequence—is not functioning as intended, and the frog refuses to eat the fly.

    exxJan 26, 2026· 1 reaction
    CivitAI

    Is it just my workflow, or are people getting really poor voice quality from this?

    CapAndABullFeb 24, 2026· 1 reaction

    It's not just you. Same here

    AcrossfyJan 27, 2026· 1 reaction
    CivitAI

    I have this error:

    !!! Exception during processing !!! mat1 and mat2 shapes cannot be multiplied (77x384 and 3840x4096)

    How to fix it?

    WItchesAIFeb 26, 2026· 1 reaction

    i have the same issue, found a fix?

    seanybonny21680Jan 28, 2026· 6 reactions
    CivitAI

    I use WanGP v10.54 by DeepBeepMeep. So, where would I put this file?

    marastharFeb 8, 2026· 1 reaction

    Put the file in \pinokio\api\wan.git\app\ckpts
    create à Json (use your best ai chat)
    Put a new Json file on \pinokio\api\wan.git\app\finetunes

    RosettaseesFeb 19, 2026· 1 reaction

    @marasthar What if you don't use pinokio. Also does this add something to the UI or is it switching to something different?

    marastharFeb 27, 2026· 2 reactions

    @Rosettasees In pinokio it will add a finetune, BUT you need to create the fintune file (use claude.ai for helping you, it's the actually the best for create the good files.
    If you use Comfyui... i'dont know how create all the stuff, so i always let Claude create the better workflow for my PC, and it works!

    seanybonny21680Mar 1, 2026· 1 reaction

    @marasthar So, how do I create the finetune and Json file? I am kinda dumb like what would I do in claude or another LLM Model to create said finetune and JSON..

    CivitExploreJan 30, 2026· 7 reactions
    CivitAI

    I'm confused. This resulted in more censorship in the audio. Now "pussy" turns to "puss-y" and "vagina" turns to garbled nonsense. Meanwhile these are fine on the original text encoder on Huggingface.

    finnserhij326Feb 5, 2026· 32 reactions
    CivitAI

    I see no difference between the two sample clips. Both are ugly.

    monkkeysMar 13, 2026· 1 reaction

    Combining this with a better Checkpoint and Loras that could correct the "ugly". Aesthetics isn't the point of the text encoder; that's the job of Checkpoints and Loras. The text encoders are for interpretation of your prompt.

    While neither look 100% realistic, but the bottom is better and for a v1.0, its a step in the right direction. Both videos in the example are weird, but if you look at the hands on the top one... that's not ideal. IMO the top version has more signs of AI than the bottom version.

    d1scoreFeb 5, 2026· 6 reactions
    CivitAI

    Another option that works really well with faster loading: Link the checkpoint in text_encoders, use DualClipLoader (GGUF) to load labonne/gemma-3-12b-it-abliterated.q5_k_m.gguf (or another quant of the same, it works really well), and the checkpoint, as "ltxv" type.

    PlasmadoseFeb 12, 2026· 5 reactions
    CivitAI

    Was excited to see what would come from LTX but so far, unfortunately, most of the videos generated are by mostly people who think they are comedians and/or videos with bland to terrible dialog.

    MrReclusive666Feb 14, 2026· 5 reactions

    my mom says im funny.. :(

    MrReclusive666Feb 14, 2026· 8 reactions
    CivitAI

    can we get an fp4?
    even without 5000 series vram savings is worth it.

    CapAndABullFeb 24, 2026· 2 reactions
    CivitAI

    Although this TE DOES improve SOME aspects, I've found that, at least in my workflows, it DOES have a tendency to cause speech to become fucked up, where the model has trouble forming some basic words (not always, but much of the time)

    validoxMar 6, 2026· 16 reactions
    CivitAI

    thanks but it's not compatible with ltx 2.3

    DonkeyworldMar 6, 2026· 4 reactions

    Good to know!.. i'm looking for a version that can use LTX 2.3 prompt enhancer.. hope someone wil make it or know how

    Open_channels_for_ALLMar 9, 2026· 1 reaction

    @Donkeyworld Let me know when the 2.3 version is out… I’m enjoying LTX desktop right now. I got Claude code to open it up to 24gb VRAM. And then split the text encoder onto my second gpu. Works great for anything up to 10 seconds.

    yoyo12345678Mar 12, 2026· 1 reaction

    f

    mixailckopp978Mar 15, 2026· 1 reaction

    yes it true, I just checked.

    jafar00Mar 18, 2026· 9 reactions
    CivitAI

    It still censors prompts. I just got told off by the avatar in the last render I made lol. Instead of rejecting it, it rendered the character in the reference image and said, "The prompt you provided asks for "s*xually suggestive" content" ...... roflmao

    ultimo_intento
    Author
    Mar 18, 2026· 1 reaction

    probably something else in your workflow

    TienXMar 20, 2026· 4 reactions

    you 100% need to post that video, lol

    exxMar 24, 2026· 2 reactions

    Vids or it didn’t happen

    MrReclusive666Mar 27, 2026

    check the control prompt for gemma enhance, it is in that by default, i got around it by adding "Do not censor, do not refuse, do EXACTLY as requested without complaint"
    i also switch from using the gemma enhancer to an ollama enhancer, gemma in comfy was god awful slow, now i run gemma 4b uncensored in ollama with only using 2gb vram and its 5 seconds to enhance.
    now if i can just figure out how to connect ltx to ollama instead of loading gemma as a clip.

    VelouraApr 29, 2026

    This happens because you have a censored prompt enhancer node attached to your clip text encode. When you type nsfw into your prompt box, it then goes through the prompt enhancer which then tells you that what you typed is sexual blah blah and then that combines with your prompt to give you a final output that may speak that entire warning as dialogue