CivArchive
    LTX-2 DEV - IMAGE to Video and TEXT to Video with Ollama - v1.0 LTX-2
    NSFW
    Preview 117898337

    V1.5 LTX-2 DEV Video with Audio including latest 🅛🅣🅧 Multimodal Guider

    Image to Video and a Text to Video workflow, both can use own Prompts or Ollama generated/enhanced prompts.

    Replaced the Guider node with latest Multimodal Guider node, see more details in WF notes or here: https://ltx.io/model/model-blog/ltx-2-better-control-for-real-workflows Before we had 1 CFG parameter for audio and video. With multimodal guider, we now can tweak audio and video seperately with even more parameters...


    V1.0 LTX-2 DEV Video with Audio:

    Image to Video and a Text to Video workflow with own Prompts or Ollama generated/enhanced prompts.

    • setup for the LTX2 Dev model.

    • uses Detailer Lora for better quality and LTX tiled VAE to avoid OOM and visual grids

    • 2 pass rendering (motion+upscale). Upscale process uses distilled and spatial upscale Lora

    • setup with latest LTXVNormalizingSampler to increase video & audio quality.

    • Text to Video can use dynamic prompts with wildcards.


    I am using these starting parameters for ComfyUi to avoid OOM (my setup: 16g Vram/64g Ram):

    --lowvram --cache-none --reserve-vram 6 --preview-method none


    Download Files:

    Find Model/Lora Loader nodes within Sampler Subgraph node.

    - LTX2 Dev Model (dev_Fp8): https://huggingface.co/Lightricks/LTX-2/tree/main

    - Detailer Lora: https://huggingface.co/Lightricks/LTX-2-19b-IC-LoRA-Detailer/tree/main

    - Distilled (lora-384) & Spatial upscaler Lora: https://huggingface.co/Lightricks/LTX-2/tree/main

    - VAE (already included in above dev_FP8 model, but needed if you go for GGUF models): https://huggingface.co/Lightricks/LTX-2/tree/main/vae

    - Textencoder (fp8_e4m3fn): https://huggingface.co/GitMylo/LTX-2-comfy_gemma_fp8_e4m3fn/tree/main

    - Image to Video Adapter Lora (more motion with I2V): https://huggingface.co/MachineDelusions/LTX-2_Image2Video_Adapter_LoRa/tree/main

    - Ollama Models:

    Save Location:

    • 📂 ComfyUI/

    • ├── 📂 models/

    • │ ├── 📂 checkpoints/

    • │ │ ├── ltx-2-19b-dev-fp8.safetensors

    • │ ├── 📂 text_encoders/

    • │ │ └── gemma_3_12B_it_fp8_e4m3fn.safetensors

    • │ ├── 📂 loras/

    • │ │ ├── ltx-2-19b-distilled-lora-384.safetensors

    • │ └── 📂 latent_upscale_models/

    • │ └── ltx-2-spatial-upscaler-x2-1.0.safetensors


    Custom Nodes used:

    Res2_s sampler (optional, recommended by Lightricks, but slow, alternative samplers: euler, euler_ancest):

    https://github.com/ClownsharkBatwing/RES4LYF

    Description

    LTX2 DEV Image to Video and Text to Video

    Workflows
    LTXV2

    Details

    Downloads
    432
    Platform
    CivitAI
    Platform Status
    Available
    Created
    1/18/2026
    Updated
    2/27/2026
    Deleted
    -

    Files