CivArchive
    LTX-2 DEV - IMAGE to Video and TEXT to Video with Ollama - v1.5 LTX-2
    NSFW
    Preview 119334503

    V1.5 LTX-2 DEV Video with Audio including latest 🅛🅣🅧 Multimodal Guider

    Image to Video and a Text to Video workflow, both can use own Prompts or Ollama generated/enhanced prompts.

    Replaced the Guider node with latest Multimodal Guider node, see more details in WF notes or here: https://ltx.io/model/model-blog/ltx-2-better-control-for-real-workflows Before we had 1 CFG parameter for audio and video. With multimodal guider, we now can tweak audio and video seperately with even more parameters...


    V1.0 LTX-2 DEV Video with Audio:

    Image to Video and a Text to Video workflow with own Prompts or Ollama generated/enhanced prompts.

    • setup for the LTX2 Dev model.

    • uses Detailer Lora for better quality and LTX tiled VAE to avoid OOM and visual grids

    • 2 pass rendering (motion+upscale). Upscale process uses distilled and spatial upscale Lora

    • setup with latest LTXVNormalizingSampler to increase video & audio quality.

    • Text to Video can use dynamic prompts with wildcards.


    I am using these starting parameters for ComfyUi to avoid OOM (my setup: 16g Vram/64g Ram):

    --lowvram --cache-none --reserve-vram 6 --preview-method none


    Download Files:

    Find Model/Lora Loader nodes within Sampler Subgraph node.

    - LTX2 Dev Model (dev_Fp8): https://huggingface.co/Lightricks/LTX-2/tree/main

    - Detailer Lora: https://huggingface.co/Lightricks/LTX-2-19b-IC-LoRA-Detailer/tree/main

    - Distilled (lora-384) & Spatial upscaler Lora: https://huggingface.co/Lightricks/LTX-2/tree/main

    - VAE (already included in above dev_FP8 model, but needed if you go for GGUF models): https://huggingface.co/Lightricks/LTX-2/tree/main/vae

    - Textencoder (fp8_e4m3fn): https://huggingface.co/GitMylo/LTX-2-comfy_gemma_fp8_e4m3fn/tree/main

    - Image to Video Adapter Lora (more motion with I2V): https://huggingface.co/MachineDelusions/LTX-2_Image2Video_Adapter_LoRa/tree/main

    - Ollama Models:

    Save Location:

    📂 ComfyUI/

    ├── 📂 models/

    │ ├── 📂 checkpoints/

    │ │ ├── ltx-2-19b-dev-fp8.safetensors

    │ ├── 📂 text_encoders/

    │ │ └── gemma_3_12B_it_fp8_e4m3fn.safetensors

    │ ├── 📂 loras/

    │ │ ├── ltx-2-19b-distilled-lora-384.safetensors

    │ └── 📂 latent_upscale_models/

    │ └── ltx-2-spatial-upscaler-x2-1.0.safetensors


    Custom Nodes used:

    https://github.com/Lightricks/ComfyUI-LTXVideo

    https://github.com/rgthree/rgthree-comfy

    https://github.com/yolain/ComfyUI-Easy-Use

    https://github.com/stavsap/comfyui-ollama

    https://github.com/evanspearman/ComfyMath (V1.0 only, not required as of V1.5)

    Text 2 Video only:

    https://github.com/KoinnAI/ComfyUI-DynPromptSimplified

    Res2_s sampler (optional, recommended by Lightricks, but slow, alternative samplers: euler, euler_ancest):

    https://github.com/ClownsharkBatwing/RES4LYF

    Description

    LTX2 DEV Image to Video and Text to Video

    (including Multimodal guider and LTXV Normalizing Sampler)

    Workflows
    LTXV2

    Details

    Downloads
    617
    Platform
    CivitAI
    Platform Status
    Available
    Created
    1/30/2026
    Updated
    2/8/2026
    Deleted
    -

    Files

    ltx2DEVIMAGEToVideoAnd_v15LTX2.zip

    Mirrors