CivArchive
    Preview 2

    .

    ★Training Wan2.2 Lora with 16GB of VRAM

    .

    !!! No download required! This is a technical showcase! !!!

    .

    →→→Original / Training dataset containing configurations provider→→→ https://civitai.com/models/1944129?modelVersionId=2200388

    .

    ●Differences from the original:

    ・4070 Ti Super 16GB / Mem ??GB

    ・This is my first training (Really the first run)

    ・(Replace musubi_tuner_gui.py)(I forgot to do this so i2v didn't work)

    ・t2v (Unchecked I2V Training)

    ・The model was also changed from i2v to t2v

    ・Changed the path of other models to my environment

    .

    ・Install Triton and Sageattention2 (SDPA / I'm not sure about the effect, but I think Xformers would be good too.)

    pip install -U "triton-windows<3.3"

    python -s -m pip install .\triton-3.2.0-cp312-cp312-win_amd??.whl

    →→→Guide is here→→→https://civitai.com/articles/12848

    .

    ・Blocks to Swap (As far as I know, this model has 40 blocks)

    35 (original is 10)

    .

    ●VRAM required for training

    30 percent of the time, I was using 15GB, and the rest of the time it was around 10GB.

    .

    As I write this, I set it to 40 to see if I could train with 12GB, but I got an error.

    "AssertionError: Cannot swap more than 39 blocks. Requested 40 blocks to swap."

    .

    ●Time required for training

    The high was 2 hours and 19 minutes, and Low was 2 hours and 17 minutes.

    I trained for 20 epochs, but it took 5 epochs to see results.

    So if I do it right, I can train in just over an hour. (Wow, I can't believe that!)

    .

    .

    .

    Description

    .

    LORA
    Wan Video 2.2 T2V-A14B

    Details

    Downloads
    66
    Platform
    SeaArt
    Platform Status
    Available
    Created
    9/20/2025
    Updated
    9/20/2025
    Deleted
    -

    Files

    Available On (1 platform)

    Same model published on other platforms. May have additional downloads or version variants.