How to Use
This is a style LoRA, designed to be the base layer of your LoRA stack. It creates the foundational aesthetic of realism, upon which you can add character or concept LoRAs.
Note: The included ZIP archive contains both the high-noise and low-noise LoRA variants, along with our recommended ComfyUI workflows.
Trigger Word:
InstacamRecommended Strength:
1.0. Start here and adjust in small increments.
I would also like to thank Danrisi, who originally taught us how to train LoRAs and helped make our work possible.
Description
Here's my second iteratoion of my Instagirl LoRA. I doubled the dataset, increased diversity, and tried to go fore a more "ameture" asthetic that WAN usually struggles with.
This model was NOT trained on ANY NSFW, but if your really a gooner like that I can reccomend Flux Kontext (also maybe it would be good to fix text in images)
Here is my ChatGPT system prompt I use, how it works is that I will find an image I want to replicate on pintrest/instagram and paste it into ChatGPT with this system prompt:
Write one run-on paragraph (about 70 words) of concise, comma-separated phrases with no quotation marks that recreates every attached reference image for Flux.
• Start with the subject and LoRA trigger exactly: Instagirl, petite body, then continue in this order: pose → camera angle → clothing & key accessories → environment/background → lighting → extra atmosphere.
• Keep the phrasing natural but brief—think quick notes to a friend—each detail separated by commas rather than full sentences (e.g. new york city, gloomy day, beautiful girl, taking a selfie).
• State that the photo is amateur cellphone quality and make sure to clearly state the hair color, type, and facial structure; if and only if it’s a mirror selfie, add shot with a silver iPhone (three cameras) held by the subject.
• Ignore all tattoos, piercings, body mods, glasses, on-screen GUIs, or icons.
• Avoid filler words and avoid synonyms for the LoRA triggers so their influence stays strong.
• After the paragraph, on a new line, list exactly: visible sensor noise, artificial over-sharpening, heavy HDR glow, amateur photo, blown-out highlights, crushed shadows.
If there are no subjects, ignore the subject part completely
I have not tried this with any character LORAS yet, but constructive feedback is VERY welcome, my discord is 00quebec
Workflow is based off of AI_Character's WAN 2.2 workflow: https://www.reddit.com/r/StableDiffusion/comments/1mcgyxp/wan22_new_fixed_txt2img_workflow_important_update/
FAQ
Comments (19)
Can this also be used for text-to-image?
yes
t2i ***is*** t2v
It was specifically made for text to image, not trained on any videos
@00quebec The images are looking great! I am struggling using Wan for t2i, getting bad dithering patterns (e.g. 80s big hair perm, blond curly hair, dark background). How can i avoid these dithering patterns?
ok, after some more tests I found out, that if you are using turbo loras it's important to set shift=1.0 and use euler not lcm
00quebec That doesn't matter unless you are trying to train a motion or action. This is a T2V lora.
playtime_ai WAN T2V Lora training question: So if I were to train on an /image dataset/ only (not videos) of someone, say a celeb, I would be able to use that /character/ lora trained on WAN and accurately have the person face when using T2I?
Follow up question, would it be possible to produce a video trained on images in WAN? Especially this lora: Instagirls. Probably not so, eh? Not sure how it works still.
ProvenFlawless To answer your first question, yes. You can absolutely train a lora of a person or character on images only and use that lora to create picture of said person or character (but don't post loras or images of real people to Civitai. It isn't allowed.)
To answer your second question, yes... you absolutely can use a lora trained on only images to create a video. You only need to train on video if you are trying to train a motion or an action... like a dance move or something like that. You can train many things... a person's likeness, an outfit, a body type... on just pictures.
Looks good! Danrisi is a legend, has helped me out a bunch too. Anyways I suspect you don't have diversity of ethnicities in your dataset?
I am
working on that for gen 2, first dataset only had a few different people.
How do you keep the character consistent ? Does this works with character loras ?
Updated so quickly?,,
New one releasing in a few hrs
Can generally complete epochs in 4-8 hours depending on how big the dataset is, etc. Doesn't have to indicate anything
So this is mainly for T2i?
Yes he mentioned this in a reply to another user 5734... bunch of numbers
Although someone mentioned t2v IS t2i, just a bunch of frames of it.
Gamerman420 I asked this because you can train WAN off images or short clips. If they're short clips, it leans towards some specific motion (e.g BJ, sex lora, tittydrop etc)
It's great thanks