CivArchive
    ComfyUI workflow for Flux (simple + Lora) - Original
    NSFW
    Preview 22403597
    Preview 22597695

    It is a simple workflow of Flux AI on ComfyUI.

    Actually there are many other beginners who don't know how to add LORA node and wire it, so I put it here to make it easier for you to get started and focus on your testing.

    And we have new way to run Flux ez with 1 click: https://civarchive.com/models/628682/flux-1-checkpoint-easy-to-use

    Check out more detailed instructions here: https://maitruclam.com/flux-ai-la-gi/

    To summarize my experience with it:

    You will need at least 30 GB to use them :)

    ***

    If you are a newbie like me, you will be less confused when trying to figure out how to use Flux on ComfyUI.

    In addition to this workflow, you will also need:

    Download Model:

    1. Model: flux1-dev.sft: 23.8 GB

    Link:  https://huggingface.co/black-forest-labs/FLUX.1-dev/tree/main

    Location: ComfyUI/models/unet/

    Download CLIP:

    1. t5xxl_fp16.safetensors: 9.79 GB

    2. clip_l.safetensors: 246 MB

    3. (optional if your machine has less than 32GB of TvT ram) t5xxl_fp8_e4m3fn.safetensors: 4.89 GB

    Link: https://huggingface.co/comfyanonymous/flux_text_encoders/tree/main

    Location: ComfyUI/models/clip/

    Download VAE:

    1. ae.sft: 335 MB

    Link: https://huggingface.co/black-forest-labs/FLUX.1-schnell/blob/main/ae.safetensors

    Location: ComfyUI/models/vae/

    If you are using an Ubuntu VPS like me, the command is as simple as this:

    # Download t5xxl_fp16.safetensors to the directory ComfyUI/models/clip/

    wget -P /home/ubuntu/ComfyUI/models/clip/ https://huggingface.co/comfyanonymous/flux_text_encoders/resolve/main/t5xxl_fp16.safetensors

    # Download tp clip_l.safetensors to ComfyUI/models/clip/

    wget -P /home/ubuntu/ComfyUI/models/clip/ https://huggingface.co/comfyanonymous/flux_text_encoders/resolve/main/clip_l.safetensors

    # (Optional) Download t5xxl_fp8_e4m3fn.safetensors to ComfyUI/models/clip/

    wget -P /home/ubuntu/ComfyUI/models/clip/ https://huggingface.co/comfyanonymous/flux_text_encoders/resolve/main/t5xxl_fp8_e4m3fn.safetensors

    # Download ae.sft to ComfyUI/models/vae/

    wget -P /home/ubuntu/ComfyUI/models/vae/ https://huggingface.co/black-forest-labs/FLUX.1-schnell/resolve/main/ae.safetensors

    For the model, you will need to learn how to generate Huggingface Access Tokens and add them to download and use like this:

    I don't know much about them so you can find out more.

    Why don't I make tutorial for Windows 10, 11 or XP? What do you expect from a Mario 64 laptop :)

    Original tutorial: https://comfyanonymous.github.io/ComfyUI_examples/flux/

    Cách sử dụng:

    • •Sampling method: Euler

    • •Schedule type: Simple

    • •Sampling steps: 30

    • Weight: 0.8 - 1.2. Best: 0.8

    Instructions for use:

    • •Sampling method: Euler a

    • •Schedule type: Simple

    • •Sampling steps: 30

    • Weight: 0.8 - 1.2. Best: 0.8

    In the prompt, you should use the word: Woman using the word Girl will create many body anatomy errors.

    0-0-0-0-0-0-0-0-0-0-0-0

    Actually, you will not need the activation keyword for it to work, but you can add it to make Flux understand faster and give better results. ⚡

    Note: It works well with FLUX.1-Turbo-Alpha, LORA human face. 👤✨

    Useful and FREE resources:

    ❤️Free server to make art with Flux: Shakker

    More FLUX LORA? List and detailed description of each LORA I implement here: https://maitruclam.com/lora

    🆕 First time using FLUX? Explanation and tutorial with A1111 forge offline and Comfy UI here: https://maitruclam.com/flux-ai-la-gi/

    🛠️ How to train your LORA with Flux? My detailed instructions are here: https://maitruclam.com/training-flux/

    ❤️ Donate me (I would be really surprised if you did that! 😄): https://maitruclam.com/donate

    Find me / Contact for work on:

    📱 Facebook: @maitruclam4real

    💬 Discord: @maitruclam

    🌐 Web: maitruclam.com

    Description

    Flux dev AI

    FAQ

    Comments (38)

    JohnnyB1Aug 2, 2024
    CivitAI

    30 GB of what do we need?

    maitruclam
    Author
    Aug 2, 2024

    model + vae + clip = 30gb++ storage

    DarknoiceAug 2, 2024· 1 reaction

    Also decent amount of Ram and Vram

    maitruclam
    Author
    Aug 2, 2024· 1 reaction

    @Darknoice oh yeah i have 24gb vram and 32gb ram total average time for dev version is 30s to 47s

    tdfilmstudioAug 2, 2024

    @maitruclam is that time taken for 1024 by 1024 res images?

    JohnnyB1Aug 2, 2024

    @maitruclam with the same setup it takes me a little longer, but the latest comy UI seems to have a bug or something, getting an error regarding FP8 (even with FP16 clip :( ) well I will try again later.

    slumberjohnAug 3, 2024

    @JohnnyB1 same for me

    maitruclam
    Author
    Aug 3, 2024

    @JohnnyB1 try reducing the steps to 12 and the image to 512x768, some of my friends tried it and it was ok on the dev version

    filmingtonAug 2, 2024
    CivitAI

    I've done some experiments and still can't figure out how to add in a negative prompt. Perhaps it can't be done right now?

    maitruclam
    Author
    Aug 2, 2024

    basically it's not necessary, unless you combine it with SD to improve face or inpaint or img to img (someone will do workflow for these eventually)

    filmingtonAug 3, 2024· 4 reactions

    @maitruclam It's not necessary for your workflow, no. You're right. But it is necessary if you want to remove reoccurring objects that keep popping up.

    civit910Aug 4, 2024

    @filmington Yeah i'd love to know too! No matter how much i emphasise on an image to look like a digital oil painting with heavy brush strokes, they always turn out like real photographs.

    maitruclam
    Author
    Aug 2, 2024· 2 reactions
    CivitAI

    I don't know if the dev team likes BlackPink, but if you prompt Asian or Korean, the face will look very similar to Jennie in BlackPink :)

    meikerAI3dAug 2, 2024· 2 reactions
    CivitAI

    Потребляет около 36 Гб. Результат с реализмом хороший. С аниме неплохой. Из минусов нету персонажей в базе данных не Nier 2b, не Konosuba. и да кошко деффок тоже не хочет не видит промт neko

    changpeidong2148Aug 4, 2024· 1 reaction
    CivitAI

    My comfyui is the latest version, and both the large model and VAE model of sft format have been placed in the specified position, but when I select the large model and VAE in the comfyui workflow, I do not know the reason why I cannot see the sft model.

    maitruclam
    Author
    Aug 4, 2024

    try reloading and switching between .safetensors and .sft. Also, do you have comfy connected to A1111 to save space? If so, try disconnecting them or putting the model file in the A1111 folder.

    Also, you can check this post to see if it has any suggestions for you: https://civitai.com/articles/6479/flux-error-and-how-to-fix-it-or-not

    efefeAug 7, 2024

    same here, couldnt find the solution,
    I change the extention to safetensors,
    then i press the queue after that it gave an error


    im getting this error when i change the unet flux1-dev from .sft to safetensors

    Error occurred when executing UNETLoader: 'conv_in.weight' File "D:\_software\TyDiffusion\Engines\ComfyUI\ComfyUI\execution.py", line 151, in recursive_execute output_data, output_ui = get_output_data(obj, input_data_all) File "D:\_software\TyDiffusion\Engines\ComfyUI\ComfyUI\execution.py", line 81, in get_output_data return_values = map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True) File "D:\_software\TyDiffusion\Engines\ComfyUI\ComfyUI\execution.py", line 74, in map_node_over_list results.append(getattr(obj, func)(**slice_dict(input_data_all, i))) File "D:\_software\TyDiffusion\Engines\ComfyUI\ComfyUI\nodes.py", line 814, in load_unet model = comfy.sd.load_unet(unet_path) File "D:\_software\TyDiffusion\Engines\ComfyUI\ComfyUI\comfy\sd.py", line 565, in load_unet model = load_unet_state_dict(sd) File "D:\_software\TyDiffusion\Engines\ComfyUI\ComfyUI\comfy\sd.py", line 538, in load_unet_state_dict model_config = model_detection.model_config_from_diffusers_unet(sd) File "D:\_software\TyDiffusion\Engines\ComfyUI\ComfyUI\comfy\model_detection.py", line 380, in model_config_from_diffusers_unet unet_config = unet_config_from_diffusers_unet(state_dict) File "D:\_software\TyDiffusion\Engines\ComfyUI\ComfyUI\comfy\model_detection.py", line 262, in unet_config_from_diffusers_unet match["model_channels"] = state_dict["conv_in.weight"].shape[0]

    FejjCAug 6, 2024
    CivitAI

    The VAE link points to a 404 page, is there an alternative link?

    InfiniteFantasyArtAug 7, 2024· 3 reactions

    You have to get it here https://huggingface.co/black-forest-labs/FLUX.1-dev/tree/main , it is called ae.safetensors as of today

    FejjCAug 7, 2024

    @InfiniteFantasyArt Thank you!!

    cathylevermanAug 13, 2024
    CivitAI

    im trying to make a simple spiral using flux. It cant draw a simple spiral. It only does round circles, but it cant do a spiral :/

    Anyone else think this is weird? Shouldn't these models easily generate simple primitives like this?

    trying to generate something like this:
    https://openclipart.org/image/2400px/svg_to_png/131347/1302066690.png

    SirColinJan 22, 2025

    I am surprised it doesn't recognize a spiral in the prompt - but you could also train a LoRA to teach it to draw spirals, worst case!

    AaliShineAug 15, 2024
    CivitAI

    Sooooooooooooooo if i download it and use it inside of A1111 and generate iamges did i able to generaste any image with text effect now ??

    maitruclam
    Author
    Aug 20, 2024

    if u use a1111 forge it will work

    RuinDwellerAug 24, 2024
    CivitAI

    Can anyone please help me update dual clip loader? Mine only has options for SD3 and SDXL. :(

    RuinDwellerAug 24, 2024

    I needed to update-all, which took all night, but now I've gotten the option for FLUX to appear. The problem now is that 6gb VRAM is apparently not enough to run it, I need at least 8. It keeps telling me the "header is too large" when I try to load ae.safetensors.

    The links to the 1-click installer are still helpful so thank you, but I imagine that 8gb limit will still be in effect.

    maitruclam
    Author
    Aug 28, 2024

    @RuinDweller i think u need to try g4 ver, i can run it in 6gb vram 3060 laptop

    barlogspam176Aug 27, 2024
    CivitAI

    please change the vae link. ae.safetsnor moved to another folder

    maitruclam
    Author
    Aug 28, 2024

    ok done bro

    barlogspam176Aug 28, 2024

    @maitruclam thanks

    rachelcenterSep 3, 2024
    CivitAI

    can you recommend a flux workflow to me that lets me create handwriting that looks like it was drawn on a whiteboard? I tried it myself in flux and it doesnt look photo realistic

    maitruclam
    Author
    Sep 8, 2024

    hmm i think you will need to train a lora yourself for that, as of now it is still the best model for text generation.

    shinypubes2626Sep 6, 2024· 13 reactions
    CivitAI

    So far, this is the only Flux workflow in civit AI I got to run. And unfortunately, it's a little to simplistic for me. sorry.

    maitruclam
    Author
    Sep 8, 2024

    like its name dude, simple and for beginners, i make it more complex by adding 1 more node to load lora V:

    shinypubes2626Sep 8, 2024

    @maitruclam I didn't mean to come off as rude. But it was a great starting point. I have since added elements to this workflow and made it my own. Thanks for that.

    maitruclam
    Author
    Sep 9, 2024· 1 reaction

    @shinypubes2626 it's normal, i don't find it offensive or rude hahaha

    blhllNov 22, 2024

    Lucky you, I just get black square. Back to forge I guess.

    Workflows
    Flux.1 D

    Details

    Downloads
    6,665
    Platform
    CivitAI
    Platform Status
    Available
    Created
    8/2/2024
    Updated
    4/30/2026
    Deleted
    -

    Files

    comfyuiWorkflowForFlux_original.zip

    Mirrors