We are thrilled to release Qwen-Image, an image generation foundation model in the Qwen series that achieves significant advances in complex text rendering and precise image editing. Experiments show strong general capabilities in both image generation and editing, with exceptional performance in text rendering, especially for Chinese.
One of its standout capabilities is high-fidelity text rendering across diverse images. Whether it’s alphabetic languages like English or logographic scripts like Chinese, Qwen-Image preserves typographic details, layout coherence, and contextual harmony with stunning accuracy. Text isn’t just overlaid—it’s seamlessly integrated into the visual fabric.
Beyond text, Qwen-Image excels at general image generation with support for a wide range of artistic styles. From photorealistic scenes to impressionist paintings, from anime aesthetics to minimalist design, the model adapts fluidly to creative prompts, making it a versatile tool for artists, designers, and storytellers.
When it comes to image editing, Qwen-Image goes far beyond simple adjustments. It enables advanced operations such as style transfer, object insertion or removal, detail enhancement, text editing within images, and even human pose manipulation—all with intuitive input and coherent output. This level of control brings professional-grade editing within reach of everyday users.
But Qwen-Image doesn’t just create or edit—it understands. It supports a suite of image understanding tasks, including object detection, semantic segmentation, depth and edge (Canny) estimation, novel view synthesis, and super-resolution. These capabilities, while technically distinct, can all be seen as specialized forms of intelligent image editing, powered by deep visual comprehension.
Together, these features make Qwen-Image not just a tool for generating pretty pictures, but a comprehensive foundation model for intelligent visual creation and manipulation—where language, layout, and imagery converge.
License Agreement
Qwen-Image is licensed under Apache 2.0.
Original Text and Models: https://huggingface.co/Qwen/Qwen-Image
Description
Qwen-Image Full BF16 (non-distilled), originally posted https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/tree/main/split_files/diffusion_models
FAQ
Comments (39)
Would the BF16 model work on system with an RTX 5090 32GB and 96GB of system RAM?
No, the original Qwen-Image_bf16 is ~40GB, which will not fit in VRAM of a 5090. The fp8 version will (~20GB), and produces great results locally. If you need to go smaller, there are a number of GGUF Quants now too.
theally Oh, ok. I see there's a Q8 quant. Is that better quality than FP8?
Also, I'm getting halftone pattern texture all over my images. It's not always noticeable, but since Qwen produces quite soft images I'm sometimes using image sharpening nodes in ComfyUI and this really makes this pattern visible. Do you know how I can get rid of this pattern?
theally On a 5070 ti, with 16gb vram, are there gguf options?
Thank you so much for making this available to a wider audience :)
theally Ok, so I've tried the BF16 model and it did work on my setup! Not sure how, maybe Comfy offloads parts of the model to system RAM on the fly. I didn't notice major slowdowns in generation time. Maybe some slight delay before it initiates inference.
I am using 4bit gguf K_S version of qwen image on my RTX4060 8GB with 16GB RAM, it take about 330 seconds to generate an image.
Wow, finally a dedicated model category for Qwen.
It feels like the most underrated model right now, and at the same time the one with the greatest potential.
I simply can't wait for CIVITAI to roll out LORA training for QWEN Image and stable generation for QWEN Image.. I did generate some sample images from CivitAI's "Dorment" QWEN Image with the exact same complex prompt without LORAs that I had difficulty doing it with Flux, had to train my own LORAs.
Why are you taking so long to add Chroma support?
This and chroma are the more promising models. Waiting for more checkpoint and loras.
can you add image2image support now?
We'll be adding Qwen-Image-Edit support soon!
I can't try it with the civitai generator
Why not? It seems to be working at the moment.
can somone create some lora with tongue? like ahegao , biting maybe licking? i wanted the lora for that
There is a ahego qwen lora but it looks lower quality. The current MCNL lora doesnt seem to do tongues well either
Following the prompts and configs of the mirror dress one to verify my setup, why is my result similar but more blurry?
If you have their workflow and are literally running it but aren't getting the same results they could still be using different loras, different prompt, or different models. They also could have upscaled it, or altered the prompt or the negatives in ways that weren't shared.
I can get perfectly clear photos with no particular positive brute-forcing with 8-steps, res_2s, bong_tangent. Also (eta 0.5, steps_to_run -1, denoise 1.17, cfg 1.14, model shift 1.04, sampler standard). (yes those denoise and cfg are pretty schizo values, but I got this workflow from here: https://civitai.com/models/1851673/mcnl-multi-concept-nsfw-lora-qwen-image)
pretty sure this image specifically: https://civitai.com/images/95469890
Also, a common strategy right now is to create an image with Qwen and then upscale it with Wan2.2. I don't exactly know the photo you are referring to, so I don't know your issue. But I'd say what got my generations to the clarity of Wan2.2 was a combination of the Lightning Lora and realism loras. Otherwise, I don't think any of my phrases or negatives are all that important.
Here's a brute-force negative: "blurry, out of focus, shallow depth of field, low quality, bad quality, low detail, mutated, jpeg artifacts, compression artifacts, bright colors, overexposed, static, blurred details, subtitles, style, artwork, painting, overall gray, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, malformed limbs, fused fingers, three legs, bad neck anatomy, walking backwards, model, idol, face too pretty, neck completely rotated, head completely backwards,"
Is anyone else having issues downloading the Full Bf16 version? it just times out and says: civitai-delivery-worker-prod.5ac0637cfd0766c97916cefa3764fbdf.r2.cloudflarestorage.com didn’t send any data.
What happened to Qwen? Why has the category disappeared?
Мм хуета. Проще на луну слетать чем эту херь запустить здесь. Пробовал соединить две лоры не запустилось. В Comfyui без проблем.
что-нибудь поменялось?
@prot1wow37 Нет она выдает ошибку при создании запроса. Красным значок постоянно.
@meikerAI3d можешь пожалуйста скинуть свой workflow? у меня RTX 3050 8gb, но ComfyUI постоянно ошибку "переподключение" выдает. Flux, Chroma без проблем работают, только Qwen нихуя не запускается
Я на Forge запустил
SOLVED-answer inside Im struggling to find where to download CLIP that is in this workflow. huggingface has 3 part thing which is weird. how to load it in workflow? please any hints?
Check the official ComfyUI documentation (https://docs.comfy.org/tutorials/image/qwen/qwen-image). There should be an example workflow that automatically downloads the text encoder, as well as a direct download link for the text encoder (https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors).
@VV24 im sorry how did you found the link of that fp8 text encoder? i was looking for an hour and found only 3split part 1 part 2 part3.. and i was confused. thanks!!!
@gemstonebro No worries, this stuff can be pretty confusing. The download link is right there in the documentation I linked, in the section "Model Download", along with all the other required model files: https://docs.comfy.org/tutorials/image/qwen/qwen-image#2-model-download
When will onsite Generation for this Qwen model be fixed? It hasn't worked in around 40 days and is one of the best model we have right now.
Please add built-in 4/8steps lightning lora support. Those would be game changer. 🙏
No longer available for the Civitai generator, it seems it wasn't fixed
Not yet, no. It's being worked on!
@theally Must be a really big issues if it has take over 4 months to fix when it was working before! ?....
@J1B Yup, difficult issue to solve. We've had to bring in more compute (GPUs), and we're close to relaunching, with the latest model, and Qwen Image Edit support.
Can Forge use this? How to use it?
Can RTX 4050 with 6GB VRAM run for Forge Neo?
Details
Files
Available On (1 platform)
Same model published on other platforms. May have additional downloads or version variants.


