Prompt: Transform the image into realistic image.

Change backgrounds. Swap objects. Add stuff. Remove stuff. Adjust styles. All through simple text prompts instead of wrestling with complicated tools.

Qwen is Alibaba's image editing model, built on their 20B-parameter foundation. It handles object manipulation, style transfers, and even text editing inside images. The results are surprisingly realistic, and it keeps context better than you'd expect.

Useful if you're creating content, designing stuff, running social media, or just want to edit images without learning Photoshop.

What we'll cover

  1. What Qwen Image Edit actually is and what makes it different
  2. Getting the workflow running on ThinkDiffusion
  3. Installing the models and custom nodes you need
  4. Walking through the workflow step-by-step
  5. Real examples of what it can do
  6. Common issues and how to fix them

What is Qwen Image Edit?

Source: Qwen Image

Qwen Image Edit is a model developed by Alibaba's Qwen team, built upon their robust 20B-parameter Qwen-Image foundation. This model brings precise object manipulation, accurate style and background transfer, and dual-language text editing directly within images. It does a solid job with realism and keeping details intact, even when you're asking it to do tricky edits.

Ideal for content creators, designers, marketers, social media teams, localization experts, e-commerce businesses, and anyone seeking intuitive, professional-grade image editing through the power of natural language.

Prompt: Transform the image crochet style.

So go ahead: give your edits a voice, and see just how far your words can take your next photo adventure!

Download Workflow

Installation guide

  1. Download the workflow file
  2. Open ComfyUI (local or ThinkDiffusion)
  3. Drag the workflow file into the ComfyUI window
  4. If you see red nodes, install missing components:
  • ComfyUI Manager > Install Missing Custom Nodes

Verified to work on ThinkDiffusion Build: September 5, 2025

ComfyUI v0.3.57 with the use qwen_image_edit_fp8_e4m3fn.safetensors
and qwen_2.5_vl_7b_fp8_scaled.safetensors

Note: We specify the build date because ComfyUI and custom node versions updated after this date may change the behavior or outputs of the workflow.

Minimum Machine Size: Ultra

Use the specified machine size or higher to ensure it meets the VRAM and performance requirements of the workflow

💡
Download the workflow and drag & drop it into your ComfyUI window, whether locally or on ThinkDiffusion. If you're using ThinkDiffusion, minimum requirement is the Turbo 24gb machine, but we do recommend the Ultra 48gb machine.

Custom Nodes

If there are red nodes in the workflow, it means that the workflow lacks the certain required nodes. Install the custom nodes in order for the workflow to work.

  1. Go to the ComfyUI Manager  > Click Install Missing Custom Nodes
  1. Check the list below if there's a list of custom nodes that needs to be installed and click the install.

Required Models

For this guide you'll need to download these 3 recommended models.

1. qwen_image_fp8_e4m3fn.safetensors
2. qwen_2.5_vl_7b_fp8_scaled.safetensors
3. qwen_image_vae.safetensors
  1. Go to ComfyUI Manager  > Click Model Manager
  1. Search for the models above and when you find the exact model that you're looking for, click install, and make sure to press refresh when you are finished.

If Model Manager doesn't have them: Use direct download links (included with workflow) and upload through ThinkDiffusion MyFiles > Upload URL. Refer our docs for more guidance on this.

You could also use the model path source instead: by pasting the model's link address into ThinkDiffusion MyFiles using upload URL.

Model Name Model Link Address ThinkDiffusion Upload Directory
qwen_image_fp8_e4m3fn.safetensors
📋 Copy Path
.../comfyui/models/diffusion_models/
qwen_2.5_vl_7b_fp8_scaled.safetensors
📋 Copy Path
.../comfyui/models/text_encoders/
qwen_image_vae.safetensors
📋 Copy Path
.../comfyui/models/vae/

Step-by-step Workflow Guide

This workflow was pretty easy to set up and runs well from the default settings. Here are a few steps where you might want to take extra note.

Steps Recommended Nodes
1. Load an Input Image

Load an input image. Any minimal resolution will do as long it has a high quality image.
2. Set the Models

Set the models as seen on the image. If you have good hardware you can use the full model but needs to be manual download.
3. Write the Prompt

Write a simple which serves as instruction of what kind image edit you want to initiate.
4. Check the Sampling

Set the settings as seen on the image.
5. Check the Output

💡
When I try complex edits or chain too many instructions, the results can include artifacts, context loss, or image offsets, especially in heavy transformation scenarios.
💡
While text rendering and semantic edits are strong, I have to watch out for occasional mismatches, artifacts, or unexpected outcomes until newer versions address these issues.
💡
Sometimes, when I edit an image, Qwen Image Edit changes the aspect ratio or introduces zoom, which means the output doesn’t perfectly match my original framing or pixel dimensions.

Examples

IP Creation

Prompt: This dog wears a doctor suit, no helmet on its head and wears a stethoscope.

Novel View Synthesis

Prompt: Obtain the back-side of toy.

Avatar Creator

Prompt: Transform the image into Ghibli style.

Object Add

Prompt: Add a realistic cat beside the dog.

Object Removal

Prompt: Remove the bird.

Object Replace

Prompt: Replace the coffee with coke in can.

Background Swap

Prompt: Replace the background with beach.

Virtual Try-On

Prompt: Replace the woman dress into a futuristic cyberpunk dress.

Text Editing

Prompt: Replace the 'Hard Rock' to ThinkDiffusion'

Troubleshooting

Red Nodes: Install missing custom nodes through ComfyUI Manager
Out of Memory: Use smaller expansion factors or switch to Ultra machine
Poor Quality: Check input image resolution and adjust kontext strength
Visible Seams: Lower strength and ensure good prompt description

If you’re having issues with installation or slow hardware, you can try any of these workflows on a more powerful GPU in your browser with ThinkDiffusion.

Join the ThinkDiffusion Discord Server!
ThinkDiffusion is your Stable Diffusion workspace in the cloud with unrestricted, bleeding edge opensource AI art tools. | 5510 members