The world of open-source AI image generation is moving at a breakneck pace, and the release of Qwen-Image-Edit has provided a powerful new tool for creators. This model brings sophisticated editing capabilities, previously the domain of proprietary cloud-based services, directly to your local machine. However, setting up these powerful models can be a significant hurdle. This guide provides a comprehensive, step-by-step walkthrough for installing and running the Qwen-Image-Edit model using ComfyUI, a flexible and powerful node-based interface for Stable Diffusion and related models. By following this tutorial, you’ll unlock advanced features like enhanced character consistency and precise industrial design editing, giving you full control over your creative projects without relying on external services.
What is Qwen-Image-Edit?
Qwen-Image-Edit is a specialized large vision language model (LVLM) developed by Alibaba Cloud. Unlike standard text-to-image models that generate pictures from scratch, Qwen-Image-Edit excels at modifying existing images based on natural language instructions. It’s designed for high-precision editing tasks where you need to make specific changes while preserving the rest of the image context. This makes it an incredibly powerful tool for designers, artists, and developers who need more granular control than simple generation offers.
Key capabilities of the model include:
- Instruction-Based Editing: You can give it commands like “change the color of the car to blue,” “add sunglasses to the man,” or “make the background a forest.”
- High Coherence: The model is adept at maintaining the original image’s style, lighting, and composition, ensuring that edits look natural and seamless.
- Complex Edits: It can handle multi-step or complex instructions, making it suitable for professional workflows in areas like product design, character art, and architectural visualization.
- Local and Private: By running it on your own hardware, you maintain complete privacy and control over your images and data, with no recurring subscription fees.
Why use ComfyUI for your local setup?
ComfyUI is a graphical user interface for AI image models that uses a node-based workflow. Instead of a linear series of settings, you connect different functional blocks (nodes) to build a processing pipeline. While this might seem intimidating at first, it offers unparalleled flexibility and insight into the image generation process. For a model like Qwen-Image-Edit, which has specific input requirements, ComfyUI is the ideal environment.
The primary benefits of using ComfyUI include:
- Modularity and Control: Each step of the process—loading the model, inputting an image, writing a prompt, and saving the output—is a separate node. This allows you to precisely control the data flow and troubleshoot issues easily.
- Efficiency: ComfyUI only re-runs the parts of the workflow that have changed, saving significant time and computational resources when you’re iterating on an idea.
- Community and Extensibility: A vast library of custom nodes developed by the community allows you to integrate almost any new model or technique, including Qwen-Image-Edit.
- Transparency: You can see exactly how your image is being processed, making it a fantastic tool for learning and experimentation.

Prerequisites for installation
Before diving into the installation, ensure your system meets the necessary hardware and software requirements. Running large AI models locally is resource-intensive.
| Component | Requirement |
|---|---|
| GPU | An NVIDIA GPU with at least 8 GB of VRAM is recommended for optimal performance. AMD and Apple Silicon GPUs may work but often require additional configuration. |
| RAM | A minimum of 16 GB of system RAM, with 32 GB being ideal for handling larger models and images. |
| Storage | At least 25 GB of free disk space. The Qwen-Image-Edit model itself is over 10 GB. |
| Software | Git and Python installed on your system. These are essential for installing ComfyUI and its dependencies. |
Step-by-step installation guide
Follow these steps carefully to set up your local Qwen-Image-Edit environment with ComfyUI.
1. Install ComfyUI
If you don’t already have ComfyUI installed, the easiest way is via their official GitHub repository.
- Open a command prompt or terminal.
- Navigate to the directory where you want to install ComfyUI.
- Clone the repository by running the following command:
git clone https://github.com/comfyanonymous/ComfyUI.git- Navigate into the newly created ComfyUI directory:
cd ComfyUI. - Install the required Python dependencies by running:
pip install -r requirements.txt.
2. Install the ComfyUI Manager
The ComfyUI Manager is an essential custom node that simplifies the installation of other custom nodes and models. It’s the first thing you should add to any new ComfyUI setup.
- In your terminal, navigate to the
ComfyUI/custom_nodes/directory. - Clone the ComfyUI Manager repository:
git clone https://github.com/ltdrdata/ComfyUI-Manager.git- Restart ComfyUI. You should now see a “Manager” button in the main menu.
3. Download the Qwen-Image-Edit model
The model is available on Hugging Face. You’ll need to download the main model file.
- Go to the official model page:
Qwen/Qwen-VL-Chaton Hugging Face. - Navigate to the “Files and versions” tab.
- Download the
qwen-vl-chat.fp16.safetensorsfile. This is a large file (around 10.5 GB), so the download may take some time. - Once downloaded, move this
.safetensorsfile into theComfyUI/models/checkpoints/directory. This is where ComfyUI looks for its main models.
4. Install the required custom nodes
To run the Qwen model, you need a specific custom node that knows how to interact with it. We can use the ComfyUI Manager for this.
- Start ComfyUI (by running
python main.pyin the ComfyUI directory if it’s not already running). - Click the “Manager” button in the side panel.
- Click “Install Custom Nodes.”
- Search for “ComfyUI-Qwen-VL” and click the “Install” button next to it.
- After the installation is complete, close the manager and restart ComfyUI. This allows it to load the new nodes.
Your first Qwen-Image-Edit workflow
With everything installed, it’s time to build a basic workflow to test the model. A typical Qwen editing workflow involves loading the model, providing an input image, and giving it a text instruction.

Here is how to structure your node graph:
- Load Checkpoint: This node should already be on your canvas. Use it to select the
qwen-vl-chat.fp16.safetensorsmodel you downloaded. - Qwen-VL Loader: Add this new node (right-click > Add Node > loaders > Qwen-VL Loader). Connect the “MODEL” output from the Load Checkpoint node to the “ckpt” input of this loader node.
- Load Image: Add a “Load Image” node (Add Node > image > Load Image) and choose a picture you want to edit.
- Qwen-VL Chat: This is the main editing node. Add it from “Add Node > Qwen > Qwen-VL Chat”. It has several inputs:
- Connect the “model” output from the Qwen-VL Loader to its “model” input.
- Connect the “image” output from your Load Image node to its “image” input.
- In the “prompt” text box, type your editing instruction (e.g., “add a hat on the person”).
- Save Image: Add a “Save Image” node (Add Node > image > Save Image). Connect the “image” output of the Qwen-VL Chat node to the “images” input of the Save Image node.
Once all nodes are connected, click “Queue Prompt.” The model will process your request, and the final edited image will appear in the Save Image node. You can now experiment with different images and prompts to explore the model’s capabilities.
Conclusion
Setting up a powerful local AI environment like Qwen-Image-Edit with ComfyUI is a rewarding process that gives you complete creative freedom. By following this guide, you have successfully installed a state-of-the-art image editing model and a flexible interface to control it. You are no longer tethered to cloud services, subscriptions, or privacy concerns. The true power of this setup lies in its modularity; you can now integrate other models, experiment with complex workflows, and push the boundaries of AI-assisted creativity.
Your next steps should be to explore more advanced workflows, experiment with different editing instructions, and engage with the vibrant ComfyUI community to discover new custom nodes and techniques. Welcome to the future of local, open-source AI image editing.

