Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,115 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
library_name: diffusers
|
6 |
+
pipeline_tag: image-to-image
|
7 |
+
tags:
|
8 |
+
- Image-to-Image
|
9 |
+
- ControlNet
|
10 |
+
- Diffusers
|
11 |
+
- QwenImageControlNetPipeline
|
12 |
+
- Qwen-Image
|
13 |
+
base_model: Qwen/Qwen-Image
|
14 |
+
---
|
15 |
+
|
16 |
+
# Qwen-Image-ControlNet-Union
|
17 |
+
This repository provides a unified ControlNet that supports 4 control types (canny, soft edge, depth, pose) for [Qwen-Image](https://huggingface.co/Qwen/Qwen-Image).
|
18 |
+
|
19 |
+
|
20 |
+
# Model Cards
|
21 |
+
- This ControlNet consists of 5 double blocks copied from the pretrained transformer layers.
|
22 |
+
- We train the model from scratch for 50K steps using a dataset of 10M high-quality general and human images.
|
23 |
+
- We train at 1328x1328 resolution in BFloat16, batch size=64, learning rate=4e-5. We set the text drop ratio to 0.10.
|
24 |
+
- This model supports multiple control modes, including canny, soft edge, depth, pose. You can use it just as a normal ControlNet.
|
25 |
+
|
26 |
+
# Showcases
|
27 |
+
<table style="width:100%; table-layout:fixed;">
|
28 |
+
<tr>
|
29 |
+
<td><img src="./conds/canny2.png" alt="canny"></td>
|
30 |
+
<td><img src="./outputs/canny2.png" alt="softedge"></td>
|
31 |
+
</tr>
|
32 |
+
<tr>
|
33 |
+
<td><img src="./conds/soft_edge.png" alt="pose"></td>
|
34 |
+
<td><img src="./outputs/soft_edge.png" alt="depth"></td>
|
35 |
+
</tr>
|
36 |
+
<tr>
|
37 |
+
<td><img src="./conds/depth.png" alt="pose"></td>
|
38 |
+
<td><img src="./outputs/depth.png" alt="depth"></td>
|
39 |
+
</tr>
|
40 |
+
<tr>
|
41 |
+
<td><img src="./conds/pose.png" alt="pose"></td>
|
42 |
+
<td><img src="./outputs/pose.png" alt="depth"></td>
|
43 |
+
</tr>
|
44 |
+
</table>
|
45 |
+
|
46 |
+
# Inference
|
47 |
+
```python
|
48 |
+
import torch
|
49 |
+
from diffusers.utils import load_image
|
50 |
+
|
51 |
+
# before merging, please import via local path
|
52 |
+
from controlnet_qwenimage import QwenImageControlNetModel
|
53 |
+
from transformer_qwenimage import QwenImageTransformer2DModel
|
54 |
+
from pipeline_qwenimage_controlnet import QwenImageControlNetPipeline
|
55 |
+
|
56 |
+
base_model = "Qwen/Qwen-Image"
|
57 |
+
controlnet_model = "InstantX/Qwen-Image-ControlNet-Union"
|
58 |
+
|
59 |
+
controlnet = QwenImageControlNetModel.from_pretrained(controlnet_model, torch_dtype=torch.bfloat16)
|
60 |
+
transformer = QwenImageTransformer2DModel.from_pretrained(base_model, subfolder="transformer", torch_dtype=torch.bfloat16)
|
61 |
+
|
62 |
+
pipe = QwenImageControlNetPipeline.from_pretrained(
|
63 |
+
base_model, controlnet=controlnet, transformer=transformer, torch_dtype=torch.bfloat16
|
64 |
+
)
|
65 |
+
pipe.to("cuda")
|
66 |
+
|
67 |
+
# canny
|
68 |
+
# it is highly suggested to add 'TEXT' into prompt if there are text elements
|
69 |
+
control_image = load_image("conds/canny.png")
|
70 |
+
prompt = "Aesthetics art, traditional asian pagoda, elaborate golden accents, sky blue and white color palette, swirling cloud pattern, digital illustration, east asian architecture, ornamental rooftop, intricate detailing on building, cultural representation."
|
71 |
+
controlnet_conditioning_scale = 1.0
|
72 |
+
|
73 |
+
# soft edge, recommended scale: 0.8 - 1.0
|
74 |
+
# control_image = load_image("conds/soft_edge.png")
|
75 |
+
# prompt = "Photograph of a young man with light brown hair jumping mid-air off a large, reddish-brown rock. He's wearing a navy blue sweater, light blue shirt, gray pants, and brown shoes. His arms are outstretched, and he has a slight smile on his face. The background features a cloudy sky and a distant, leafless tree line. The grass around the rock is patchy."
|
76 |
+
# controlnet_conditioning_scale = 0.9
|
77 |
+
|
78 |
+
# depth
|
79 |
+
# control_image = load_image("conds/depth.png")
|
80 |
+
# prompt = "A swanky, minimalist living room with a huge floor-to-ceiling window letting in loads of natural light. A beige couch with white cushions sits on a wooden floor, with a matching coffee table in front. The walls are a soft, warm beige, decorated with two framed botanical prints. A potted plant chills in the corner near the window. Sunlight pours through the leaves outside, casting cool shadows on the floor."
|
81 |
+
# controlnet_conditioning_scale = 0.9
|
82 |
+
|
83 |
+
# pose
|
84 |
+
# control_image = load_image("conds/pose.png")
|
85 |
+
# prompt = "Photograph of a young man with light brown hair and a beard, wearing a beige flat cap, black leather jacket, gray shirt, brown pants, and white sneakers. He's sitting on a concrete ledge in front of a large circular window, with a cityscape reflected in the glass. The wall is cream-colored, and the sky is clear blue. His shadow is cast on the wall."
|
86 |
+
# controlnet_conditioning_scale = 1.0
|
87 |
+
|
88 |
+
image = pipe(
|
89 |
+
prompt=prompt,
|
90 |
+
negative_prompt=" ",
|
91 |
+
control_image=control_image,
|
92 |
+
controlnet_conditioning_scale=controlnet_conditioning_scale,
|
93 |
+
width=control_image.size[0],
|
94 |
+
height=control_image.size[1],
|
95 |
+
num_inference_steps=30,
|
96 |
+
true_cfg_scale=4.0,
|
97 |
+
generator=torch.Generator(device="cuda").manual_seed(42),
|
98 |
+
).images[0]
|
99 |
+
image.save(f"qwenimage_cn_union_result.png")
|
100 |
+
```
|
101 |
+
|
102 |
+
# Recommended Parameters
|
103 |
+
You can adjust control strength via controlnet_conditioning_scale.
|
104 |
+
- Canny: use cv2.Canny, set controlnet_conditioning_scale in [0.8, 1.0]
|
105 |
+
- Soft Edge: use [AnylineDetector](https://github.com/huggingface/controlnet_aux), set controlnet_conditioning_scale in [0.8, 1.0]
|
106 |
+
- Depth: use [depth-anything](https://github.com/DepthAnything/Depth-Anything-V2), set controlnet_conditioning_scale in [0.8, 1.0]
|
107 |
+
- Pose: use [DWPose](https://github.com/IDEA-Research/DWPose/tree/onnx), set controlnet_conditioning_scale in [0.8, 1.0]
|
108 |
+
|
109 |
+
We strongly recommend using detailed prompts, especially when include text elements. For example, use prompt "A poster with a wilderness scene in the background. In the lower right corner, it says 'InstantX Team. All copyright reserved.'' The headlines are 'Qwen-Image' and 'ControlNet-Union', and the date is '2025.8'." instead of "a poster".
|
110 |
+
|
111 |
+
# Limitations
|
112 |
+
We find that the model was unable to preserve some details, such as small font text.
|
113 |
+
|
114 |
+
# Acknowledgements
|
115 |
+
This model is developed by InstantX Team. All copyright reserved.
|