Gerold Meisinger
commited on
Commit
·
0bb179b
1
Parent(s):
129f9e8
control-edgedrawing-cv480edpf-drop50-fp16-checkpoint-90000
Browse files
README.md
CHANGED
@@ -4,9 +4,11 @@ datasets:
|
|
4 |
- ChristophSchuhmann/improved_aesthetics_6.5plus
|
5 |
language:
|
6 |
- en
|
|
|
|
|
7 |
---
|
8 |
|
9 |
-
Controls image generation by edge maps generated with [Edge Drawing](https://github.com/CihanTopal/ED_Lib). Edge Drawing comes in different flavors: original (_ed_), parameter
|
10 |
|
11 |
* Based on my monologs at [github.com - Edge Drawing](https://github.com/lllyasviel/ControlNet/discussions/318)
|
12 |
* For usage see the model page on [civitai.com - Model](https://civitai.com/models/149740).
|
@@ -26,11 +28,13 @@ prompt: _a detailed high-quality professional photo of swedish woman standing in
|
|
26 |
|
27 |

|
28 |
|
|
|
|
|
29 |
**Canndy Edge for comparison (default in Automatic1111)**
|
30 |
|
31 |

|
32 |
|
33 |
-
|
34 |
|
35 |
# Image dataset
|
36 |
|
@@ -61,7 +65,7 @@ accelerate launch train_controlnet.py ^
|
|
61 |
|
62 |
To evaluate the model it makes sense to compare it with the original Canny model. Original evaluations and comparisons are available at [ControlNet 1.0 repo](https://github.com/lllyasviel/ControlNet), [ControlNet 1.1 repo](https://github.com/lllyasviel/ControlNet-v1-1-nightly), [ControlNet paper v1](https://arxiv.org/abs/2302.05543v1), [ControlNet paper v2](https://arxiv.org/abs/2302.05543) and [Diffusers implementation](https://huggingface.co/takuma104/controlnet_dev/tree/main). Some points we have to keep in mind when comparing canny with edpf in order not to compare apples with oranges:
|
63 |
* canny 1.0 model was trained on 3M images with fp32, canny 1.1 model on even more, while edpf model so far is only trained on a 180k-360k with fp16.
|
64 |
-
* canny edge-detector requires parameter tuning while edpf is parameter
|
65 |
* Do we manually fine-tune canny to find the perfect input image or do we leave it at default? We could argue that "no fine-tuning required" is the usp of edpf and we want to compare in the default setting, whereas canny fine-tuning is subjective.
|
66 |
* Would the canny model actually benefit from a edpf pre-processor and we might not even require a edpf model? (2023-09-25: see `eval_canny_edpf.zip` but it seems as if it doesn't work and the edpf model may be justified)
|
67 |
* When evaluating human images we need to be aware of Stable Diffusion's inherent limits, like disformed faces and hands, and don't attribute them to the control net.
|
@@ -109,7 +113,7 @@ edges = ed.detectEdges(image)
|
|
109 |
edge_map = ed.getEdgeImage(edges)
|
110 |
```
|
111 |
|
112 |
-
45000 steps => looks good. released as **version 0.1 on civitai**.
|
113 |
|
114 |
**Experiment 3.1 - 2023-09-24 - control-edgedrawing-cv480edpf-drop0-fp16-checkpoint-90000**
|
115 |
|
@@ -123,7 +127,9 @@ resumed with epoch 2 from 90000 using `--proportion_empty_prompts=0.5` => result
|
|
123 |
|
124 |
see experiment 3.0. restarted from 0 with `--proportion_empty_prompts=0.5` => results are not good, 50% is probably too much for 45k steps. guessmode still doesn't work and tends to produces humans. resuming until 90k with right-left flipped in the hope it will get better with more images.
|
125 |
|
126 |
-
**Experiment 4.1 - control-edgedrawing-cv480edpf-drop50-fp16-checkpoint-
|
|
|
|
|
127 |
|
128 |
# Ideas
|
129 |
|
|
|
4 |
- ChristophSchuhmann/improved_aesthetics_6.5plus
|
5 |
language:
|
6 |
- en
|
7 |
+
tags:
|
8 |
+
- controlnet
|
9 |
---
|
10 |
|
11 |
+
Controls image generation by edge maps generated with [Edge Drawing](https://github.com/CihanTopal/ED_Lib). Edge Drawing comes in different flavors: original (_ed_), parameter free (_edpf_), color (_edcolor_).
|
12 |
|
13 |
* Based on my monologs at [github.com - Edge Drawing](https://github.com/lllyasviel/ControlNet/discussions/318)
|
14 |
* For usage see the model page on [civitai.com - Model](https://civitai.com/models/149740).
|
|
|
28 |
|
29 |

|
30 |
|
31 |
+
_Clear and pristine! Wooow!_
|
32 |
+
|
33 |
**Canndy Edge for comparison (default in Automatic1111)**
|
34 |
|
35 |

|
36 |
|
37 |
+
_Noise, artifacts and missing edges. Yuck! Ugh!_
|
38 |
|
39 |
# Image dataset
|
40 |
|
|
|
65 |
|
66 |
To evaluate the model it makes sense to compare it with the original Canny model. Original evaluations and comparisons are available at [ControlNet 1.0 repo](https://github.com/lllyasviel/ControlNet), [ControlNet 1.1 repo](https://github.com/lllyasviel/ControlNet-v1-1-nightly), [ControlNet paper v1](https://arxiv.org/abs/2302.05543v1), [ControlNet paper v2](https://arxiv.org/abs/2302.05543) and [Diffusers implementation](https://huggingface.co/takuma104/controlnet_dev/tree/main). Some points we have to keep in mind when comparing canny with edpf in order not to compare apples with oranges:
|
67 |
* canny 1.0 model was trained on 3M images with fp32, canny 1.1 model on even more, while edpf model so far is only trained on a 180k-360k with fp16.
|
68 |
+
* canny edge-detector requires parameter tuning while edpf is parameter free.
|
69 |
* Do we manually fine-tune canny to find the perfect input image or do we leave it at default? We could argue that "no fine-tuning required" is the usp of edpf and we want to compare in the default setting, whereas canny fine-tuning is subjective.
|
70 |
* Would the canny model actually benefit from a edpf pre-processor and we might not even require a edpf model? (2023-09-25: see `eval_canny_edpf.zip` but it seems as if it doesn't work and the edpf model may be justified)
|
71 |
* When evaluating human images we need to be aware of Stable Diffusion's inherent limits, like disformed faces and hands, and don't attribute them to the control net.
|
|
|
113 |
edge_map = ed.getEdgeImage(edges)
|
114 |
```
|
115 |
|
116 |
+
45000 steps => looks good. resuming with left-right flipped images. released as **version 0.1 on civitai**.
|
117 |
|
118 |
**Experiment 3.1 - 2023-09-24 - control-edgedrawing-cv480edpf-drop0-fp16-checkpoint-90000**
|
119 |
|
|
|
127 |
|
128 |
see experiment 3.0. restarted from 0 with `--proportion_empty_prompts=0.5` => results are not good, 50% is probably too much for 45k steps. guessmode still doesn't work and tends to produces humans. resuming until 90k with right-left flipped in the hope it will get better with more images.
|
129 |
|
130 |
+
**Experiment 4.1 - 2023-09-26 - control-edgedrawing-cv480edpf-drop50-fp16-checkpoint-90000**
|
131 |
+
|
132 |
+
resumed from 45000 steps with left-right flipped images => results are still not good, 50% is probably too much for 45k steps. guessmode still doesn't work and tends to produces humans. abort.
|
133 |
|
134 |
# Ideas
|
135 |
|
control-edgedrawing-cv480edpf-drop50-fp16-checkpoint-90000.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8a65184060f15e136ef5871c240d433a70bee6e61c1f9c1731d9e0c5783c40b2
|
3 |
+
size 722598616
|
control-edgedrawing-cv480edpf-drop50-fp16-checkpoint-90000.zip
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a27b63038afcde03e686d924d662bbd5d4c950b0019f6423c38e6854c36a79b
|
3 |
+
size 5469226
|