Text-to-Image
Diffusers
English
Not-For-All-Audiences
art
RedHotTensors commited on
Commit
f85cc49
1 Parent(s): bd173c1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -52,8 +52,8 @@ The model is zero-terminal-SNR with V-prediction. Use the ModelSamplingDiscrete
52
  Experimental textual inversion embeddings in a similar vein to the [Boring Embeddings](https://huggingface.co/FoodDesert/Boring_Embeddings) are provided above.
53
  They're intended to improve quality while not drastically altering image content. They should be used as part of a negative prompt, although using them in the positive prompt can be fun too.
54
 
55
- - The "lite" version is 6 tokens wide and is initialized on the values of ``by <|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|>``, which is very close to a "blank slate".
56
- - The "plus" version is trained on the same dataset, is 8 tokens wide, and is initialized on an average vector of 100 low-scoring artists. Currently, the "lite" version is recommended.
57
 
58
  ## Training Details
59
  - Adaptive timestep weighting: Timesteps are weighted using a similar method to what the EDM2 paper used, according to the homoscedastic uncertainty of MSE loss on each timestep, thereby equalizing the contribution of each timestep. Loss weight was also conditioned on resolution in order to equalize the contribution of each resolution group. The overall effect of this is that the model is now very good at both high- and low-frequency details, and is not as biased towards blurry backgrounds.
 
52
  Experimental textual inversion embeddings in a similar vein to the [Boring Embeddings](https://huggingface.co/FoodDesert/Boring_Embeddings) are provided above.
53
  They're intended to improve quality while not drastically altering image content. They should be used as part of a negative prompt, although using them in the positive prompt can be fun too.
54
 
55
+ - The "lite" version is 6 tokens wide and is initialized on the values of ``by <|endoftext|><|endoftext|><|endoftext|><|endoftext|><|endoftext|>``, which is very close to a "blank slate". Current, this version is recommended.
56
+ - The "plus" version is trained on the same dataset, is 8 tokens wide, and is initialized on an average vector of 100 low-scoring artists.
57
 
58
  ## Training Details
59
  - Adaptive timestep weighting: Timesteps are weighted using a similar method to what the EDM2 paper used, according to the homoscedastic uncertainty of MSE loss on each timestep, thereby equalizing the contribution of each timestep. Loss weight was also conditioned on resolution in order to equalize the contribution of each resolution group. The overall effect of this is that the model is now very good at both high- and low-frequency details, and is not as biased towards blurry backgrounds.