Update README.md
Browse files
README.md
CHANGED
@@ -6,23 +6,22 @@ license_link: https://freedevproject.org/faipl-1.0-sd/
|
|
6 |
language:
|
7 |
- en
|
8 |
library_name: diffusers
|
9 |
-
base_model: Laxhar/noobai-XL-Vpred-0
|
10 |
---
|
11 |
|
12 |
The LHC (Large Heap o' Chuubas) is aiming to be the model for all of your VTuber needs.
|
13 |
|
14 |
-
|
15 |
-
|
16 |
-
The LHC models are a series of Vtuber centric finetunes.
|
17 |
-
As opposed to many smallscale finetunes, where the aim is to improve aesthetics or a general concept like backgrounds, the aim of the LHC models is primarily to add specific characters while preserving as much of the base model as possible.
|
18 |
-
## The Why
|
19 |
-
The usual way of adding characters to an already trained model is LoRAs or similar methods, where you end up with a small model that can be applied to existing models, adding concepts in a very plug and play way. While this is a very convenient way of achieving this, that most modern consumer GPUs are capable of training, they come with several downsides to a model that has been fully trained on these concepts.
|
20 |
-
1. Loras will have an effect on composition, style and character knowledge even outside of their intended concept. This will be especially apparent in incorrectly trained loras, where the concept will always be applied, causing poses to stiffen or characters taking on attributes of the new character. While this bleeding can and does happen on finetuned models as well, the effect is drastically smaller.
|
21 |
-
2. Full finetunes usually result in the model being very capable of abstracting, meaning even if a specific combination of concepts isn't present in the training data, a well trained finetune will be able to combine these concepts in an almost logical way.
|
22 |
-
3. As a specific example of 2.: while style and concept Loras usually work quite well when used together with other Loras of their type, character Loras tend to be much less capable of being used together with other character Loras for generating images of multiple new characters. While there are ways to make this work, and some Loras are more capable than others, this also limits the results in my experience. One workaround to this is training a Lora on multiple characters as well as images of these characters together, which does work quite well in my experience, however finding artwork of specific characters together is not always possible and when adding more and more characters, the one must also increased the size of the resulting Lora in order to learn more and more characters. This make Loras less and less effective the more characters one wants to add at once.
|
23 |
-
4. It is possible to extract a Lora from a model, meaning the things the finetune learned can be applied to models that have a similar base. This means that even if the resulting finetune doesn't work for every application, the extract can profit from most of the advantages of a Lora.
|
24 |
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
language:
|
7 |
- en
|
8 |
library_name: diffusers
|
9 |
+
base_model: Laxhar/noobai-XL-Vpred-1.0
|
10 |
---
|
11 |
|
12 |
The LHC (Large Heap o' Chuubas) is aiming to be the model for all of your VTuber needs.
|
13 |
|
14 |
+
Currently on Version 0.5 (considered alpha). It is a v-prediction model, with the necessary settings enabled to allow for automatic detection in ComfyUI and (Re)Forge.
|
15 |
+
For a comprehensive overview of all vtubers that were trained on, refer to https://catbox.moe/c/pjfwt1 or the txt file, detailing the vtubers, as well as useful/necessary tagging.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
+
# Training Details
|
18 |
+
|
19 |
+
Due to some major changes between version 0.4 and 0.5, the training parameters were not yet dialed in, resulting in a very long training process that was resumed multiple times from intermediate checkpoints that weren't trained to satisfaction. Therefore, exact training parameters aren't stated, instead the training logs were uploaded, and a general overview is given here.
|
20 |
+
|
21 |
+
## Parameters
|
22 |
+
|
23 |
+
* Total Epochs: 102
|
24 |
+
* Effective Batchsize: 32 (or 16, or 8)
|
25 |
+
* Initial Learning Rate: 5e-5
|
26 |
+
* Scheduler: Cosine
|
27 |
+
* TE Training Epochs: ~10
|