DavidAU commited on
Commit
1f6f411
1 Parent(s): d087184

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -14
README.md CHANGED
@@ -44,20 +44,7 @@ pipeline_tag: text-generation
44
  I took the original models in "L3-Stheno-Maid-Blackroot 8B" and completely rebuilt it a new pass-through merge (everything preserved)
45
  and blew it out to over 20.7 billion parameters at full float 32 precision.
46
 
47
- From these 3 core models I build TWO custom Llama3 models:
48
-
49
- Grand Horror 16.5B ( <A href="https://huggingface.co/DavidAU/L3-Stheno-Maid-Blackroot-Grand-HORROR-16B-GGUF"> here </a> ) and
50
- Grand Story 16.5B Alpha (unreleased) then merged these together with a "smoothing step" captured at F32 precision.
51
-
52
- The result is a model that is far more stable, far more capable than any of the 3 models originally and it is more than the "sum"
53
- of the two 16.5B models noted. Compared to Grand Horror 16.5B is it over 25000 points lower (IQ4XS) in perplexity
54
- (lower is better) or 2.5 full levels of magnitude lower.
55
-
56
- It is tougher, stronger and can handle a far wider range of operating conditions - from temp .1 to temp 5 all day long.
57
-
58
- The F32 precision (along with full F32 transfer to the ggufs) increases the performance even further.
59
-
60
- Then I added the NEW 10X Brainstorm Adapter called "Foreshadow".
61
 
62
  This punched the model out to 20.7B and added unique feature to the model: "unbalanced"
63
 
 
44
  I took the original models in "L3-Stheno-Maid-Blackroot 8B" and completely rebuilt it a new pass-through merge (everything preserved)
45
  and blew it out to over 20.7 billion parameters at full float 32 precision.
46
 
47
+ Then I added the NEW 20X Brainstorm Adapter called "Unbalanced".
 
 
 
 
 
 
 
 
 
 
 
 
 
48
 
49
  This punched the model out to 20.7B and added unique feature to the model: "unbalanced"
50