JosephusCheung
commited on
Commit
·
3f97d34
1
Parent(s):
330f2fa
Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ license: gpl-3.0
|
|
5 |
|
6 |
A Chat Model, Testing only, no performance guaranteeeee...
|
7 |
|
8 |
-
In short: CausalLM / Qwen 8x7B MoE in Mixtral Arch, 8 real experts in different domains.
|
9 |
|
10 |
Only intended for conceptual validation, however the expert models do not seem to be working as expected. The model could output text and complete the conversation normally, but the performance of the expert model was not significant.
|
11 |
|
|
|
5 |
|
6 |
A Chat Model, Testing only, no performance guaranteeeee...
|
7 |
|
8 |
+
In short: CausalLM / Qwen 8x7B MoE in Mixtral Arch, 8 real experts in different domains. Trained, not a merge.
|
9 |
|
10 |
Only intended for conceptual validation, however the expert models do not seem to be working as expected. The model could output text and complete the conversation normally, but the performance of the expert model was not significant.
|
11 |
|