|
--- |
|
license: other |
|
language: |
|
- en |
|
--- |
|
|
|
# Detail |
|
|
|
Quantized at 6.13bpw h8, as always |
|
|
|
# Original model card |
|
|
|
I received a request to upload the FP16 version of my first [MergeMonster](https://github.com/Gryphe/MergeMonster/tree/main) 13b merge, which had the sole focus of reducing overused ChatGPT roleplaying phrases. Many ministrations were murdered throughout this process. |
|
|
|
GGUF quantizations can be found in my main [MergeMonster](https://huggingface.co/Gryphe/MergeMonster) repo. |
|
|
|
## Merge Composition |
|
|
|
Unsurprisingly, not much remained of MythoMax. I'm (obviously!) very fond of it, but it's not perfect. |
|
|
|
Reminder: MergeMonster works with hundreds of pieces scattered throughout the final merge. If a model takes a large part of the composition this does not mean it is merged in a manner that might be logical to humans. |
|
``` |
|
Base model: MythoMax-L2-13b |
|
|
|
------ FINAL MERGE COMPOSITION ------ |
|
LLaMA2-13B-Tiefighter: 0.27 |
|
Emerhyst-13B: 0.26 |
|
Orca-2-13b: 0.24 |
|
Stheno-1.8-L2-13B: 0.09 |
|
MythoMax-L2-13b: 0.09 |
|
Athena-v4: 0.04 |
|
``` |
|
|
|
## Prompt Format |
|
|
|
As always, I suggest sticking to the Alpaca format. |
|
|
|
``` |
|
### Instruction: |
|
Your instruction or question here. |
|
|
|
### Response: |
|
``` |