Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,7 @@ tags:
|
|
13 |
license: llama3.3
|
14 |
---
|
15 |
After having some fair success with merging some of my favorite Llama models, particularly on
|
16 |
-
Tarek07/Progenitor-V1.1-LLaMa-70B, which was based on the excellent Steelskull/L3.3-MS-Nevoria-70b merge, just with a couple extra ingredients and different merge methods. I then decided to try my hand at working the new deepseek-ai/DeepSeek-R1-Distill-Llama-70B in. I have decided to try mix it in 3 different ways. For this first model I decided to use Deepseek as the base, using the same formula as my Progenitor V1.1 model. From early testing it seems very promising.
|
17 |
# merge
|
18 |
|
19 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
|
|
13 |
license: llama3.3
|
14 |
---
|
15 |
After having some fair success with merging some of my favorite Llama models, particularly on
|
16 |
+
Tarek07/Progenitor-V1.1-LLaMa-70B, which was based on the excellent Steelskull/L3.3-MS-Nevoria-70b merge, just with a couple extra ingredients and different merge methods. I then decided to try my hand at working the new deepseek-ai/DeepSeek-R1-Distill-Llama-70B in. I have decided to try mix it in 3 different ways. For this first model I decided to use Deepseek as the base, using the same formula as my Progenitor V1.1 model. From early testing it seems very promising. But I think Delta beat this one.
|
17 |
# merge
|
18 |
|
19 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|