Update README.md
Browse files
README.md
CHANGED
@@ -28,7 +28,7 @@ Nous Hermes 2 Mixtral 8x7B DPO is the new flagship Nous Research model trained o
|
|
28 |
|
29 |
The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.
|
30 |
|
31 |
-
This is the SFT + DPO version of Mixtral Hermes 2, we
|
32 |
|
33 |
## We are grateful to Together.ai for sponsoring our compute during the many experiments both training Mixtral and working on DPO!
|
34 |
|
|
|
28 |
|
29 |
The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.
|
30 |
|
31 |
+
This is the SFT + DPO version of Mixtral Hermes 2, we have also released an SFT only version, for people to find which works best for them, which can be found here: https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT
|
32 |
|
33 |
## We are grateful to Together.ai for sponsoring our compute during the many experiments both training Mixtral and working on DPO!
|
34 |
|