Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
---
|
2 |
-
base_model:
|
3 |
tags:
|
4 |
- Mixtral
|
5 |
- instruct
|
@@ -18,7 +18,7 @@ language:
|
|
18 |
- en
|
19 |
---
|
20 |
|
21 |
-
# Nous Hermes 2 -
|
22 |
|
23 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/qVRnEDL_BUEWulUvWBD95.png)
|
24 |
|
@@ -28,6 +28,8 @@ Nous Hermes 2 Mixtral 7bx8 DPO is the new flagship Nous Research model trained o
|
|
28 |
|
29 |
The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.
|
30 |
|
|
|
|
|
31 |
# Table of Contents
|
32 |
1. [Example Outputs](#example-outputs)
|
33 |
2. [Benchmark Results](#benchmark-results)
|
|
|
1 |
---
|
2 |
+
base_model: mistralai/Mixtral-8x7B-v0.1
|
3 |
tags:
|
4 |
- Mixtral
|
5 |
- instruct
|
|
|
18 |
- en
|
19 |
---
|
20 |
|
21 |
+
# Nous Hermes 2 - Mixtral 8x7B-DPO
|
22 |
|
23 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/qVRnEDL_BUEWulUvWBD95.png)
|
24 |
|
|
|
28 |
|
29 |
The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.
|
30 |
|
31 |
+
This is the SFT + DPO version of Mixtral Hermes 2, we will also be providing an SFT only version, for people to find which works best for them.
|
32 |
+
|
33 |
# Table of Contents
|
34 |
1. [Example Outputs](#example-outputs)
|
35 |
2. [Benchmark Results](#benchmark-results)
|