Update README.md
Browse files
README.md
CHANGED
@@ -15,12 +15,19 @@ tags:
|
|
15 |
- NLP
|
16 |
- DPO
|
17 |
---
|
18 |
-
# Model
|
19 |
-
This model is a result of merging two models A and B.
|
20 |
-
The method used for merging is "slerp" with [mergekit](https://github.com/cg123/mergekit).
|
21 |
|
22 |
-
A
|
23 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
|
25 |
### Jinja Prompt Template
|
26 |
```
|
|
|
15 |
- NLP
|
16 |
- DPO
|
17 |
---
|
18 |
+
# Model Overview
|
|
|
|
|
19 |
|
20 |
+
This model is a result of a sophisticated merging process involving two distinct models, Model A and Model B. The merging methodology employed is the "slerp" technique, facilitated by the use of [mergekit](https://github.com/cg123/mergekit).
|
21 |
+
|
22 |
+
## Component Models
|
23 |
+
|
24 |
+
### Model A
|
25 |
+
- **Source**: [mistralai/Mistral-7B-instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
|
26 |
+
- **Description**: This model is designed to provide instruction-based outputs, enhancing the clarity and precision in response generation.
|
27 |
+
|
28 |
+
### Model B
|
29 |
+
- **Source**: Based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)
|
30 |
+
- **Enhancements**: Fine-tuned using DPO (Direct Preference Optimization) training, augmenting its capabilities for more adaptive and context-aware responses.
|
31 |
|
32 |
### Jinja Prompt Template
|
33 |
```
|