thesunday commited on
Commit
aaddbb1
·
1 Parent(s): 2275d8b

Update model card

Browse files
Files changed (1) hide show
  1. README.md +95 -0
README.md CHANGED
@@ -1,3 +1,98 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - en
5
  ---
6
+
7
+ # Model Description
8
+ This is an experiment to test merging 14 models using DARE TIES 🦙
9
+
10
+ The result is a base model that performs quite well but requires some further instruction fine-tuning.
11
+
12
+ The 14 models are as follows:
13
+ 1. [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
14
+ 2. [ehartford/dolphin-2.2.1-mistral-7b](https://huggingface.co/ehartford/dolphin-2.2.1-mistral-7b)
15
+ 3. [SciPhi/SciPhi-Mistral-7B-32k](https://huggingface.co/SciPhi/SciPhi-Mistral-7B-32k)
16
+ 4. [ehartford/samantha-1.2-mistral-7b](https://huggingface.co/ehartford/samantha-1.2-mistral-7b)
17
+ 5. [Arc53/docsgpt-7b-mistral](https://huggingface.co/Arc53/docsgpt-7b-mistral)
18
+ 6. [berkeley-nest/Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha)
19
+ 7. [Q-bert/MetaMath-Cybertron-Starling](https://huggingface.co/Q-bert/MetaMath-Cybertron-Starling)
20
+ 8. [Open-Orca/Mistral-7B-OpenOrca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca)
21
+ 9. [v1olet/v1olet_marcoroni-go-bruins-merge-7B](https://huggingface.co/v1olet/v1olet_marcoroni-go-bruins-merge-7B)
22
+ 10. [beowolx/MistralHermes-CodePro-7B-v1](https://huggingface.co/beowolx/MistralHermes-CodePro-7B-v1)
23
+ 11. [TIGER-Lab/MAmmoTH-7B-Mistral](https://huggingface.co/TIGER-Lab/MAmmoTH-7B-Mistral)
24
+ 12. [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B)
25
+ 13. [Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp](https://huggingface.co/Weyaxi/OpenHermes-2.5-neural-chat-v3-3-Slerp)
26
+ 14. [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B)
27
+
28
+ - base model: [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)
29
+
30
+ The yaml config file for this model is here:
31
+
32
+ ```yaml
33
+ models:
34
+ - model: /media/data5/hf_models/Mistral-7B-v0.1
35
+ # no parameters necessary for base model
36
+ - model: /media/data5/hf_models/dolphin-2.2.1-mistral-7b
37
+ parameters:
38
+ weight: 0.08
39
+ density: 0.4
40
+ - model: /media/data5/hf_models/SciPhi-Mistral-7B-32k
41
+ parameters:
42
+ weight: 0.08
43
+ density: 0.4
44
+ - model: /media/data5/hf_models/samantha-1.2-mistral-7b
45
+ parameters:
46
+ weight: 0.08
47
+ density: 0.4
48
+ - model: /media/data5/hf_models/docsgpt-7b-mistral
49
+ parameters:
50
+ weight: 0.08
51
+ density: 0.4
52
+ - model: /media/data5/hf_models/Starling-LM-7B-alpha
53
+ parameters:
54
+ weight: 0.08
55
+ density: 0.4
56
+ - model: /media/data5/hf_models/MetaMath-Cybertron-Starling
57
+ parameters:
58
+ weight: 0.08
59
+ density: 0.4
60
+ - model: /media/data5/hf_models/Mistral-7B-OpenOrca
61
+ parameters:
62
+ weight: 0.08
63
+ density: 0.4
64
+ - model: /media/data5/hf_models/v1olet_marcoroni-go-bruins-merge-7B
65
+ parameters:
66
+ weight: 0.08
67
+ density: 0.4
68
+ - model: /media/data5/hf_models/MistralHermes-CodePro-7B-v1
69
+ parameters:
70
+ weight: 0.08
71
+ density: 0.4
72
+ - model: /media/data5/hf_models/MAmmoTH-7B-Mistral
73
+ parameters:
74
+ weight: 0.08
75
+ density: 0.4
76
+ - model: /media/data5/hf_models/OpenHermes-2.5-Mistral-7B
77
+ parameters:
78
+ weight: 0.08
79
+ density: 0.4
80
+ - model: /media/data5/hf_models/OpenHermes-2.5-neural-chat-v3-3-Slerp
81
+ parameters:
82
+ weight: 0.08
83
+ density: 0.4
84
+ - model: /media/data5/hf_models/NeuralHermes-2.5-Mistral-7B
85
+ parameters:
86
+ weight: 0.08
87
+ density: 0.4
88
+ - model: /media/data5/hf_models/Mistral-7B-Instruct-v0.2
89
+ parameters:
90
+ weight: 0.08
91
+ density: 0.5
92
+ merge_method: dare_ties
93
+ base_model: /media/data5/hf_models/Mistral-7B-v0.1
94
+ parameters:
95
+ int8_mask: true
96
+ dtype: bfloat16
97
+
98
+ ```