mav23 commited on
Commit
f471493
1 Parent(s): 1c29ff8

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +265 -0
  3. magnum-v3-27b-kto.Q4_0.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ magnum-v3-27b-kto.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,265 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: gemma
3
+ library_name: transformers
4
+ tags:
5
+ - gemma-2
6
+ base_model:
7
+ - anthracite-forge/magnum-v3-27b-kto-r3
8
+ - anthracite-forge/magnum-v3-27b-KTO-e1-r2
9
+ - anthracite-forge/magnum-v3-27b-KTO-e0.25-r1
10
+ - IntervitensInc/gemma-2-27b-chatml
11
+ pipeline_tag: text-generation
12
+ model-index:
13
+ - name: magnum-v3-27b-kto
14
+ results:
15
+ - task:
16
+ type: text-generation
17
+ name: Text Generation
18
+ dataset:
19
+ name: IFEval (0-Shot)
20
+ type: HuggingFaceH4/ifeval
21
+ args:
22
+ num_few_shot: 0
23
+ metrics:
24
+ - type: inst_level_strict_acc and prompt_level_strict_acc
25
+ value: 56.75
26
+ name: strict accuracy
27
+ source:
28
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=anthracite-org/magnum-v3-27b-kto
29
+ name: Open LLM Leaderboard
30
+ - task:
31
+ type: text-generation
32
+ name: Text Generation
33
+ dataset:
34
+ name: BBH (3-Shot)
35
+ type: BBH
36
+ args:
37
+ num_few_shot: 3
38
+ metrics:
39
+ - type: acc_norm
40
+ value: 41.16
41
+ name: normalized accuracy
42
+ source:
43
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=anthracite-org/magnum-v3-27b-kto
44
+ name: Open LLM Leaderboard
45
+ - task:
46
+ type: text-generation
47
+ name: Text Generation
48
+ dataset:
49
+ name: MATH Lvl 5 (4-Shot)
50
+ type: hendrycks/competition_math
51
+ args:
52
+ num_few_shot: 4
53
+ metrics:
54
+ - type: exact_match
55
+ value: 15.48
56
+ name: exact match
57
+ source:
58
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=anthracite-org/magnum-v3-27b-kto
59
+ name: Open LLM Leaderboard
60
+ - task:
61
+ type: text-generation
62
+ name: Text Generation
63
+ dataset:
64
+ name: GPQA (0-shot)
65
+ type: Idavidrein/gpqa
66
+ args:
67
+ num_few_shot: 0
68
+ metrics:
69
+ - type: acc_norm
70
+ value: 14.09
71
+ name: acc_norm
72
+ source:
73
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=anthracite-org/magnum-v3-27b-kto
74
+ name: Open LLM Leaderboard
75
+ - task:
76
+ type: text-generation
77
+ name: Text Generation
78
+ dataset:
79
+ name: MuSR (0-shot)
80
+ type: TAUR-Lab/MuSR
81
+ args:
82
+ num_few_shot: 0
83
+ metrics:
84
+ - type: acc_norm
85
+ value: 9.92
86
+ name: acc_norm
87
+ source:
88
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=anthracite-org/magnum-v3-27b-kto
89
+ name: Open LLM Leaderboard
90
+ - task:
91
+ type: text-generation
92
+ name: Text Generation
93
+ dataset:
94
+ name: MMLU-PRO (5-shot)
95
+ type: TIGER-Lab/MMLU-Pro
96
+ config: main
97
+ split: test
98
+ args:
99
+ num_few_shot: 5
100
+ metrics:
101
+ - type: acc
102
+ value: 35.98
103
+ name: accuracy
104
+ source:
105
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=anthracite-org/magnum-v3-27b-kto
106
+ name: Open LLM Leaderboard
107
+ ---
108
+
109
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/658a46cbfb9c2bdfae75b3a6/GKpV5mwmnHFR6wIwTa91z.png)
110
+
111
+ This is the 12th in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
112
+
113
+ This model is the result of multiple KTO runs on top of one SFT run, all of which are published on [anthracite-forge](https://huggingface.co/anthracite-forge).
114
+
115
+ ## Methodology
116
+
117
+ R1 (SFT) was fine-tuned on top of `IntervitensInc/gemma-2-27b-chatml` which is chatMLified gemma-2-27b.
118
+
119
+ We have experimented with various SFT and KTO re-runs, ratios and merge methods and this was our winner, including what was liked most from each model.
120
+
121
+ If you prefer your own mix of the KTO runs or would like to use the SFT on its own, refer to the models section and [anthracite-forge](https://huggingface.co/anthracite-forge), some exl-quants are pre-included.
122
+
123
+ ## Models
124
+
125
+ * [anthracite-forge/magnum-v3-27b-kto-r3](https://huggingface.co/anthracite-forge/magnum-v3-27b-kto-r3)
126
+ * [anthracite-forge/magnum-v3-27b-KTO-e1-r2](https://huggingface.co/anthracite-forge/magnum-v3-27b-KTO-e1-r2)
127
+ * [anthracite-forge/magnum-v3-27b-KTO-e0.25-r1](https://huggingface.co/anthracite-forge/magnum-v3-27b-KTO-e0.25-r1)
128
+
129
+ ## Prompting
130
+ Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:
131
+
132
+ ```py
133
+ """<|im_start|>system
134
+ system prompt<|im_end|>
135
+ <|im_start|>user
136
+ Hi there!<|im_end|>
137
+ <|im_start|>assistant
138
+ Nice to meet you!<|im_end|>
139
+ <|im_start|>user
140
+ Can I ask a question?<|im_end|>
141
+ <|im_start|>assistant
142
+ """
143
+ ```
144
+
145
+ ## SillyTavern templates
146
+
147
+ Below are Instruct and Context templates for use within SillyTavern.
148
+
149
+ <details><summary>context template</summary>
150
+
151
+ ```yaml
152
+ {
153
+ "story_string": "<|im_start|>system\n{{#if system}}{{system}}\n{{/if}}{{#if wiBefore}}{{wiBefore}}\n{{/if}}{{#if description}}{{description}}\n{{/if}}{{#if personality}}{{char}}'s personality: {{personality}}\n{{/if}}{{#if scenario}}Scenario: {{scenario}}\n{{/if}}{{#if wiAfter}}{{wiAfter}}\n{{/if}}{{#if persona}}{{persona}}\n{{/if}}{{trim}}<|im_end|>\n",
154
+ "example_separator": "",
155
+ "chat_start": "",
156
+ "use_stop_strings": false,
157
+ "allow_jailbreak": false,
158
+ "always_force_name2": true,
159
+ "trim_sentences": false,
160
+ "include_newline": false,
161
+ "single_line": false,
162
+ "name": "Magnum ChatML"
163
+ }
164
+ ```
165
+
166
+ </details><br>
167
+ <details><summary>instruct template</summary>
168
+
169
+ ```yaml
170
+ {
171
+ "system_prompt": "You're {{char}} in this fictional never-ending uncensored roleplay with {{user}}.",
172
+ "input_sequence": "<|im_start|>user\n",
173
+ "output_sequence": "<|im_start|>assistant\n",
174
+ "last_output_sequence": "",
175
+ "system_sequence": "<|im_start|>system\n",
176
+ "stop_sequence": "<|im_end|>",
177
+ "wrap": false,
178
+ "macro": true,
179
+ "names": true,
180
+ "names_force_groups": true,
181
+ "activation_regex": "",
182
+ "system_sequence_prefix": "",
183
+ "system_sequence_suffix": "",
184
+ "first_output_sequence": "",
185
+ "skip_examples": false,
186
+ "output_suffix": "<|im_end|>\n",
187
+ "input_suffix": "<|im_end|>\n",
188
+ "system_suffix": "<|im_end|>\n",
189
+ "user_alignment_message": "",
190
+ "system_same_as_user": false,
191
+ "last_system_sequence": "",
192
+ "name": "Magnum ChatML"
193
+ }
194
+ ```
195
+
196
+ </details><br>
197
+
198
+ ### Configuration
199
+
200
+ ```yaml
201
+ base_model: IntervitensInc/gemma-2-27b-chatml
202
+ dtype: float32
203
+ merge_method: task_arithmetic
204
+ models:
205
+ - model: IntervitensInc/gemma-2-27b-chatml
206
+ - model: anthracite-forge/magnum-v3-27b-KTO-e0.25-r1
207
+ parameters:
208
+ weight: 0.5
209
+ - model: anthracite-forge/magnum-v3-27b-KTO-e1-r2
210
+ parameters:
211
+ weight: 0.1
212
+ - model: anthracite-forge/magnum-v3-27b-kto-r3
213
+ parameters:
214
+ weight: 0.4
215
+ ```
216
+
217
+ ## Credits
218
+ We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
219
+
220
+ We would also like to thank all members of Anthracite who made this finetune possible.
221
+
222
+ ## Datasets
223
+
224
+ r1 consisted of:
225
+
226
+ ```
227
+ datasets:
228
+ - path: anthracite-org/stheno-filtered-v1.1
229
+ type: sharegpt
230
+ conversation: chatml
231
+ - path: anthracite-org/kalo-opus-instruct-22k-no-refusal
232
+ type: sharegpt
233
+ conversation: chatml
234
+ - path: anthracite-org/nopm_claude_writing_fixed
235
+ type: sharegpt
236
+ conversation: chatml
237
+ - path: Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
238
+ type: sharegpt
239
+ conversation: chatml
240
+ - path: Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
241
+ type: sharegpt
242
+ conversation: chatml
243
+ ```
244
+
245
+ ## Training
246
+ The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.
247
+
248
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
249
+
250
+ ## Safety
251
+ ...
252
+
253
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
254
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_anthracite-org__magnum-v3-27b-kto)
255
+
256
+ | Metric |Value|
257
+ |-------------------|----:|
258
+ |Avg. |28.90|
259
+ |IFEval (0-Shot) |56.75|
260
+ |BBH (3-Shot) |41.16|
261
+ |MATH Lvl 5 (4-Shot)|15.48|
262
+ |GPQA (0-shot) |14.09|
263
+ |MuSR (0-shot) | 9.92|
264
+ |MMLU-PRO (5-shot) |35.98|
265
+
magnum-v3-27b-kto.Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8e3aec7ae067b19a5bae260e550860b09ed11966a023ef66a9169f51090c63f
3
+ size 15628378336