morriszms commited on
Commit
1d75667
1 Parent(s): 502b452

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ zephyr-7b-beta-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ zephyr-7b-beta-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ zephyr-7b-beta-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ zephyr-7b-beta-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ zephyr-7b-beta-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ zephyr-7b-beta-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ zephyr-7b-beta-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ zephyr-7b-beta-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ zephyr-7b-beta-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ zephyr-7b-beta-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ zephyr-7b-beta-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ zephyr-7b-beta-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,238 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - generated_from_trainer
4
+ - TensorBlock
5
+ - GGUF
6
+ license: mit
7
+ datasets:
8
+ - HuggingFaceH4/ultrachat_200k
9
+ - HuggingFaceH4/ultrafeedback_binarized
10
+ language:
11
+ - en
12
+ base_model: HuggingFaceH4/zephyr-7b-beta
13
+ widget:
14
+ - example_title: Pirate!
15
+ messages:
16
+ - role: system
17
+ content: You are a pirate chatbot who always responds with Arr!
18
+ - role: user
19
+ content: There's a llama on my lawn, how can I get rid of him?
20
+ output:
21
+ text: Arr! 'Tis a puzzlin' matter, me hearty! A llama on yer lawn be a rare sight,
22
+ but I've got a plan that might help ye get rid of 'im. Ye'll need to gather
23
+ some carrots and hay, and then lure the llama away with the promise of a tasty
24
+ treat. Once he's gone, ye can clean up yer lawn and enjoy the peace and quiet
25
+ once again. But beware, me hearty, for there may be more llamas where that one
26
+ came from! Arr!
27
+ pipeline_tag: text-generation
28
+ model-index:
29
+ - name: zephyr-7b-beta
30
+ results:
31
+ - task:
32
+ type: text-generation
33
+ name: Text Generation
34
+ dataset:
35
+ name: AI2 Reasoning Challenge (25-Shot)
36
+ type: ai2_arc
37
+ config: ARC-Challenge
38
+ split: test
39
+ args:
40
+ num_few_shot: 25
41
+ metrics:
42
+ - type: acc_norm
43
+ value: 62.03071672354948
44
+ name: normalized accuracy
45
+ source:
46
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceH4/zephyr-7b-beta
47
+ name: Open LLM Leaderboard
48
+ - task:
49
+ type: text-generation
50
+ name: Text Generation
51
+ dataset:
52
+ name: HellaSwag (10-Shot)
53
+ type: hellaswag
54
+ split: validation
55
+ args:
56
+ num_few_shot: 10
57
+ metrics:
58
+ - type: acc_norm
59
+ value: 84.35570603465445
60
+ name: normalized accuracy
61
+ source:
62
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceH4/zephyr-7b-beta
63
+ name: Open LLM Leaderboard
64
+ - task:
65
+ type: text-generation
66
+ name: Text Generation
67
+ dataset:
68
+ name: Drop (3-Shot)
69
+ type: drop
70
+ split: validation
71
+ args:
72
+ num_few_shot: 3
73
+ metrics:
74
+ - type: f1
75
+ value: 9.66243708053691
76
+ name: f1 score
77
+ source:
78
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceH4/zephyr-7b-beta
79
+ name: Open LLM Leaderboard
80
+ - task:
81
+ type: text-generation
82
+ name: Text Generation
83
+ dataset:
84
+ name: TruthfulQA (0-shot)
85
+ type: truthful_qa
86
+ config: multiple_choice
87
+ split: validation
88
+ args:
89
+ num_few_shot: 0
90
+ metrics:
91
+ - type: mc2
92
+ value: 57.44916942762855
93
+ source:
94
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceH4/zephyr-7b-beta
95
+ name: Open LLM Leaderboard
96
+ - task:
97
+ type: text-generation
98
+ name: Text Generation
99
+ dataset:
100
+ name: GSM8k (5-shot)
101
+ type: gsm8k
102
+ config: main
103
+ split: test
104
+ args:
105
+ num_few_shot: 5
106
+ metrics:
107
+ - type: acc
108
+ value: 12.736921910538287
109
+ name: accuracy
110
+ source:
111
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceH4/zephyr-7b-beta
112
+ name: Open LLM Leaderboard
113
+ - task:
114
+ type: text-generation
115
+ name: Text Generation
116
+ dataset:
117
+ name: MMLU (5-Shot)
118
+ type: cais/mmlu
119
+ config: all
120
+ split: test
121
+ args:
122
+ num_few_shot: 5
123
+ metrics:
124
+ - type: acc
125
+ value: 61.07
126
+ name: accuracy
127
+ source:
128
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceH4/zephyr-7b-beta
129
+ name: Open LLM Leaderboard
130
+ - task:
131
+ type: text-generation
132
+ name: Text Generation
133
+ dataset:
134
+ name: Winogrande (5-shot)
135
+ type: winogrande
136
+ config: winogrande_xl
137
+ split: validation
138
+ args:
139
+ num_few_shot: 5
140
+ metrics:
141
+ - type: acc
142
+ value: 77.7426992896606
143
+ name: accuracy
144
+ source:
145
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=HuggingFaceH4/zephyr-7b-beta
146
+ name: Open LLM Leaderboard
147
+ - task:
148
+ type: text-generation
149
+ name: Text Generation
150
+ dataset:
151
+ name: AlpacaEval
152
+ type: tatsu-lab/alpaca_eval
153
+ metrics:
154
+ - type: unknown
155
+ value: 0.906
156
+ name: win rate
157
+ source:
158
+ url: https://tatsu-lab.github.io/alpaca_eval/
159
+ - task:
160
+ type: text-generation
161
+ name: Text Generation
162
+ dataset:
163
+ name: MT-Bench
164
+ type: unknown
165
+ metrics:
166
+ - type: unknown
167
+ value: 7.34
168
+ name: score
169
+ source:
170
+ url: https://huggingface.co/spaces/lmsys/mt-bench
171
+ ---
172
+
173
+ <div style="width: auto; margin-left: auto; margin-right: auto">
174
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
175
+ </div>
176
+ <div style="display: flex; justify-content: space-between; width: 100%;">
177
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
178
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
179
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
180
+ </p>
181
+ </div>
182
+ </div>
183
+
184
+ ## HuggingFaceH4/zephyr-7b-beta - GGUF
185
+
186
+ This repo contains GGUF format model files for [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta).
187
+
188
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
189
+
190
+ ## Prompt template
191
+
192
+ ```
193
+ <|system|>
194
+ {system_prompt}</s>
195
+ <|user|>
196
+ {prompt}</s>
197
+ <|assistant|>
198
+ ```
199
+
200
+ ## Model file specification
201
+
202
+ | Filename | Quant type | File Size | Description |
203
+ | -------- | ---------- | --------- | ----------- |
204
+ | [zephyr-7b-beta-Q2_K.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q2_K.gguf) | Q2_K | 2.532 GB | smallest, significant quality loss - not recommended for most purposes |
205
+ | [zephyr-7b-beta-Q3_K_S.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q3_K_S.gguf) | Q3_K_S | 2.947 GB | very small, high quality loss |
206
+ | [zephyr-7b-beta-Q3_K_M.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q3_K_M.gguf) | Q3_K_M | 3.277 GB | very small, high quality loss |
207
+ | [zephyr-7b-beta-Q3_K_L.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q3_K_L.gguf) | Q3_K_L | 3.560 GB | small, substantial quality loss |
208
+ | [zephyr-7b-beta-Q4_0.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q4_0.gguf) | Q4_0 | 3.827 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
209
+ | [zephyr-7b-beta-Q4_K_S.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q4_K_S.gguf) | Q4_K_S | 3.856 GB | small, greater quality loss |
210
+ | [zephyr-7b-beta-Q4_K_M.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q4_K_M.gguf) | Q4_K_M | 4.068 GB | medium, balanced quality - recommended |
211
+ | [zephyr-7b-beta-Q5_0.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q5_0.gguf) | Q5_0 | 4.654 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
212
+ | [zephyr-7b-beta-Q5_K_S.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q5_K_S.gguf) | Q5_K_S | 4.654 GB | large, low quality loss - recommended |
213
+ | [zephyr-7b-beta-Q5_K_M.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q5_K_M.gguf) | Q5_K_M | 4.779 GB | large, very low quality loss - recommended |
214
+ | [zephyr-7b-beta-Q6_K.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q6_K.gguf) | Q6_K | 5.534 GB | very large, extremely low quality loss |
215
+ | [zephyr-7b-beta-Q8_0.gguf](https://huggingface.co/tensorblock/zephyr-7b-beta-GGUF/tree/main/zephyr-7b-beta-Q8_0.gguf) | Q8_0 | 7.167 GB | very large, extremely low quality loss - not recommended |
216
+
217
+
218
+ ## Downloading instruction
219
+
220
+ ### Command line
221
+
222
+ Firstly, install Huggingface Client
223
+
224
+ ```shell
225
+ pip install -U "huggingface_hub[cli]"
226
+ ```
227
+
228
+ Then, downoad the individual model file the a local directory
229
+
230
+ ```shell
231
+ huggingface-cli download tensorblock/zephyr-7b-beta-GGUF --include "zephyr-7b-beta-Q2_K.gguf" --local-dir MY_LOCAL_DIR
232
+ ```
233
+
234
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
235
+
236
+ ```shell
237
+ huggingface-cli download tensorblock/zephyr-7b-beta-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
238
+ ```
zephyr-7b-beta-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f5a1dca3543a5487c0de63a377d9614e68b2bc2e4df87bc422db752b6bcfc52
3
+ size 2719243264
zephyr-7b-beta-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32433d71d0b423eada1626608a7ed7a7e0be4bfc2790578276fde59e8804489a
3
+ size 3822025728
zephyr-7b-beta-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f6519295258361da96da2e51f8e4733301e5ca26cdd15876b01350a88d6072a
3
+ size 3518987264
zephyr-7b-beta-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc8f38905cf4665c5e3fcbff7a94c419ddd6437d56a3c7eda069a8a648adb616
3
+ size 3164568576
zephyr-7b-beta-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a9895a27bbc4a0ec3b7adfe17cd90c85a12bbfb245677004258ecae6e13d33b
3
+ size 4108917760
zephyr-7b-beta-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0ecd7186ec22b778831dc274db95755d2feaa63cf44289ec3aa60d81729d1a2
3
+ size 4368440320
zephyr-7b-beta-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d44a6b89eb341e09d2ecc13fd6f9d009e275f3546ce3a95110f0decdb648747
3
+ size 4140375040
zephyr-7b-beta-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af6fd69af5b0f834dc02bd5591f7d21ca98f01abc8fcc0985a5fb6de75039eb7
3
+ size 4997716992
zephyr-7b-beta-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56d437bea0ff62e2565529813441cdf56353d7f6d2392bdf9eba6a87a4bfbed3
3
+ size 5131410432
zephyr-7b-beta-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8146e47e3b24f56d926548898d1853360a0a61780d56c307d7ca39bdf5c29578
3
+ size 4997716992
zephyr-7b-beta-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9416265d60c7f69d139a660cf3ba26009492fb818162c1f193be6c3150d6c7ab
3
+ size 5942066176
zephyr-7b-beta-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aa693a2fb69d7587b0c664a0e7d0f7fdc18eb0bb3ff4acec57add63a782c273
3
+ size 7695858688