morriszms commited on
Commit
df4a979
1 Parent(s): 23308c4

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Tucano-2b4-Instruct-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Tucano-2b4-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Tucano-2b4-Instruct-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Tucano-2b4-Instruct-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Tucano-2b4-Instruct-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Tucano-2b4-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Tucano-2b4-Instruct-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Tucano-2b4-Instruct-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Tucano-2b4-Instruct-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Tucano-2b4-Instruct-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Tucano-2b4-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Tucano-2b4-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,350 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - pt
4
+ license: apache-2.0
5
+ library_name: transformers
6
+ tags:
7
+ - text-generation-inference
8
+ - TensorBlock
9
+ - GGUF
10
+ datasets:
11
+ - nicholasKluge/instruct-aira-dataset-v3
12
+ - cnmoro/GPT4-500k-Augmented-PTBR-Clean
13
+ - rhaymison/orca-math-portuguese-64k
14
+ - nicholasKluge/reward-aira-dataset
15
+ metrics:
16
+ - perplexity
17
+ pipeline_tag: text-generation
18
+ widget:
19
+ - text: <instruction>Cite algumas bandas de rock brasileiras famosas.</instruction>
20
+ example_title: Exemplo
21
+ - text: <instruction>Invente uma história sobre um encanador com poderes mágicos.</instruction>
22
+ example_title: Exemplo
23
+ - text: <instruction>Qual cidade é a capital do estado do Rio Grande do Sul?</instruction>
24
+ example_title: Exemplo
25
+ - text: <instruction>Diga o nome de uma maravilha culinária característica da cosinha
26
+ Portuguesa?</instruction>
27
+ example_title: Exemplo
28
+ inference:
29
+ parameters:
30
+ repetition_penalty: 1.2
31
+ temperature: 0.2
32
+ top_k: 20
33
+ top_p: 0.2
34
+ max_new_tokens: 150
35
+ co2_eq_emissions:
36
+ emissions: 42270
37
+ source: CodeCarbon
38
+ training_type: pre-training
39
+ geographical_location: Germany
40
+ hardware_used: NVIDIA A100-SXM4-80GB
41
+ base_model: TucanoBR/Tucano-2b4-Instruct
42
+ model-index:
43
+ - name: Tucano-2b4-Instruct
44
+ results:
45
+ - task:
46
+ type: text-generation
47
+ name: Text Generation
48
+ dataset:
49
+ name: CALAME-PT
50
+ type: NOVA-vision-language/calame-pt
51
+ split: all
52
+ args:
53
+ num_few_shot: 0
54
+ metrics:
55
+ - type: acc
56
+ value: 57.66
57
+ name: accuracy
58
+ source:
59
+ url: https://huggingface.co/datasets/NOVA-vision-language/calame-pt
60
+ name: Context-Aware LAnguage Modeling Evaluation for Portuguese
61
+ - task:
62
+ type: text-generation
63
+ name: Text Generation
64
+ dataset:
65
+ name: LAMBADA-PT
66
+ type: TucanoBR/lambada-pt
67
+ split: train
68
+ args:
69
+ num_few_shot: 0
70
+ metrics:
71
+ - type: acc
72
+ value: 39.92
73
+ name: accuracy
74
+ source:
75
+ url: https://huggingface.co/datasets/TucanoBR/lambada-pt
76
+ name: LAMBADA-PT
77
+ - task:
78
+ type: text-generation
79
+ name: Text Generation
80
+ dataset:
81
+ name: ENEM Challenge (No Images)
82
+ type: eduagarcia/enem_challenge
83
+ split: train
84
+ args:
85
+ num_few_shot: 3
86
+ metrics:
87
+ - type: acc
88
+ value: 20.43
89
+ name: accuracy
90
+ source:
91
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
92
+ name: Open Portuguese LLM Leaderboard
93
+ - task:
94
+ type: text-generation
95
+ name: Text Generation
96
+ dataset:
97
+ name: BLUEX (No Images)
98
+ type: eduagarcia-temp/BLUEX_without_images
99
+ split: train
100
+ args:
101
+ num_few_shot: 3
102
+ metrics:
103
+ - type: acc
104
+ value: 22.81
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
108
+ name: Open Portuguese LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: OAB Exams
114
+ type: eduagarcia/oab_exams
115
+ split: train
116
+ args:
117
+ num_few_shot: 3
118
+ metrics:
119
+ - type: acc
120
+ value: 24.83
121
+ name: accuracy
122
+ source:
123
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
124
+ name: Open Portuguese LLM Leaderboard
125
+ - task:
126
+ type: text-generation
127
+ name: Text Generation
128
+ dataset:
129
+ name: Assin2 RTE
130
+ type: assin2
131
+ split: test
132
+ args:
133
+ num_few_shot: 15
134
+ metrics:
135
+ - type: f1_macro
136
+ value: 43.39
137
+ name: f1-macro
138
+ source:
139
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
140
+ name: Open Portuguese LLM Leaderboard
141
+ - task:
142
+ type: text-generation
143
+ name: Text Generation
144
+ dataset:
145
+ name: Assin2 STS
146
+ type: eduagarcia/portuguese_benchmark
147
+ split: test
148
+ args:
149
+ num_few_shot: 10
150
+ metrics:
151
+ - type: pearson
152
+ value: 6.31
153
+ name: pearson
154
+ source:
155
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
156
+ name: Open Portuguese LLM Leaderboard
157
+ - task:
158
+ type: text-generation
159
+ name: Text Generation
160
+ dataset:
161
+ name: FaQuAD NLI
162
+ type: ruanchaves/faquad-nli
163
+ split: test
164
+ args:
165
+ num_few_shot: 15
166
+ metrics:
167
+ - type: f1_macro
168
+ value: 43.97
169
+ name: f1-macro
170
+ source:
171
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
172
+ name: Open Portuguese LLM Leaderboard
173
+ - task:
174
+ type: text-generation
175
+ name: Text Generation
176
+ dataset:
177
+ name: HateBR Binary
178
+ type: ruanchaves/hatebr
179
+ split: test
180
+ args:
181
+ num_few_shot: 25
182
+ metrics:
183
+ - type: f1_macro
184
+ value: 27.7
185
+ name: f1-macro
186
+ source:
187
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
188
+ name: Open Portuguese LLM Leaderboard
189
+ - task:
190
+ type: text-generation
191
+ name: Text Generation
192
+ dataset:
193
+ name: PT Hate Speech Binary
194
+ type: hate_speech_portuguese
195
+ split: test
196
+ args:
197
+ num_few_shot: 25
198
+ metrics:
199
+ - type: f1_macro
200
+ value: 29.18
201
+ name: f1-macro
202
+ source:
203
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
204
+ name: Open Portuguese LLM Leaderboard
205
+ - task:
206
+ type: text-generation
207
+ name: Text Generation
208
+ dataset:
209
+ name: tweetSentBR
210
+ type: eduagarcia-temp/tweetsentbr
211
+ split: test
212
+ args:
213
+ num_few_shot: 25
214
+ metrics:
215
+ - type: f1_macro
216
+ value: 43.11
217
+ name: f1-macro
218
+ source:
219
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
220
+ name: Open Portuguese LLM Leaderboard
221
+ - task:
222
+ type: text-generation
223
+ name: Text Generation
224
+ dataset:
225
+ name: ARC-Challenge (PT)
226
+ type: arc_pt
227
+ args:
228
+ num_few_shot: 25
229
+ metrics:
230
+ - type: acc_norm
231
+ value: 32.05
232
+ name: normalized accuracy
233
+ source:
234
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
235
+ name: Evaluation Framework for Multilingual Large Language Models
236
+ - task:
237
+ type: text-generation
238
+ name: Text Generation
239
+ dataset:
240
+ name: HellaSwag (PT)
241
+ type: hellaswag_pt
242
+ args:
243
+ num_few_shot: 10
244
+ metrics:
245
+ - type: acc_norm
246
+ value: 48.28
247
+ name: normalized accuracy
248
+ source:
249
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
250
+ name: Evaluation Framework for Multilingual Large Language Models
251
+ - task:
252
+ type: text-generation
253
+ name: Text Generation
254
+ dataset:
255
+ name: TruthfulQA (PT)
256
+ type: truthfulqa_pt
257
+ args:
258
+ num_few_shot: 0
259
+ metrics:
260
+ - type: mc2
261
+ value: 38.44
262
+ name: bleurt
263
+ source:
264
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
265
+ name: Evaluation Framework for Multilingual Large Language Models
266
+ - task:
267
+ type: text-generation
268
+ name: Text Generation
269
+ dataset:
270
+ name: Alpaca-Eval (PT)
271
+ type: alpaca_eval_pt
272
+ args:
273
+ num_few_shot: 0
274
+ metrics:
275
+ - type: lc_winrate
276
+ value: 13.0
277
+ name: length controlled winrate
278
+ source:
279
+ url: https://github.com/tatsu-lab/alpaca_eval
280
+ name: AlpacaEval
281
+ ---
282
+
283
+ <div style="width: auto; margin-left: auto; margin-right: auto">
284
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
285
+ </div>
286
+ <div style="display: flex; justify-content: space-between; width: 100%;">
287
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
288
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
289
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
290
+ </p>
291
+ </div>
292
+ </div>
293
+
294
+ ## TucanoBR/Tucano-2b4-Instruct - GGUF
295
+
296
+ This repo contains GGUF format model files for [TucanoBR/Tucano-2b4-Instruct](https://huggingface.co/TucanoBR/Tucano-2b4-Instruct).
297
+
298
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4242](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
299
+
300
+ <div style="text-align: left; margin: 20px 0;">
301
+ <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
302
+ Run them on the TensorBlock client using your local machine ↗
303
+ </a>
304
+ </div>
305
+
306
+ ## Prompt template
307
+
308
+ ```
309
+ <instruction>{prompt}</instruction>
310
+ ```
311
+
312
+ ## Model file specification
313
+
314
+ | Filename | Quant type | File Size | Description |
315
+ | -------- | ---------- | --------- | ----------- |
316
+ | [Tucano-2b4-Instruct-Q2_K.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q2_K.gguf) | Q2_K | 0.933 GB | smallest, significant quality loss - not recommended for most purposes |
317
+ | [Tucano-2b4-Instruct-Q3_K_S.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q3_K_S.gguf) | Q3_K_S | 1.084 GB | very small, high quality loss |
318
+ | [Tucano-2b4-Instruct-Q3_K_M.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q3_K_M.gguf) | Q3_K_M | 1.197 GB | very small, high quality loss |
319
+ | [Tucano-2b4-Instruct-Q3_K_L.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q3_K_L.gguf) | Q3_K_L | 1.297 GB | small, substantial quality loss |
320
+ | [Tucano-2b4-Instruct-Q4_0.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q4_0.gguf) | Q4_0 | 1.397 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
321
+ | [Tucano-2b4-Instruct-Q4_K_S.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q4_K_S.gguf) | Q4_K_S | 1.408 GB | small, greater quality loss |
322
+ | [Tucano-2b4-Instruct-Q4_K_M.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q4_K_M.gguf) | Q4_K_M | 1.484 GB | medium, balanced quality - recommended |
323
+ | [Tucano-2b4-Instruct-Q5_0.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q5_0.gguf) | Q5_0 | 1.693 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
324
+ | [Tucano-2b4-Instruct-Q5_K_S.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q5_K_S.gguf) | Q5_K_S | 1.693 GB | large, low quality loss - recommended |
325
+ | [Tucano-2b4-Instruct-Q5_K_M.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q5_K_M.gguf) | Q5_K_M | 1.737 GB | large, very low quality loss - recommended |
326
+ | [Tucano-2b4-Instruct-Q6_K.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q6_K.gguf) | Q6_K | 2.007 GB | very large, extremely low quality loss |
327
+ | [Tucano-2b4-Instruct-Q8_0.gguf](https://huggingface.co/tensorblock/Tucano-2b4-Instruct-GGUF/blob/main/Tucano-2b4-Instruct-Q8_0.gguf) | Q8_0 | 2.599 GB | very large, extremely low quality loss - not recommended |
328
+
329
+
330
+ ## Downloading instruction
331
+
332
+ ### Command line
333
+
334
+ Firstly, install Huggingface Client
335
+
336
+ ```shell
337
+ pip install -U "huggingface_hub[cli]"
338
+ ```
339
+
340
+ Then, downoad the individual model file the a local directory
341
+
342
+ ```shell
343
+ huggingface-cli download tensorblock/Tucano-2b4-Instruct-GGUF --include "Tucano-2b4-Instruct-Q2_K.gguf" --local-dir MY_LOCAL_DIR
344
+ ```
345
+
346
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
347
+
348
+ ```shell
349
+ huggingface-cli download tensorblock/Tucano-2b4-Instruct-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
350
+ ```
Tucano-2b4-Instruct-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d352773c4316a6971ac0a9b0fa93a234dbe8dacc53b0b19c2a87e293cecfaa6
3
+ size 932786752
Tucano-2b4-Instruct-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77c64701c44d2c9fbe6b462a3b898a65f154a3e8e24652a4e8a1345a41a192ba
3
+ size 1296537664
Tucano-2b4-Instruct-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee4d1d3c2eaf7845c4a057fe3594bf3058029e73a836c799503d69f987322415
3
+ size 1197004864
Tucano-2b4-Instruct-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc99698095abcdd4909142107be1634fad960e2b2d618947397ec52ccb472bf8
3
+ size 1083648064
Tucano-2b4-Instruct-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:305ee92b44bed979b85fae4d3bb7b666b6aa40ce00b6b1f0a78ee167a86d453c
3
+ size 1397427936
Tucano-2b4-Instruct-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef3e03b603ec789d2ec781a922d3aeced03df210b22c31d3b6d3cdbf23a9fcd0
3
+ size 1483597536
Tucano-2b4-Instruct-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a29fb574a21d8b0d2f001a5a392a3e18feb02862fb1506bcd528878f76830f81
3
+ size 1408077536
Tucano-2b4-Instruct-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67cc4b9baad6faa7d07aa5829401a680bb9584013a37d557fd3560f358504ab1
3
+ size 1692750176
Tucano-2b4-Instruct-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:699e571f060410b96191c78263f7db44fb2d030c0e06c214d996dcb9e30853ce
3
+ size 1737140576
Tucano-2b4-Instruct-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb4a90c95de426cdb78f026665ca62ba766dae90c01b7fae3c8c1ed7d9e22fab
3
+ size 1692750176
Tucano-2b4-Instruct-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cab18a823ea06776439e58c3066e78184ec993d96c6a8dcc967b763a6eeb24b
3
+ size 2006530080
Tucano-2b4-Instruct-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fddba27cf73774bef811310dc14f1e762ec35709b0c4dd1685fcd7ca7e3938f
3
+ size 2598558112