morriszms commited on
Commit
3211dfd
·
verified ·
1 Parent(s): fb3083f

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Tucano-1b1-Instruct-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Tucano-1b1-Instruct-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Tucano-1b1-Instruct-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Tucano-1b1-Instruct-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Tucano-1b1-Instruct-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Tucano-1b1-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Tucano-1b1-Instruct-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Tucano-1b1-Instruct-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Tucano-1b1-Instruct-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Tucano-1b1-Instruct-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Tucano-1b1-Instruct-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Tucano-1b1-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,350 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - pt
4
+ license: apache-2.0
5
+ library_name: transformers
6
+ tags:
7
+ - text-generation-inference
8
+ - TensorBlock
9
+ - GGUF
10
+ datasets:
11
+ - nicholasKluge/instruct-aira-dataset-v3
12
+ - cnmoro/GPT4-500k-Augmented-PTBR-Clean
13
+ - rhaymison/orca-math-portuguese-64k
14
+ - nicholasKluge/reward-aira-dataset
15
+ metrics:
16
+ - perplexity
17
+ pipeline_tag: text-generation
18
+ widget:
19
+ - text: <instruction>Cite algumas bandas de rock brasileiras famosas.</instruction>
20
+ example_title: Exemplo
21
+ - text: <instruction>Invente uma história sobre um encanador com poderes mágicos.</instruction>
22
+ example_title: Exemplo
23
+ - text: <instruction>Qual cidade é a capital do estado do Rio Grande do Sul?</instruction>
24
+ example_title: Exemplo
25
+ - text: <instruction>Diga o nome de uma maravilha culinária característica da cosinha
26
+ Portuguesa?</instruction>
27
+ example_title: Exemplo
28
+ inference:
29
+ parameters:
30
+ repetition_penalty: 1.2
31
+ temperature: 0.2
32
+ top_k: 20
33
+ top_p: 0.2
34
+ max_new_tokens: 150
35
+ co2_eq_emissions:
36
+ emissions: 21890
37
+ source: CodeCarbon
38
+ training_type: pre-training
39
+ geographical_location: Germany
40
+ hardware_used: NVIDIA A100-SXM4-80GB
41
+ base_model: TucanoBR/Tucano-1b1-Instruct
42
+ model-index:
43
+ - name: Tucano-1b1-Instruct
44
+ results:
45
+ - task:
46
+ type: text-generation
47
+ name: Text Generation
48
+ dataset:
49
+ name: CALAME-PT
50
+ type: NOVA-vision-language/calame-pt
51
+ split: all
52
+ args:
53
+ num_few_shot: 0
54
+ metrics:
55
+ - type: acc
56
+ value: 56.55
57
+ name: accuracy
58
+ source:
59
+ url: https://huggingface.co/datasets/NOVA-vision-language/calame-pt
60
+ name: Context-Aware LAnguage Modeling Evaluation for Portuguese
61
+ - task:
62
+ type: text-generation
63
+ name: Text Generation
64
+ dataset:
65
+ name: LAMBADA-PT
66
+ type: TucanoBR/lambada-pt
67
+ split: train
68
+ args:
69
+ num_few_shot: 0
70
+ metrics:
71
+ - type: acc
72
+ value: 35.53
73
+ name: accuracy
74
+ source:
75
+ url: https://huggingface.co/datasets/TucanoBR/lambada-pt
76
+ name: LAMBADA-PT
77
+ - task:
78
+ type: text-generation
79
+ name: Text Generation
80
+ dataset:
81
+ name: ENEM Challenge (No Images)
82
+ type: eduagarcia/enem_challenge
83
+ split: train
84
+ args:
85
+ num_few_shot: 3
86
+ metrics:
87
+ - type: acc
88
+ value: 21.06
89
+ name: accuracy
90
+ source:
91
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
92
+ name: Open Portuguese LLM Leaderboard
93
+ - task:
94
+ type: text-generation
95
+ name: Text Generation
96
+ dataset:
97
+ name: BLUEX (No Images)
98
+ type: eduagarcia-temp/BLUEX_without_images
99
+ split: train
100
+ args:
101
+ num_few_shot: 3
102
+ metrics:
103
+ - type: acc
104
+ value: 26.01
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
108
+ name: Open Portuguese LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: OAB Exams
114
+ type: eduagarcia/oab_exams
115
+ split: train
116
+ args:
117
+ num_few_shot: 3
118
+ metrics:
119
+ - type: acc
120
+ value: 26.47
121
+ name: accuracy
122
+ source:
123
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
124
+ name: Open Portuguese LLM Leaderboard
125
+ - task:
126
+ type: text-generation
127
+ name: Text Generation
128
+ dataset:
129
+ name: Assin2 RTE
130
+ type: assin2
131
+ split: test
132
+ args:
133
+ num_few_shot: 15
134
+ metrics:
135
+ - type: f1_macro
136
+ value: 67.78
137
+ name: f1-macro
138
+ source:
139
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
140
+ name: Open Portuguese LLM Leaderboard
141
+ - task:
142
+ type: text-generation
143
+ name: Text Generation
144
+ dataset:
145
+ name: Assin2 STS
146
+ type: eduagarcia/portuguese_benchmark
147
+ split: test
148
+ args:
149
+ num_few_shot: 10
150
+ metrics:
151
+ - type: pearson
152
+ value: 8.88
153
+ name: pearson
154
+ source:
155
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
156
+ name: Open Portuguese LLM Leaderboard
157
+ - task:
158
+ type: text-generation
159
+ name: Text Generation
160
+ dataset:
161
+ name: FaQuAD NLI
162
+ type: ruanchaves/faquad-nli
163
+ split: test
164
+ args:
165
+ num_few_shot: 15
166
+ metrics:
167
+ - type: f1_macro
168
+ value: 43.97
169
+ name: f1-macro
170
+ source:
171
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
172
+ name: Open Portuguese LLM Leaderboard
173
+ - task:
174
+ type: text-generation
175
+ name: Text Generation
176
+ dataset:
177
+ name: HateBR Binary
178
+ type: ruanchaves/hatebr
179
+ split: test
180
+ args:
181
+ num_few_shot: 25
182
+ metrics:
183
+ - type: f1_macro
184
+ value: 31.28
185
+ name: f1-macro
186
+ source:
187
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
188
+ name: Open Portuguese LLM Leaderboard
189
+ - task:
190
+ type: text-generation
191
+ name: Text Generation
192
+ dataset:
193
+ name: PT Hate Speech Binary
194
+ type: hate_speech_portuguese
195
+ split: test
196
+ args:
197
+ num_few_shot: 25
198
+ metrics:
199
+ - type: f1_macro
200
+ value: 41.23
201
+ name: f1-macro
202
+ source:
203
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
204
+ name: Open Portuguese LLM Leaderboard
205
+ - task:
206
+ type: text-generation
207
+ name: Text Generation
208
+ dataset:
209
+ name: tweetSentBR
210
+ type: eduagarcia-temp/tweetsentbr
211
+ split: test
212
+ args:
213
+ num_few_shot: 25
214
+ metrics:
215
+ - type: f1_macro
216
+ value: 22.03
217
+ name: f1-macro
218
+ source:
219
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
220
+ name: Open Portuguese LLM Leaderboard
221
+ - task:
222
+ type: text-generation
223
+ name: Text Generation
224
+ dataset:
225
+ name: ARC-Challenge (PT)
226
+ type: arc_pt
227
+ args:
228
+ num_few_shot: 25
229
+ metrics:
230
+ - type: acc_norm
231
+ value: 30.77
232
+ name: normalized accuracy
233
+ source:
234
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
235
+ name: Evaluation Framework for Multilingual Large Language Models
236
+ - task:
237
+ type: text-generation
238
+ name: Text Generation
239
+ dataset:
240
+ name: HellaSwag (PT)
241
+ type: hellaswag_pt
242
+ args:
243
+ num_few_shot: 10
244
+ metrics:
245
+ - type: acc_norm
246
+ value: 43.5
247
+ name: normalized accuracy
248
+ source:
249
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
250
+ name: Evaluation Framework for Multilingual Large Language Models
251
+ - task:
252
+ type: text-generation
253
+ name: Text Generation
254
+ dataset:
255
+ name: TruthfulQA (PT)
256
+ type: truthfulqa_pt
257
+ args:
258
+ num_few_shot: 0
259
+ metrics:
260
+ - type: mc2
261
+ value: 41.14
262
+ name: bleurt
263
+ source:
264
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
265
+ name: Evaluation Framework for Multilingual Large Language Models
266
+ - task:
267
+ type: text-generation
268
+ name: Text Generation
269
+ dataset:
270
+ name: Alpaca-Eval (PT)
271
+ type: alpaca_eval_pt
272
+ args:
273
+ num_few_shot: 0
274
+ metrics:
275
+ - type: lc_winrate
276
+ value: 8.8
277
+ name: length controlled winrate
278
+ source:
279
+ url: https://github.com/tatsu-lab/alpaca_eval
280
+ name: AlpacaEval
281
+ ---
282
+
283
+ <div style="width: auto; margin-left: auto; margin-right: auto">
284
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
285
+ </div>
286
+ <div style="display: flex; justify-content: space-between; width: 100%;">
287
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
288
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
289
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
290
+ </p>
291
+ </div>
292
+ </div>
293
+
294
+ ## TucanoBR/Tucano-1b1-Instruct - GGUF
295
+
296
+ This repo contains GGUF format model files for [TucanoBR/Tucano-1b1-Instruct](https://huggingface.co/TucanoBR/Tucano-1b1-Instruct).
297
+
298
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4242](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
299
+
300
+ <div style="text-align: left; margin: 20px 0;">
301
+ <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
302
+ Run them on the TensorBlock client using your local machine ↗
303
+ </a>
304
+ </div>
305
+
306
+ ## Prompt template
307
+
308
+ ```
309
+ <instruction>{prompt}</instruction>
310
+ ```
311
+
312
+ ## Model file specification
313
+
314
+ | Filename | Quant type | File Size | Description |
315
+ | -------- | ---------- | --------- | ----------- |
316
+ | [Tucano-1b1-Instruct-Q2_K.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q2_K.gguf) | Q2_K | 0.432 GB | smallest, significant quality loss - not recommended for most purposes |
317
+ | [Tucano-1b1-Instruct-Q3_K_S.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q3_K_S.gguf) | Q3_K_S | 0.499 GB | very small, high quality loss |
318
+ | [Tucano-1b1-Instruct-Q3_K_M.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q3_K_M.gguf) | Q3_K_M | 0.548 GB | very small, high quality loss |
319
+ | [Tucano-1b1-Instruct-Q3_K_L.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q3_K_L.gguf) | Q3_K_L | 0.592 GB | small, substantial quality loss |
320
+ | [Tucano-1b1-Instruct-Q4_0.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q4_0.gguf) | Q4_0 | 0.637 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
321
+ | [Tucano-1b1-Instruct-Q4_K_S.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q4_K_S.gguf) | Q4_K_S | 0.640 GB | small, greater quality loss |
322
+ | [Tucano-1b1-Instruct-Q4_K_M.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q4_K_M.gguf) | Q4_K_M | 0.668 GB | medium, balanced quality - recommended |
323
+ | [Tucano-1b1-Instruct-Q5_0.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q5_0.gguf) | Q5_0 | 0.766 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
324
+ | [Tucano-1b1-Instruct-Q5_K_S.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q5_K_S.gguf) | Q5_K_S | 0.766 GB | large, low quality loss - recommended |
325
+ | [Tucano-1b1-Instruct-Q5_K_M.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q5_K_M.gguf) | Q5_K_M | 0.782 GB | large, very low quality loss - recommended |
326
+ | [Tucano-1b1-Instruct-Q6_K.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q6_K.gguf) | Q6_K | 0.903 GB | very large, extremely low quality loss |
327
+ | [Tucano-1b1-Instruct-Q8_0.gguf](https://huggingface.co/tensorblock/Tucano-1b1-Instruct-GGUF/blob/main/Tucano-1b1-Instruct-Q8_0.gguf) | Q8_0 | 1.170 GB | very large, extremely low quality loss - not recommended |
328
+
329
+
330
+ ## Downloading instruction
331
+
332
+ ### Command line
333
+
334
+ Firstly, install Huggingface Client
335
+
336
+ ```shell
337
+ pip install -U "huggingface_hub[cli]"
338
+ ```
339
+
340
+ Then, downoad the individual model file the a local directory
341
+
342
+ ```shell
343
+ huggingface-cli download tensorblock/Tucano-1b1-Instruct-GGUF --include "Tucano-1b1-Instruct-Q2_K.gguf" --local-dir MY_LOCAL_DIR
344
+ ```
345
+
346
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
347
+
348
+ ```shell
349
+ huggingface-cli download tensorblock/Tucano-1b1-Instruct-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
350
+ ```
Tucano-1b1-Instruct-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cca08f26baa3348c487d55d7fe3400073f6858fb6f4d2140b30592a870f5394
3
+ size 432170304
Tucano-1b1-Instruct-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9173b23a89e0b7e9b5968d88030d71928208426e1966f1bcea6195c8e24b5816
3
+ size 591566560
Tucano-1b1-Instruct-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84b559a4955923e98411f2edda650f8d54e8c0eebe9f22717b746eb69315df38
3
+ size 548443872
Tucano-1b1-Instruct-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6acec27b9ce5d765a8ad8ff8aa4963b917f23ae7e7ea36e06987b19058f428ef
3
+ size 499381984
Tucano-1b1-Instruct-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af9b367df4187cc3fea095971999ea87999f5123e9c7c546ebe19dec610d15fc
3
+ size 636766464
Tucano-1b1-Instruct-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eba53cf4e4d8d069137156bd1e6958d839f449cd07f6aaa581e2cb7bc4d7c582
3
+ size 667855104
Tucano-1b1-Instruct-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5040406412c1c667c17a128726c003f445c8961d7ef20b7acb5f572a25f37baa
3
+ size 639912192
Tucano-1b1-Instruct-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bd3e75df21ac0f8ac4aee28d6176ec6e29a7e83715400d698f9f9ee3997c856
3
+ size 766069504
Tucano-1b1-Instruct-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b8b287595c8264e34e1759e739b1d639d56c49964f7a60af6d8992d4f7a54d5
3
+ size 782084864
Tucano-1b1-Instruct-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c3ae55231a9e325dfcba16a72847518fb397ef046abe76645b463297f0dc6b2
3
+ size 766069504
Tucano-1b1-Instruct-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ba6b77dd0d6243eda85586b42cab9380e00fee73d036f4ed370148decc3fa05
3
+ size 903453984
Tucano-1b1-Instruct-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d17925bd8f1e1e66c31f4bbb859112ec2dadb47dba409866fc30192697f95211
3
+ size 1169851616