mradermacher commited on
Commit
6f3820d
·
verified ·
1 Parent(s): 65f884d

auto-patch README.md

Browse files
Files changed (1) hide show
  1. README.md +157 -0
README.md CHANGED
@@ -1,6 +1,163 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  <!-- ### quantize_version: 2 -->
2
  <!-- ### output_tensor_quantised: 1 -->
3
  <!-- ### convert_type: hf -->
4
  <!-- ### vocab_type: -->
5
  <!-- ### tags: -->
6
  static quants of https://huggingface.co/LLaMAX/LLaMAX2-7B-XNLI
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: LLaMAX/LLaMAX2-7B-XNLI
3
+ language:
4
+ - af
5
+ - am
6
+ - ar
7
+ - hy
8
+ - as
9
+ - ast
10
+ - az
11
+ - be
12
+ - bn
13
+ - bs
14
+ - bg
15
+ - my
16
+ - ca
17
+ - ceb
18
+ - zho
19
+ - hr
20
+ - cs
21
+ - da
22
+ - nl
23
+ - en
24
+ - et
25
+ - tl
26
+ - fi
27
+ - fr
28
+ - ff
29
+ - gl
30
+ - lg
31
+ - ka
32
+ - de
33
+ - el
34
+ - gu
35
+ - ha
36
+ - he
37
+ - hi
38
+ - hu
39
+ - is
40
+ - ig
41
+ - id
42
+ - ga
43
+ - it
44
+ - ja
45
+ - jv
46
+ - kea
47
+ - kam
48
+ - kn
49
+ - kk
50
+ - km
51
+ - ko
52
+ - ky
53
+ - lo
54
+ - lv
55
+ - ln
56
+ - lt
57
+ - luo
58
+ - lb
59
+ - mk
60
+ - ms
61
+ - ml
62
+ - mt
63
+ - mi
64
+ - mr
65
+ - mn
66
+ - ne
67
+ - ns
68
+ - no
69
+ - ny
70
+ - oc
71
+ - or
72
+ - om
73
+ - ps
74
+ - fa
75
+ - pl
76
+ - pt
77
+ - pa
78
+ - ro
79
+ - ru
80
+ - sr
81
+ - sn
82
+ - sd
83
+ - sk
84
+ - sl
85
+ - so
86
+ - ku
87
+ - es
88
+ - sw
89
+ - sv
90
+ - tg
91
+ - ta
92
+ - te
93
+ - th
94
+ - tr
95
+ - uk
96
+ - umb
97
+ - ur
98
+ - uz
99
+ - vi
100
+ - cy
101
+ - wo
102
+ - xh
103
+ - yo
104
+ - zu
105
+ library_name: transformers
106
+ license: mit
107
+ quantized_by: mradermacher
108
+ tags:
109
+ - Multilingual
110
+ ---
111
+ ## About
112
+
113
  <!-- ### quantize_version: 2 -->
114
  <!-- ### output_tensor_quantised: 1 -->
115
  <!-- ### convert_type: hf -->
116
  <!-- ### vocab_type: -->
117
  <!-- ### tags: -->
118
  static quants of https://huggingface.co/LLaMAX/LLaMAX2-7B-XNLI
119
+
120
+ <!-- provided-files -->
121
+ weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
122
+ ## Usage
123
+
124
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
125
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
126
+ more details, including on how to concatenate multi-part files.
127
+
128
+ ## Provided Quants
129
+
130
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
131
+
132
+ | Link | Type | Size/GB | Notes |
133
+ |:-----|:-----|--------:|:------|
134
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.Q2_K.gguf) | Q2_K | 2.6 | |
135
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.Q3_K_S.gguf) | Q3_K_S | 3.0 | |
136
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.Q3_K_M.gguf) | Q3_K_M | 3.4 | lower quality |
137
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.Q3_K_L.gguf) | Q3_K_L | 3.7 | |
138
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.Q4_K_S.gguf) | Q4_K_S | 4.0 | fast, recommended |
139
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.Q4_K_M.gguf) | Q4_K_M | 4.2 | fast, recommended |
140
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.Q6_K.gguf) | Q6_K | 5.6 | very good quality |
141
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.Q8_0.gguf) | Q8_0 | 7.3 | fast, best quality |
142
+ | [GGUF](https://huggingface.co/mradermacher/LLaMAX2-7B-XNLI-GGUF/resolve/main/LLaMAX2-7B-XNLI.f16.gguf) | f16 | 13.6 | 16 bpw, overkill |
143
+
144
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
145
+ types (lower is better):
146
+
147
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
148
+
149
+ And here are Artefact2's thoughts on the matter:
150
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
151
+
152
+ ## FAQ / Model Request
153
+
154
+ See https://huggingface.co/mradermacher/model_requests for some answers to
155
+ questions you might have and/or if you want some other model quantized.
156
+
157
+ ## Thanks
158
+
159
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
160
+ me use its servers and providing upgrades to my workstation to enable
161
+ this work in my free time.
162
+
163
+ <!-- end -->