Upload folder using huggingface_hub
Browse files- .gitattributes +14 -0
- README.md +84 -0
- llamantino-2-chat-13b-hf-ita.Q2_K.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q3_K_L.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q3_K_M.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q3_K_S.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q4_0.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q4_1.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q4_K_M.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q4_K_S.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q5_0.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q5_1.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q5_K_M.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q5_K_S.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q6_K.gguf +3 -0
- llamantino-2-chat-13b-hf-ita.Q8_0.gguf +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
llamantino-2-chat-13b-hf-ita.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
llamantino-2-chat-13b-hf-ita.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
llamantino-2-chat-13b-hf-ita.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
llamantino-2-chat-13b-hf-ita.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
llamantino-2-chat-13b-hf-ita.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
llamantino-2-chat-13b-hf-ita.Q4_1.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
llamantino-2-chat-13b-hf-ita.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
llamantino-2-chat-13b-hf-ita.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
llamantino-2-chat-13b-hf-ita.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
+
llamantino-2-chat-13b-hf-ita.Q5_1.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
llamantino-2-chat-13b-hf-ita.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
llamantino-2-chat-13b-hf-ita.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
48 |
+
llamantino-2-chat-13b-hf-ita.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
49 |
+
llamantino-2-chat-13b-hf-ita.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,84 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
title: "LLaMAntino-2-chat-13b-hf-ITA Quantized in GGUF"
|
3 |
+
tags:
|
4 |
+
- GGUF
|
5 |
+
language: en
|
6 |
+
---
|
7 |
+

|
8 |
+
|
9 |
+
# Tsunemoto GGUF's of LLaMAntino-2-chat-13b-hf-ITA
|
10 |
+
|
11 |
+
This is a GGUF quantization of LLaMAntino-2-chat-13b-hf-ITA.
|
12 |
+
|
13 |
+
## Original Repo Link:
|
14 |
+
[Original Repository](https://huggingface.co/swap-uniba/LLaMAntino-2-chat-13b-hf-ITA)
|
15 |
+
|
16 |
+
## Original Model Card:
|
17 |
+
---
|
18 |
+
# Model Card for LLaMAntino-2-chat-13b-ITA
|
19 |
+
|
20 |
+
## Model description
|
21 |
+
|
22 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
23 |
+
|
24 |
+
**LLaMAntino-2-chat-13b** is a *Large Language Model (LLM)* that is an italian-adapted **LLaMA 2 chat**.
|
25 |
+
This model aims to provide Italian NLP researchers with a base model for italian dialogue use cases.
|
26 |
+
|
27 |
+
The model was trained using *QLora* and using as training data [clean_mc4_it medium](https://huggingface.co/datasets/gsarti/clean_mc4_it/viewer/medium).
|
28 |
+
If you are interested in more details regarding the training procedure, you can find the code we used at the following link:
|
29 |
+
- **Repository:** https://github.com/swapUniba/LLaMAntino
|
30 |
+
|
31 |
+
**NOTICE**: the code has not been released yet, we apologize for the delay, it will be available asap!
|
32 |
+
|
33 |
+
- **Developed by:** Pierpaolo Basile, Elio Musacchio, Marco Polignano, Lucia Siciliani, Giuseppe Fiameni, Giovanni Semeraro
|
34 |
+
- **Funded by:** PNRR project FAIR - Future AI Research
|
35 |
+
- **Compute infrastructure:** [Leonardo](https://www.hpc.cineca.it/systems/hardware/leonardo/) supercomputer
|
36 |
+
- **Model type:** LLaMA 2 chat
|
37 |
+
- **Language(s) (NLP):** Italian
|
38 |
+
- **License:** Llama 2 Community License
|
39 |
+
- **Finetuned from model:** [NousResearch/Llama-2-13b-chat-hf](https://huggingface.co/NousResearch/Llama-2-13b-chat-hf)
|
40 |
+
|
41 |
+
## How to Get Started with the Model
|
42 |
+
|
43 |
+
Below you can find an example of model usage:
|
44 |
+
|
45 |
+
```python
|
46 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
47 |
+
|
48 |
+
model_id = "swap-uniba/LLaMAntino-2-chat-13b-hf-ITA"
|
49 |
+
|
50 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
51 |
+
model = AutoModelForCausalLM.from_pretrained(model_id)
|
52 |
+
|
53 |
+
prompt = "Scrivi qui un possibile prompt"
|
54 |
+
|
55 |
+
input_ids = tokenizer(prompt, return_tensors="pt").input_ids
|
56 |
+
outputs = model.generate(input_ids=input_ids)
|
57 |
+
|
58 |
+
print(tokenizer.batch_decode(outputs.detach().cpu().numpy()[:, input_ids.shape[1]:], skip_special_tokens=True)[0])
|
59 |
+
```
|
60 |
+
|
61 |
+
If you are facing issues when loading the model, you can try to load it quantized:
|
62 |
+
|
63 |
+
```python
|
64 |
+
model = AutoModelForCausalLM.from_pretrained(model_id, load_in_8bit=True)
|
65 |
+
```
|
66 |
+
|
67 |
+
*Note*: The model loading strategy above requires the [*bitsandbytes*](https://pypi.org/project/bitsandbytes/) and [*accelerate*](https://pypi.org/project/accelerate/) libraries
|
68 |
+
|
69 |
+
## Citation
|
70 |
+
|
71 |
+
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
72 |
+
|
73 |
+
If you use this model in your research, please cite the following:
|
74 |
+
|
75 |
+
```bibtex
|
76 |
+
@misc{basile2023llamantino,
|
77 |
+
title={LLaMAntino: LLaMA 2 Models for Effective Text Generation in Italian Language},
|
78 |
+
author={Pierpaolo Basile and Elio Musacchio and Marco Polignano and Lucia Siciliani and Giuseppe Fiameni and Giovanni Semeraro},
|
79 |
+
year={2023},
|
80 |
+
eprint={2312.09993},
|
81 |
+
archivePrefix={arXiv},
|
82 |
+
primaryClass={cs.CL}
|
83 |
+
}
|
84 |
+
```
|
llamantino-2-chat-13b-hf-ita.Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cdee2b4fcb568d53b449b802630ac9e7539c6bdf3a8df35eaff1aed0d147578b
|
3 |
+
size 5429348352
|
llamantino-2-chat-13b-hf-ita.Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26a28151b370ad1416c946a02861e2ed33d0b2f905da1545556e44ae26e45f38
|
3 |
+
size 6929559552
|
llamantino-2-chat-13b-hf-ita.Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2449600464cb566936888831fdfcfad6f2f471d46c722045e5aaa1eaa72a59a6
|
3 |
+
size 6337769472
|
llamantino-2-chat-13b-hf-ita.Q3_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2327eb52ae85116773cc86209ce0ebe22ab458fff54ce11a6e7ea6ab5dc14bd
|
3 |
+
size 5658980352
|
llamantino-2-chat-13b-hf-ita.Q4_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:62ffab0c7399396d26c1264cf83a3cf8d093296a7592d0d8f7f6f984623bed11
|
3 |
+
size 7365834752
|
llamantino-2-chat-13b-hf-ita.Q4_1.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e1d7e4f2614bb16d126f61359cc6c1f0caebeb092cac314f57a8b9965d6224e
|
3 |
+
size 8169060352
|
llamantino-2-chat-13b-hf-ita.Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:181b1cd5f419fe8cf40b60a6017b25ee30e712f436fa5e7b69002bb67e622505
|
3 |
+
size 7865956352
|
llamantino-2-chat-13b-hf-ita.Q4_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0022af42c01e2a385a968ffddbb16142b6f0e6584ddae20fed90a44742acbcf7
|
3 |
+
size 7414331392
|
llamantino-2-chat-13b-hf-ita.Q5_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:810f2161edd64550bcd3d15b3753b8aab15e8e2960c15222e7924f6b671cbcd3
|
3 |
+
size 8972285952
|
llamantino-2-chat-13b-hf-ita.Q5_1.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4cf9056e400cac26588788f6fd414b2a91a6033046725f252ce85f9a381ff4bb
|
3 |
+
size 9775511552
|
llamantino-2-chat-13b-hf-ita.Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:247dcc92236e498654f1db0c88e1133ff35afb337ed37da8948331888cc8de96
|
3 |
+
size 9229924352
|
llamantino-2-chat-13b-hf-ita.Q5_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37abf2e1caf3c09e4cc032e3a3207a6fb810848344d3f6fd66efb8cb3aa05939
|
3 |
+
size 8972285952
|
llamantino-2-chat-13b-hf-ita.Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3380c11cd7f8f26c5f172a0316c90fa7af79aea12927da3dc51ed32f430cf48b
|
3 |
+
size 10679140352
|
llamantino-2-chat-13b-hf-ita.Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1cf3b54be6629da89838b15ce1237c2af89bb9a7db9e40ea38777db229331856
|
3 |
+
size 13831319552
|