RioShiina commited on
Commit
73c24ac
·
verified ·
1 Parent(s): 3c0c159

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +79 -3
README.md CHANGED
@@ -1,3 +1,79 @@
1
- ---
2
- license: llama3.1
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3
3
+ base_model_relation: quantized
4
+ license:
5
+ - llama3.1
6
+ - gemma
7
+ language:
8
+ - ja
9
+ - en
10
+ ---
11
+
12
+ **[2.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.3-exl2/tree/2.0bpw)**
13
+ **[3.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.3-exl2/tree/3.0bpw)**
14
+ **[4.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.3-exl2/tree/4.0bpw)**
15
+ **[5.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.3-exl2/tree/5.0bpw)**
16
+ **[6.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.3-exl2/tree/6.0bpw)**
17
+ **[7.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.3-exl2/tree/7.0bpw)**
18
+ **[8.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.3-exl2/tree/8.0bpw)**
19
+
20
+ # Llama-3.1-Swallow-70B-Instruct-v0.3-exl2
21
+ - Model creator: [tokyotech-llm](https://huggingface.co/tokyotech-llm)
22
+ - Original model: [Llama-3.1-Swallow-70B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3)
23
+
24
+ ### License
25
+
26
+ [META LLAMA 3.1 COMMUNITY LICENSE](https://www.llama.com/llama3_1/license/) and [Gemma Terms of Use](https://ai.google.dev/gemma/terms)
27
+
28
+ ## Prompt template
29
+
30
+ ```
31
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
32
+
33
+ あなたは誠実で優秀な日本人のアシスタントです。<|eot_id|><|start_header_id|>user<|end_header_id|>
34
+
35
+ 東京の紅葉した公園で、東京タワーと高層ビルを背景に、空を舞うツバメと草地に佇むラマが出会う温かな物語を書いてください。<|eot_id|><|start_header_id|>assistant<|end_header_id|>
36
+
37
+ ```
38
+
39
+ ### Citations
40
+
41
+ ```tex
42
+ @inproceedings{Fujii:COLM2024,
43
+ title={Continual Pre-Training for Cross-Lingual LLM Adaptation:
44
+ Enhancing Japanese Language Capabilities},
45
+ author={Kazuki Fujii and Taishi Nakamura and Mengsay Loem and Hiroki
46
+ Iida and Masanari Ohi and Kakeru Hattori and Hirai Shota and Sakae
47
+ Mizuki and Rio Yokota and Naoaki Okazaki},
48
+ booktitle="Proceedings of the First Conference on Language Modeling",
49
+ series={COLM},
50
+ pages="(to appear)",
51
+ year="2024",
52
+ month=oct,
53
+ address={University of Pennsylvania, USA},
54
+ }
55
+
56
+ @inproceedings{Okazaki:COLM2024,
57
+ title={Building a Large Japanese Web Corpus for Large Language Models},
58
+ author={Naoaki Okazaki and Kakeru Hattori and Hirai Shota and Hiroki
59
+ Iida and Masanari Ohi and Kazuki Fujii and Taishi Nakamura and Mengsay
60
+ Loem and Rio Yokota and Sakae Mizuki},
61
+ booktitle="Proceedings of the First Conference on Language Modeling",
62
+ series={COLM},
63
+ pages="(to appear)",
64
+ year="2024",
65
+ month=oct,
66
+ address={University of Pennsylvania, USA},
67
+ }
68
+
69
+ @misc{dubey2024llama3herdmodels,
70
+ title={The Llama 3 Herd of Models},
71
+ author={Abhimanyu Dubey and Abhinav Jauhri and Abhinav Pandey and Abhishek Kadian and Ahmad Al-Dahle and Aiesha Letman and Akhil Mathur and Alan Schelten and Amy Yang and Angela Fan et al.},
72
+ year={2024},
73
+ eprint={2407.21783},
74
+ archivePrefix={arXiv},
75
+ primaryClass={cs.AI},
76
+ url={https://arxiv.org/abs/2407.21783},
77
+ }
78
+
79
+ ```