zhichen commited on
Commit
9e3e33a
1 Parent(s): 2c4ec93

add github

Browse files
Files changed (2) hide show
  1. README.md +44 -3
  2. README_CN.md +43 -2
README.md CHANGED
@@ -36,7 +36,7 @@
36
 
37
  **Github:** [https://github.com/seanzhang-zhichen/llama3-chinese](https://github.com/seanzhang-zhichen/llama3-chinese)
38
 
39
- ![DEMO](./images/vllm_web_demo.png)
40
 
41
 
42
  ## Download Model
@@ -63,7 +63,6 @@ git clone https://www.modelscope.cn/LLM-Research/Meta-Llama-3-8B.git
63
  ```bash
64
  git lfs install
65
  git clone https://www.modelscope.cn/seanzhang/Llama3-Chinese-Lora.git
66
-
67
  ```
68
 
69
  **From HuggingFace**
@@ -96,6 +95,48 @@ git lfs install
96
  git clone https://huggingface.co/zhichen/Llama3-Chinese
97
  ```
98
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
99
 
100
  ## VLLM WEB DEMO
101
 
@@ -131,7 +172,7 @@ If you used Llama3-Chinese in your research, cite it in the following format:
131
  ```latex
132
  @misc{Llama3-Chinese,
133
  title={Llama3-Chinese},
134
- author={Zhichen Zhang},
135
  year={2024},
136
  howpublished={\url{https://github.com/seanzhang-zhichen/llama3-chinese}},
137
  }
 
36
 
37
  **Github:** [https://github.com/seanzhang-zhichen/llama3-chinese](https://github.com/seanzhang-zhichen/llama3-chinese)
38
 
39
+ ![DEMO](./images/web_demo.png)
40
 
41
 
42
  ## Download Model
 
63
  ```bash
64
  git lfs install
65
  git clone https://www.modelscope.cn/seanzhang/Llama3-Chinese-Lora.git
 
66
  ```
67
 
68
  **From HuggingFace**
 
95
  git clone https://huggingface.co/zhichen/Llama3-Chinese
96
  ```
97
 
98
+ ## Inference
99
+
100
+ ```python
101
+ from transformers import AutoTokenizer, AutoModelForCausalLM
102
+
103
+ model_id = "zhichen/Llama3-Chinese"
104
+
105
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
106
+ model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype="auto", device_map="auto")
107
+
108
+ messages = [
109
+ {"role": "system", "content": "You are a helpful assistant."},
110
+ {"role": "user", "content": "你好"},
111
+ ]
112
+
113
+ input_ids = tokenizer.apply_chat_template(
114
+ messages, add_generation_prompt=True, return_tensors="pt"
115
+ ).to(model.device)
116
+
117
+ outputs = model.generate(
118
+ input_ids,
119
+ max_new_tokens=2048,
120
+ do_sample=True,
121
+ temperature=0.7,
122
+ top_p=0.95,
123
+ )
124
+ response = outputs[0][input_ids.shape[-1]:]
125
+ print(tokenizer.decode(response, skip_special_tokens=True))
126
+ ```
127
+
128
+ ## CLI DEMO
129
+
130
+ ```bash
131
+ python cli_demo.py --model_path zhichen/Llama3-Chinese
132
+ ```
133
+
134
+ ## WEB DEMO
135
+
136
+ ```bash
137
+ python web_demo.py --model_path zhichen/Llama3-Chinese
138
+ ```
139
+
140
 
141
  ## VLLM WEB DEMO
142
 
 
172
  ```latex
173
  @misc{Llama3-Chinese,
174
  title={Llama3-Chinese},
175
+ author={Zhichen Zhang, Xin LU, Long Chen},
176
  year={2024},
177
  howpublished={\url{https://github.com/seanzhang-zhichen/llama3-chinese}},
178
  }
README_CN.md CHANGED
@@ -37,7 +37,7 @@
37
 
38
  **Github:** [https://github.com/seanzhang-zhichen/llama3-chinese](https://github.com/seanzhang-zhichen/llama3-chinese)
39
 
40
- ![DEMO](./images/vllm_web_demo.png)
41
 
42
 
43
  ## 模型下载
@@ -96,6 +96,47 @@ git clone https://huggingface.co/zhichen/Llama3-Chinese
96
  ```
97
 
98
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
99
 
100
 
101
  ## vllm web 推理
@@ -133,7 +174,7 @@ Llama3-Chinese项目代码的授权协议为 [The Apache License 2.0](./LICENSE)
133
  ```latex
134
  @misc{Llama3-Chinese,
135
  title={Llama3-Chinese},
136
- author={Zhichen Zhang},
137
  year={2024},
138
  howpublished={\url{https://github.com/seanzhang-zhichen/llama3-chinese}},
139
  }
 
37
 
38
  **Github:** [https://github.com/seanzhang-zhichen/llama3-chinese](https://github.com/seanzhang-zhichen/llama3-chinese)
39
 
40
+ ![DEMO](./images/web_demo.png)
41
 
42
 
43
  ## 模型下载
 
96
  ```
97
 
98
 
99
+ ## 推理
100
+
101
+ ```python
102
+ from transformers import AutoTokenizer, AutoModelForCausalLM
103
+
104
+ model_id = "zhichen/Llama3-Chinese"
105
+
106
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
107
+ model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype="auto", device_map="auto")
108
+
109
+ messages = [
110
+ {"role": "system", "content": "You are a helpful assistant."},
111
+ {"role": "user", "content": "你好"},
112
+ ]
113
+
114
+ input_ids = tokenizer.apply_chat_template(
115
+ messages, add_generation_prompt=True, return_tensors="pt"
116
+ ).to(model.device)
117
+
118
+ outputs = model.generate(
119
+ input_ids,
120
+ max_new_tokens=2048,
121
+ do_sample=True,
122
+ temperature=0.7,
123
+ top_p=0.95,
124
+ )
125
+ response = outputs[0][input_ids.shape[-1]:]
126
+ print(tokenizer.decode(response, skip_special_tokens=True))
127
+ ```
128
+
129
+ ## 命令行推理
130
+
131
+ ```bash
132
+ python cli_demo.py --model_path zhichen/Llama3-Chinese
133
+ ```
134
+
135
+ ## web推理
136
+
137
+ ```bash
138
+ python web_demo.py --model_path zhichen/Llama3-Chinese
139
+ ```
140
 
141
 
142
  ## vllm web 推理
 
174
  ```latex
175
  @misc{Llama3-Chinese,
176
  title={Llama3-Chinese},
177
+ author={Zhichen Zhang, Xin LU, Long Chen},
178
  year={2024},
179
  howpublished={\url{https://github.com/seanzhang-zhichen/llama3-chinese}},
180
  }