Update app.py
Browse files
app.py
CHANGED
@@ -39,4 +39,20 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
39 |
|
40 |
rich.print(f"{model=}")
|
41 |
|
42 |
-
logger.info("done")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
39 |
|
40 |
rich.print(f"{model=}")
|
41 |
|
42 |
+
logger.info("done")
|
43 |
+
|
44 |
+
# ========
|
45 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
46 |
+
from transformers.generation.utils import GenerationConfig
|
47 |
+
tokenizer = AutoTokenizer.from_pretrained("baichuan-inc/Baichuan2-13B-Chat-4bits", use_fast=False, trust_remote_code=True)
|
48 |
+
|
49 |
+
# model = AutoModelForCausalLM.from_pretrained("baichuan-inc/Baichuan2-13B-Chat-4bits", device_map="auto", torch_dtype=torch.bfloat16, trust_remote_code=True)
|
50 |
+
|
51 |
+
model.generation_config = GenerationConfig.from_pretrained("baichuan-inc/Baichuan2-13B-Chat-4bits")
|
52 |
+
messages = []
|
53 |
+
messages.append({"role": "user", "content": "解释一下“温故而知新”"})
|
54 |
+
response = model.chat(tokenizer, messages)
|
55 |
+
|
56 |
+
rich.print(response)
|
57 |
+
|
58 |
+
logger.info(f"{response=}")
|