Commit
·
c32086b
1
Parent(s):
0e7e454
Update README.md
Browse files
README.md
CHANGED
@@ -43,11 +43,11 @@ model = transformers.LlamaForCausalLM.from_pretrained(model_name, torch_dtype=to
|
|
43 |
def stream(user_prompt):
|
44 |
runtimeFlag = "cuda:0"
|
45 |
system_prompt = ''
|
46 |
-
B_INST, E_INST = "[INST]", "[/INST]"
|
47 |
prompt = f"{system_prompt}{B_INST}{user_prompt.strip()}\n{E_INST}"
|
48 |
inputs = tokenizer([prompt], return_tensors="pt").to(runtimeFlag)
|
49 |
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
|
50 |
-
_ = model.generate(**inputs, streamer=streamer, max_new_tokens=
|
51 |
|
52 |
domanda = """Quale è la media tra questi valori?"""
|
53 |
contesto = """
|
|
|
43 |
def stream(user_prompt):
|
44 |
runtimeFlag = "cuda:0"
|
45 |
system_prompt = ''
|
46 |
+
B_INST, E_INST = "<s> [INST]", "[/INST]"
|
47 |
prompt = f"{system_prompt}{B_INST}{user_prompt.strip()}\n{E_INST}"
|
48 |
inputs = tokenizer([prompt], return_tensors="pt").to(runtimeFlag)
|
49 |
streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
|
50 |
+
_ = model.generate(**inputs, streamer=streamer, max_new_tokens=200, temperature=0.0001, eos_token_id=2, num_return_sequences=1)
|
51 |
|
52 |
domanda = """Quale è la media tra questi valori?"""
|
53 |
contesto = """
|