Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -8,7 +8,6 @@ from huggingface_hub import Repository, InferenceClient
|
|
8 |
|
9 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
10 |
API_URL = "https://api-inference.huggingface.co/models/tiiuae/falcon-180B-chat"
|
11 |
-
BOT_NAME = "Medic"
|
12 |
|
13 |
STOP_SEQUENCES = ["\nUser:", "<|endoftext|>", " User:", "###"]
|
14 |
|
@@ -17,7 +16,7 @@ client = InferenceClient(
|
|
17 |
headers={"Authorization": f"Bearer {HF_TOKEN}"},
|
18 |
)
|
19 |
|
20 |
-
def query(system_prompt, user_prompt, temperature=0.9, max_new_tokens=256, top_p=0.95, repetition_penalty=1.0):
|
21 |
print(temperature, max_new_tokens, top_p, repetition_penalty)
|
22 |
seed = 42
|
23 |
generate_kwargs = dict(
|
@@ -29,10 +28,11 @@ def query(system_prompt, user_prompt, temperature=0.9, max_new_tokens=256, top_p
|
|
29 |
do_sample=True,
|
30 |
seed=seed,
|
31 |
)
|
|
|
32 |
print(system_prompt)
|
33 |
print(user_prompt)
|
34 |
print('-' * 20)
|
35 |
-
prompt = f"System: {system_prompt}\nUser: {user_prompt}\
|
36 |
stream = client.text_generation(prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
|
37 |
output = ""
|
38 |
for response in stream:
|
|
|
8 |
|
9 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
10 |
API_URL = "https://api-inference.huggingface.co/models/tiiuae/falcon-180B-chat"
|
|
|
11 |
|
12 |
STOP_SEQUENCES = ["\nUser:", "<|endoftext|>", " User:", "###"]
|
13 |
|
|
|
16 |
headers={"Authorization": f"Bearer {HF_TOKEN}"},
|
17 |
)
|
18 |
|
19 |
+
def query(bot_name, system_prompt, user_prompt, temperature=0.9, max_new_tokens=256, top_p=0.95, repetition_penalty=1.0):
|
20 |
print(temperature, max_new_tokens, top_p, repetition_penalty)
|
21 |
seed = 42
|
22 |
generate_kwargs = dict(
|
|
|
28 |
do_sample=True,
|
29 |
seed=seed,
|
30 |
)
|
31 |
+
print(bot_name)
|
32 |
print(system_prompt)
|
33 |
print(user_prompt)
|
34 |
print('-' * 20)
|
35 |
+
prompt = f"System: {system_prompt}\nUser: {user_prompt}\n{bot_name}: "
|
36 |
stream = client.text_generation(prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
|
37 |
output = ""
|
38 |
for response in stream:
|