andyfe commited on
Commit
1c5f382
·
1 Parent(s): 59fbd32

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -8,7 +8,6 @@ from huggingface_hub import Repository, InferenceClient
8
 
9
  HF_TOKEN = os.environ.get("HF_TOKEN", None)
10
  API_URL = "https://api-inference.huggingface.co/models/tiiuae/falcon-180B-chat"
11
- BOT_NAME = "Medic"
12
 
13
  STOP_SEQUENCES = ["\nUser:", "<|endoftext|>", " User:", "###"]
14
 
@@ -17,7 +16,7 @@ client = InferenceClient(
17
  headers={"Authorization": f"Bearer {HF_TOKEN}"},
18
  )
19
 
20
- def query(system_prompt, user_prompt, temperature=0.9, max_new_tokens=256, top_p=0.95, repetition_penalty=1.0):
21
  print(temperature, max_new_tokens, top_p, repetition_penalty)
22
  seed = 42
23
  generate_kwargs = dict(
@@ -29,10 +28,11 @@ def query(system_prompt, user_prompt, temperature=0.9, max_new_tokens=256, top_p
29
  do_sample=True,
30
  seed=seed,
31
  )
 
32
  print(system_prompt)
33
  print(user_prompt)
34
  print('-' * 20)
35
- prompt = f"System: {system_prompt}\nUser: {user_prompt}\Medic:"
36
  stream = client.text_generation(prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
37
  output = ""
38
  for response in stream:
 
8
 
9
  HF_TOKEN = os.environ.get("HF_TOKEN", None)
10
  API_URL = "https://api-inference.huggingface.co/models/tiiuae/falcon-180B-chat"
 
11
 
12
  STOP_SEQUENCES = ["\nUser:", "<|endoftext|>", " User:", "###"]
13
 
 
16
  headers={"Authorization": f"Bearer {HF_TOKEN}"},
17
  )
18
 
19
+ def query(bot_name, system_prompt, user_prompt, temperature=0.9, max_new_tokens=256, top_p=0.95, repetition_penalty=1.0):
20
  print(temperature, max_new_tokens, top_p, repetition_penalty)
21
  seed = 42
22
  generate_kwargs = dict(
 
28
  do_sample=True,
29
  seed=seed,
30
  )
31
+ print(bot_name)
32
  print(system_prompt)
33
  print(user_prompt)
34
  print('-' * 20)
35
+ prompt = f"System: {system_prompt}\nUser: {user_prompt}\n{bot_name}: "
36
  stream = client.text_generation(prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
37
  output = ""
38
  for response in stream: