fullstuckdev commited on
Commit
2e3ad26
·
1 Parent(s): 60dcdbe

change model

Browse files
Files changed (1) hide show
  1. app.py +2 -6
app.py CHANGED
@@ -2,9 +2,6 @@ import gradio as gr
2
  import os
3
  from openai import OpenAI
4
 
5
- API_URL = "https://api-inference.huggingface.co/models/nvidia/Llama-3.1-Nemotron-70B-Instruct-HF"
6
- headers = {"Authorization": f"Bearer {os.getenv('HUGGINGFACE_API_KEY')}"}
7
-
8
  async def generate_response(user_input):
9
  client = OpenAI(
10
  base_url="https://api-inference.huggingface.co/v1/",
@@ -16,10 +13,9 @@ async def generate_response(user_input):
16
  ]
17
 
18
  response = client.chat.completions.create(
19
- model="meta-llama/Llama-3.1-70B-Instruct",
20
  messages=messages,
21
- max_tokens=500,
22
- stream=False
23
  )
24
 
25
  return response.choices[0].message.content
 
2
  import os
3
  from openai import OpenAI
4
 
 
 
 
5
  async def generate_response(user_input):
6
  client = OpenAI(
7
  base_url="https://api-inference.huggingface.co/v1/",
 
13
  ]
14
 
15
  response = client.chat.completions.create(
16
+ model= 'nvidia/Llama-3.1-Nemotron-70B-Instruct-HF',
17
  messages=messages,
18
+ max_tokens= 500,
 
19
  )
20
 
21
  return response.choices[0].message.content