Spaces:
Running
on
Zero
Running
on
Zero
respond_to_question_with_llama
Browse files
app.py
CHANGED
@@ -44,7 +44,7 @@ def respond_to_question_llama(transcript, question):
|
|
44 |
|
45 |
client = InferenceClient(
|
46 |
"meta-llama/Meta-Llama-3.1-8B-Instruct",
|
47 |
-
token=os.environ["
|
48 |
)
|
49 |
|
50 |
response = client.chat_completion(
|
|
|
44 |
|
45 |
client = InferenceClient(
|
46 |
"meta-llama/Meta-Llama-3.1-8B-Instruct",
|
47 |
+
token=os.environ["HUGGINGFACEHUB_API_TOKEN"],
|
48 |
)
|
49 |
|
50 |
response = client.chat_completion(
|