Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -9,7 +9,7 @@ import chromadb
|
|
9 |
import os
|
10 |
import requests
|
11 |
|
12 |
-
API_URL = "https://api-inference.huggingface.co/models/meta-llama/Llama-2-7b-
|
13 |
headers = {"Authorization": f"Bearer {os.getenv('HF_Token')}"}
|
14 |
|
15 |
def query_llama(prompt):
|
@@ -40,7 +40,7 @@ def get_embedding(text):
|
|
40 |
return outputs.last_hidden_state[:, 0, :].numpy().tolist() # Take CLS token embedding
|
41 |
|
42 |
# Load LLaMA Model (Meta LLaMA 2)
|
43 |
-
llama_pipe = pipeline("text-generation", model=AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-7b-chat-hf"))
|
44 |
|
45 |
# Load a small subset (10,000 rows)
|
46 |
#dataset = load_dataset("wiki40b", "en", split="train[:1000]")
|
|
|
9 |
import os
|
10 |
import requests
|
11 |
|
12 |
+
API_URL = "https://api-inference.huggingface.co/models/meta-llama/Llama-2-7b-hf"
|
13 |
headers = {"Authorization": f"Bearer {os.getenv('HF_Token')}"}
|
14 |
|
15 |
def query_llama(prompt):
|
|
|
40 |
return outputs.last_hidden_state[:, 0, :].numpy().tolist() # Take CLS token embedding
|
41 |
|
42 |
# Load LLaMA Model (Meta LLaMA 2)
|
43 |
+
#llama_pipe = pipeline("text-generation", model=AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-7b-chat-hf"))
|
44 |
|
45 |
# Load a small subset (10,000 rows)
|
46 |
#dataset = load_dataset("wiki40b", "en", split="train[:1000]")
|