Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -17,15 +17,15 @@ import requests
|
|
17 |
from huggingface_hub import hf_hub_download
|
18 |
from safetensors.torch import load_file
|
19 |
from langchain.schema import AIMessage
|
20 |
-
from
|
21 |
-
from
|
22 |
-
from
|
23 |
from langchain.text_splitter import CharacterTextSplitter
|
24 |
from langchain.chains import RetrievalQA
|
25 |
|
26 |
# Initialize models and clients
|
27 |
MODEL = 'llama3-groq-70b-8192-tool-use-preview'
|
28 |
-
client = Groq(
|
29 |
|
30 |
vqa_model = AutoModel.from_pretrained('openbmb/MiniCPM-V-2', trust_remote_code=True,
|
31 |
device_map="auto", torch_dtype=torch.bfloat16)
|
|
|
17 |
from huggingface_hub import hf_hub_download
|
18 |
from safetensors.torch import load_file
|
19 |
from langchain.schema import AIMessage
|
20 |
+
from langchain_community.embeddings import HuggingFaceEmbeddings
|
21 |
+
from langchain_community.vectorstores import FAISS
|
22 |
+
from langchain_community.document_loaders import TextLoader
|
23 |
from langchain.text_splitter import CharacterTextSplitter
|
24 |
from langchain.chains import RetrievalQA
|
25 |
|
26 |
# Initialize models and clients
|
27 |
MODEL = 'llama3-groq-70b-8192-tool-use-preview'
|
28 |
+
client = Groq(api_key=os.environ.get("GROQ_API_KEY"))
|
29 |
|
30 |
vqa_model = AutoModel.from_pretrained('openbmb/MiniCPM-V-2', trust_remote_code=True,
|
31 |
device_map="auto", torch_dtype=torch.bfloat16)
|