Spaces:
Sleeping
Sleeping
Update main.py
Browse files
main.py
CHANGED
@@ -1,16 +1,10 @@
|
|
1 |
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
from fastapi import FastAPI, File, UploadFile
|
9 |
import os
|
10 |
import shutil
|
11 |
import tempfile
|
12 |
from langchain_community.document_loaders import PyMuPDFLoader
|
13 |
-
from
|
14 |
from langchain_community.embeddings import HuggingFaceEmbeddings
|
15 |
from langchain_community.vectorstores import FAISS
|
16 |
from langchain.chains import ConversationalRetrievalChain
|
@@ -43,7 +37,7 @@ async def PromptLLM(file: UploadFile = File(...)):
|
|
43 |
db.save_local(DB_FAISS_PATH)
|
44 |
|
45 |
# Load the language model
|
46 |
-
llm =
|
47 |
|
48 |
# Create a conversational chain
|
49 |
chain = ConversationalRetrievalChain.from_llm(llm=llm, retriever=db.as_retriever())
|
|
|
1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
from fastapi import FastAPI, File, UploadFile
|
3 |
import os
|
4 |
import shutil
|
5 |
import tempfile
|
6 |
from langchain_community.document_loaders import PyMuPDFLoader
|
7 |
+
from langchain_community.llms import CTransformers
|
8 |
from langchain_community.embeddings import HuggingFaceEmbeddings
|
9 |
from langchain_community.vectorstores import FAISS
|
10 |
from langchain.chains import ConversationalRetrievalChain
|
|
|
37 |
db.save_local(DB_FAISS_PATH)
|
38 |
|
39 |
# Load the language model
|
40 |
+
llm = CTransformers(model="llama-2-7b-chat.ggmlv3.q4_1.bin", model_type="llama",config={'max_new_tokens': 1024, 'context_length': 2048, 'temperature': 0.01})
|
41 |
|
42 |
# Create a conversational chain
|
43 |
chain = ConversationalRetrievalChain.from_llm(llm=llm, retriever=db.as_retriever())
|