Spaces:
Sleeping
Sleeping
Commit
·
0619ee4
1
Parent(s):
a6e2b43
Using docker image
Browse files- Dockerfile +4 -3
- __pycache__/app.cpython-311.pyc +0 -0
- app.py +8 -9
- ingest.py +4 -6
Dockerfile
CHANGED
@@ -1,8 +1,9 @@
|
|
1 |
FROM python:3.11
|
2 |
FROM docker
|
3 |
|
4 |
-
RUN
|
5 |
-
RUN docker
|
|
|
6 |
|
7 |
# Set working directory
|
8 |
WORKDIR /code
|
@@ -23,4 +24,4 @@ ENV HOME=/home/user \
|
|
23 |
WORKDIR $HOME/app
|
24 |
|
25 |
# Start Qdrant and the application
|
26 |
-
CMD ["
|
|
|
1 |
FROM python:3.11
|
2 |
FROM docker
|
3 |
|
4 |
+
RUN systemctl start docker
|
5 |
+
RUN docker pull kushagrash13/patent-image:1.0
|
6 |
+
RUN docker run -p 6333:6333 kushagrash13/patent-image:1.0
|
7 |
|
8 |
# Set working directory
|
9 |
WORKDIR /code
|
|
|
24 |
WORKDIR $HOME/app
|
25 |
|
26 |
# Start Qdrant and the application
|
27 |
+
CMD ["uvicorn app:app --host 0.0.0.0 --port 7860"]
|
__pycache__/app.cpython-311.pyc
CHANGED
Binary files a/__pycache__/app.cpython-311.pyc and b/__pycache__/app.cpython-311.pyc differ
|
|
app.py
CHANGED
@@ -1,16 +1,16 @@
|
|
|
|
|
|
|
|
1 |
from langchain import PromptTemplate
|
|
|
2 |
from langchain.chains import RetrievalQA
|
3 |
-
from langchain.
|
4 |
-
from fastapi import FastAPI, Request, Form, Response
|
5 |
from fastapi.responses import HTMLResponse
|
6 |
-
from fastapi.templating import Jinja2Templates
|
7 |
from fastapi.staticfiles import StaticFiles
|
8 |
from fastapi.encoders import jsonable_encoder
|
9 |
-
from
|
10 |
-
from
|
11 |
-
import
|
12 |
-
import json
|
13 |
-
from langchain_groq import ChatGroq
|
14 |
|
15 |
os.environ["TRANSFORMERS_FORCE_CPU"] = "true"
|
16 |
|
@@ -36,7 +36,6 @@ llm = ChatGroq(
|
|
36 |
api_key=api_key,
|
37 |
)
|
38 |
|
39 |
-
|
40 |
print("LLM Initialized....")
|
41 |
|
42 |
prompt_template = """Use the following pieces of information to answer the user's question.
|
|
|
1 |
+
import os
|
2 |
+
import json
|
3 |
+
from langchain_groq import ChatGroq
|
4 |
from langchain import PromptTemplate
|
5 |
+
from qdrant_client import QdrantClient
|
6 |
from langchain.chains import RetrievalQA
|
7 |
+
from langchain.vectorstores import Qdrant
|
|
|
8 |
from fastapi.responses import HTMLResponse
|
|
|
9 |
from fastapi.staticfiles import StaticFiles
|
10 |
from fastapi.encoders import jsonable_encoder
|
11 |
+
from fastapi.templating import Jinja2Templates
|
12 |
+
from fastapi import FastAPI, Request, Form, Response
|
13 |
+
from langchain.embeddings import SentenceTransformerEmbeddings
|
|
|
|
|
14 |
|
15 |
os.environ["TRANSFORMERS_FORCE_CPU"] = "true"
|
16 |
|
|
|
36 |
api_key=api_key,
|
37 |
)
|
38 |
|
|
|
39 |
print("LLM Initialized....")
|
40 |
|
41 |
prompt_template = """Use the following pieces of information to answer the user's question.
|
ingest.py
CHANGED
@@ -1,11 +1,9 @@
|
|
1 |
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
2 |
-
from
|
3 |
-
from
|
4 |
-
from
|
5 |
-
from
|
6 |
|
7 |
-
# embeddings = SentenceTransformerEmbeddings(model_name='NeuML/pubmedbert-base-embeddings')
|
8 |
-
# embeddings = SentenceTransformerEmbeddings(model_name='mixedbread-ai/mxbai-embed-large-v1')
|
9 |
embeddings = SentenceTransformerEmbeddings(model_name='BAAI/bge-large-en')
|
10 |
|
11 |
print(embeddings)
|
|
|
1 |
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
2 |
+
from langchain_community.document_loaders import DirectoryLoader
|
3 |
+
from langchain_community.document_loaders import PyPDFLoader
|
4 |
+
from langchain_community.vectorstores import Qdrant
|
5 |
+
from langchain_community.embeddings import SentenceTransformerEmbeddings
|
6 |
|
|
|
|
|
7 |
embeddings = SentenceTransformerEmbeddings(model_name='BAAI/bge-large-en')
|
8 |
|
9 |
print(embeddings)
|