kushagrasharma-13 commited on
Commit
0619ee4
·
1 Parent(s): a6e2b43

Using docker image

Browse files
Files changed (4) hide show
  1. Dockerfile +4 -3
  2. __pycache__/app.cpython-311.pyc +0 -0
  3. app.py +8 -9
  4. ingest.py +4 -6
Dockerfile CHANGED
@@ -1,8 +1,9 @@
1
  FROM python:3.11
2
  FROM docker
3
 
4
- RUN docker pull qdrant/qdrant
5
- RUN docker run -p 6333:6333 qdrant/qdrant
 
6
 
7
  # Set working directory
8
  WORKDIR /code
@@ -23,4 +24,4 @@ ENV HOME=/home/user \
23
  WORKDIR $HOME/app
24
 
25
  # Start Qdrant and the application
26
- CMD ["python3 ingest.py", "uvicorn app:app --host 0.0.0.0 --port 7860"]
 
1
  FROM python:3.11
2
  FROM docker
3
 
4
+ RUN systemctl start docker
5
+ RUN docker pull kushagrash13/patent-image:1.0
6
+ RUN docker run -p 6333:6333 kushagrash13/patent-image:1.0
7
 
8
  # Set working directory
9
  WORKDIR /code
 
24
  WORKDIR $HOME/app
25
 
26
  # Start Qdrant and the application
27
+ CMD ["uvicorn app:app --host 0.0.0.0 --port 7860"]
__pycache__/app.cpython-311.pyc CHANGED
Binary files a/__pycache__/app.cpython-311.pyc and b/__pycache__/app.cpython-311.pyc differ
 
app.py CHANGED
@@ -1,16 +1,16 @@
 
 
 
1
  from langchain import PromptTemplate
 
2
  from langchain.chains import RetrievalQA
3
- from langchain.embeddings import SentenceTransformerEmbeddings
4
- from fastapi import FastAPI, Request, Form, Response
5
  from fastapi.responses import HTMLResponse
6
- from fastapi.templating import Jinja2Templates
7
  from fastapi.staticfiles import StaticFiles
8
  from fastapi.encoders import jsonable_encoder
9
- from qdrant_client import QdrantClient
10
- from langchain.vectorstores import Qdrant
11
- import os
12
- import json
13
- from langchain_groq import ChatGroq
14
 
15
  os.environ["TRANSFORMERS_FORCE_CPU"] = "true"
16
 
@@ -36,7 +36,6 @@ llm = ChatGroq(
36
  api_key=api_key,
37
  )
38
 
39
-
40
  print("LLM Initialized....")
41
 
42
  prompt_template = """Use the following pieces of information to answer the user's question.
 
1
+ import os
2
+ import json
3
+ from langchain_groq import ChatGroq
4
  from langchain import PromptTemplate
5
+ from qdrant_client import QdrantClient
6
  from langchain.chains import RetrievalQA
7
+ from langchain.vectorstores import Qdrant
 
8
  from fastapi.responses import HTMLResponse
 
9
  from fastapi.staticfiles import StaticFiles
10
  from fastapi.encoders import jsonable_encoder
11
+ from fastapi.templating import Jinja2Templates
12
+ from fastapi import FastAPI, Request, Form, Response
13
+ from langchain.embeddings import SentenceTransformerEmbeddings
 
 
14
 
15
  os.environ["TRANSFORMERS_FORCE_CPU"] = "true"
16
 
 
36
  api_key=api_key,
37
  )
38
 
 
39
  print("LLM Initialized....")
40
 
41
  prompt_template = """Use the following pieces of information to answer the user's question.
ingest.py CHANGED
@@ -1,11 +1,9 @@
1
  from langchain.text_splitter import RecursiveCharacterTextSplitter
2
- from langchain.document_loaders import DirectoryLoader
3
- from langchain.document_loaders import PyPDFLoader
4
- from langchain.vectorstores import Qdrant
5
- from langchain.embeddings import SentenceTransformerEmbeddings
6
 
7
- # embeddings = SentenceTransformerEmbeddings(model_name='NeuML/pubmedbert-base-embeddings')
8
- # embeddings = SentenceTransformerEmbeddings(model_name='mixedbread-ai/mxbai-embed-large-v1')
9
  embeddings = SentenceTransformerEmbeddings(model_name='BAAI/bge-large-en')
10
 
11
  print(embeddings)
 
1
  from langchain.text_splitter import RecursiveCharacterTextSplitter
2
+ from langchain_community.document_loaders import DirectoryLoader
3
+ from langchain_community.document_loaders import PyPDFLoader
4
+ from langchain_community.vectorstores import Qdrant
5
+ from langchain_community.embeddings import SentenceTransformerEmbeddings
6
 
 
 
7
  embeddings = SentenceTransformerEmbeddings(model_name='BAAI/bge-large-en')
8
 
9
  print(embeddings)