Pamudu13 commited on
Commit
0195c50
·
verified ·
1 Parent(s): dc88c5a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -4
app.py CHANGED
@@ -3,7 +3,7 @@ from langchain_community.vectorstores import FAISS
3
  from langchain_community.document_loaders import PyPDFLoader
4
  from langchain.text_splitter import RecursiveCharacterTextSplitter
5
  from langchain_community.embeddings import HuggingFaceEmbeddings
6
- from langchain.llms import HuggingFaceHub
7
  from langchain.chains import ConversationalRetrievalChain
8
  from langchain.memory import ConversationBufferMemory
9
  import os
@@ -13,6 +13,7 @@ import base64
13
  import tempfile
14
  import io
15
  from pathlib import Path
 
16
 
17
  # Load environment variables
18
  load_dotenv()
@@ -70,12 +71,17 @@ def initialize_llmchain(llm_model, temperature, max_tokens, top_k, vector_db):
70
  """Initialize the LLM chain with correct parameters"""
71
  llm = HuggingFaceHub(
72
  repo_id=llm_model,
 
 
 
 
 
73
  model_kwargs={
74
  "temperature": float(temperature),
75
  "max_new_tokens": int(max_tokens),
76
- "top_k": int(top_k)
77
- },
78
- huggingfacehub_api_token=api_token
79
  )
80
 
81
  memory = ConversationBufferMemory(
 
3
  from langchain_community.document_loaders import PyPDFLoader
4
  from langchain.text_splitter import RecursiveCharacterTextSplitter
5
  from langchain_community.embeddings import HuggingFaceEmbeddings
6
+ from langchain_community.llms import HuggingFaceHub
7
  from langchain.chains import ConversationalRetrievalChain
8
  from langchain.memory import ConversationBufferMemory
9
  import os
 
13
  import tempfile
14
  import io
15
  from pathlib import Path
16
+ from huggingface_hub import InferenceClient
17
 
18
  # Load environment variables
19
  load_dotenv()
 
71
  """Initialize the LLM chain with correct parameters"""
72
  llm = HuggingFaceHub(
73
  repo_id=llm_model,
74
+ task="text-generation",
75
+ client=InferenceClient(
76
+ model=llm_model,
77
+ token=api_token
78
+ ),
79
  model_kwargs={
80
  "temperature": float(temperature),
81
  "max_new_tokens": int(max_tokens),
82
+ "top_k": int(top_k),
83
+ "do_sample": True
84
+ }
85
  )
86
 
87
  memory = ConversationBufferMemory(