Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -3,7 +3,7 @@ from langchain_community.vectorstores import FAISS
|
|
3 |
from langchain_community.document_loaders import PyPDFLoader
|
4 |
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
5 |
from langchain_community.embeddings import HuggingFaceEmbeddings
|
6 |
-
from
|
7 |
from langchain.chains import ConversationalRetrievalChain
|
8 |
from langchain.memory import ConversationBufferMemory
|
9 |
import os
|
@@ -13,6 +13,7 @@ import base64
|
|
13 |
import tempfile
|
14 |
import io
|
15 |
from pathlib import Path
|
|
|
16 |
|
17 |
# Load environment variables
|
18 |
load_dotenv()
|
@@ -70,12 +71,17 @@ def initialize_llmchain(llm_model, temperature, max_tokens, top_k, vector_db):
|
|
70 |
"""Initialize the LLM chain with correct parameters"""
|
71 |
llm = HuggingFaceHub(
|
72 |
repo_id=llm_model,
|
|
|
|
|
|
|
|
|
|
|
73 |
model_kwargs={
|
74 |
"temperature": float(temperature),
|
75 |
"max_new_tokens": int(max_tokens),
|
76 |
-
"top_k": int(top_k)
|
77 |
-
|
78 |
-
|
79 |
)
|
80 |
|
81 |
memory = ConversationBufferMemory(
|
|
|
3 |
from langchain_community.document_loaders import PyPDFLoader
|
4 |
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
5 |
from langchain_community.embeddings import HuggingFaceEmbeddings
|
6 |
+
from langchain_community.llms import HuggingFaceHub
|
7 |
from langchain.chains import ConversationalRetrievalChain
|
8 |
from langchain.memory import ConversationBufferMemory
|
9 |
import os
|
|
|
13 |
import tempfile
|
14 |
import io
|
15 |
from pathlib import Path
|
16 |
+
from huggingface_hub import InferenceClient
|
17 |
|
18 |
# Load environment variables
|
19 |
load_dotenv()
|
|
|
71 |
"""Initialize the LLM chain with correct parameters"""
|
72 |
llm = HuggingFaceHub(
|
73 |
repo_id=llm_model,
|
74 |
+
task="text-generation",
|
75 |
+
client=InferenceClient(
|
76 |
+
model=llm_model,
|
77 |
+
token=api_token
|
78 |
+
),
|
79 |
model_kwargs={
|
80 |
"temperature": float(temperature),
|
81 |
"max_new_tokens": int(max_tokens),
|
82 |
+
"top_k": int(top_k),
|
83 |
+
"do_sample": True
|
84 |
+
}
|
85 |
)
|
86 |
|
87 |
memory = ConversationBufferMemory(
|