Spaces:
Sleeping
Sleeping
hf llm
Browse files- backend.py +2 -1
- requirements.txt +1 -1
backend.py
CHANGED
@@ -14,7 +14,8 @@ from typing import Iterator, List, Any
|
|
14 |
from llama_index.core.chat_engine import CondensePlusContextChatEngine
|
15 |
from llama_index.core.llms import ChatMessage, MessageRole , CompletionResponse
|
16 |
from IPython.display import Markdown, display
|
17 |
-
from
|
|
|
18 |
#from llama_index import LangchainEmbedding, ServiceContext
|
19 |
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
20 |
from llama_index.llms.huggingface import HuggingFaceInferenceAPI, HuggingFaceLLM
|
|
|
14 |
from llama_index.core.chat_engine import CondensePlusContextChatEngine
|
15 |
from llama_index.core.llms import ChatMessage, MessageRole , CompletionResponse
|
16 |
from IPython.display import Markdown, display
|
17 |
+
from langchain_huggingface import HuggingFaceEmbeddings
|
18 |
+
|
19 |
#from llama_index import LangchainEmbedding, ServiceContext
|
20 |
from llama_index.embeddings.huggingface import HuggingFaceEmbedding
|
21 |
from llama_index.llms.huggingface import HuggingFaceInferenceAPI, HuggingFaceLLM
|
requirements.txt
CHANGED
@@ -10,7 +10,7 @@ llama-index-readers-web
|
|
10 |
llama-index-readers-file
|
11 |
llama-index-embeddings-langchain
|
12 |
pypdf
|
13 |
-
langchain
|
14 |
langchain-community
|
15 |
#langchain-huggingface
|
16 |
bitsandbytes
|
|
|
10 |
llama-index-readers-file
|
11 |
llama-index-embeddings-langchain
|
12 |
pypdf
|
13 |
+
langchain>=0.2.3
|
14 |
langchain-community
|
15 |
#langchain-huggingface
|
16 |
bitsandbytes
|