Spaces:
Runtime error
Runtime error
# Using Ollama Embeddings with GraphRAG: A Quick Guide | |
## Problem | |
GraphRAG is designed to work with OpenAI-compatible APIs for both language models and embeddings and Ollama currently has their own way of doing embeddings. | |
## Solution: Embeddings Proxy | |
To bridge this gap, let's use an embeddings proxy. This proxy acts as a middleware between GraphRAG and Ollama, translating Ollama's embedding responses into a format that GraphRAG expects. | |
## Use the Embeddings Proxy | |
1. **Set up the proxy:** | |
- Save the provided `embedding_proxy.py` script to your project directory. | |
- Install required dependencies (not needed if you've already done this in the normal setup): `pip install fastapi uvicorn httpx` | |
2. **Run the proxy:** | |
```bash | |
python embedding_proxy.py --port 11435 --host http://localhost:11434 | |
``` | |
This starts the proxy on port 11435, connecting to Ollama at localhost:11434. | |
3. **Configure GraphRAG:** | |
Update your `settings.yaml` file to use the proxy for embeddings: | |
```yaml | |
embeddings: | |
llm: | |
api_key: ${GRAPHRAG_API_KEY} | |
type: openai_embedding | |
model: nomic-embed-text:latest | |
api_base: http://localhost:11435 # Point to your proxy | |
``` | |
4. **Run GraphRAG:** | |
With the proxy running and the configuration updated, you can now run GraphRAG as usual. It will use Ollama for embeddings through the proxy. | |