GraphRAG is designed to work with OpenAI-compatible APIs for both language models and embeddings and Ollama currently has their own way of doing embeddings.
To bridge this gap, let's use an embeddings proxy. This proxy acts as a middleware between GraphRAG and Ollama, translating Ollama's embedding responses into a format that GraphRAG expects.
-
Set up the proxy:
- Save the provided
embedding_proxy.py
script to your project directory. - Install required dependencies (not needed if you've already done this in the normal setup):
pip install fastapi uvicorn httpx
- Save the provided
-
Run the proxy:
python embedding_proxy.py --port 11435 --host http://localhost:11434
This starts the proxy on port 11435, connecting to Ollama at localhost:11434.
-
Configure GraphRAG: Update your
settings.yaml
file to use the proxy for embeddings:embeddings: llm: api_key: ${GRAPHRAG_API_KEY} type: openai_embedding model: nomic-embed-text:latest api_base: http://localhost:11435 # Point to your proxy
-
Run GraphRAG: With the proxy running and the configuration updated, you can now run GraphRAG as usual. It will use Ollama for embeddings through the proxy.