Skip to content

Question: How do I search asynchronously using llm response streaming? #283

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
bldev2473 opened this issue Feb 22, 2025 · 2 comments
Open

Comments

@bldev2473
Copy link

bldev2473 commented Feb 22, 2025

Hi there.

I'm trying to export the final response of a RAG search to SSE using content streaming from an LLM provider.

How can I integrate the generator response of generate_content_stream() func?
The search() function of GraphRAG is synchronous and the invoke() call of the llm model is synchronous.

Image

Do I need to customise GraphRAG?

@stellasia
Copy link
Contributor

Hi @bldev2473 ,

I'm afraid streaming response is not yet possible with this package. The LLMInterface which is responsible for calling the LLM is not managing streaming at the moment.

If this is important to you and you feel like you want to contribute this feature, we'd be happy to support you in the implementation.

@bldev2473
Copy link
Author

Thank you for your response. I'll use the custom implementation for now and ask for more later.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants