Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature Request] Add batch processing for input prompt data in embedding mode. #893

Closed
AL-Kost opened this issue Apr 11, 2023 · 2 comments
Labels

Comments

@AL-Kost
Copy link

AL-Kost commented Apr 11, 2023

It would be nice to add batch processing for input prompt data in embedding mode.
I.e., read prompts from a file and output a map of prompts and their embeddings.
It seems that this can be done by modifying the logic of the -f flag for embedding mode.

@AL-Kost AL-Kost changed the title [Feature Request] Add the ability to work with batches of input prompt data in embedding mode. [Feature Request] Add batch processing for input prompt data in embedding mode. Apr 11, 2023
@rjadr
Copy link
Contributor

rjadr commented Apr 13, 2023

I second this. I've been working on integrating llama.cpp in langchain, but the retrieval of embeddings is terribly slow since we can only pass single strings (for which the model is loaded anew every time). Batch processing embeddings would be very helpful here, preferably by being able to pass a list of strings in the CLI.

@github-actions github-actions bot added the stale label Mar 25, 2024
Copy link
Contributor

This issue was closed because it has been inactive for 14 days since being marked as stale.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants