You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It would be nice to add batch processing for input prompt data in embedding mode.
I.e., read prompts from a file and output a map of prompts and their embeddings.
It seems that this can be done by modifying the logic of the -f flag for embedding mode.
The text was updated successfully, but these errors were encountered:
AL-Kost
changed the title
[Feature Request] Add the ability to work with batches of input prompt data in embedding mode.
[Feature Request] Add batch processing for input prompt data in embedding mode.
Apr 11, 2023
I second this. I've been working on integrating llama.cpp in langchain, but the retrieval of embeddings is terribly slow since we can only pass single strings (for which the model is loaded anew every time). Batch processing embeddings would be very helpful here, preferably by being able to pass a list of strings in the CLI.
It would be nice to add batch processing for input prompt data in embedding mode.
I.e., read prompts from a file and output a map of prompts and their embeddings.
It seems that this can be done by modifying the logic of the
-f
flag for embedding mode.The text was updated successfully, but these errors were encountered: