-
Notifications
You must be signed in to change notification settings - Fork 254
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Local Language Model Implementation Options: Vicuna and Open-Source Alternatives #75
Comments
I recently set it up to run on my local 7B Vicuna model via a custom Langchain classes for VicunaLLM and VicunaChatModel but the issue is that the responses tend to not return in proper JSON format and the system breaks down. So just letting you know that Vicuna is probably not really good enough to work out of the box without some additional tweaking. I could potentially look into LangChain's ReLLM integration to try and get around this problem though. I'm also looking to tweak my setup to get the 13B Vicuna running with CPU assistance. Just figured I'd share my experience with Vicuna here since I saw your comment. I'm also brand new to this stuff so I could just have messed something up too. |
Well, Thank you. I would really like to see if you get it run and if you can later share the code with maybe video or tutorial it will be great. Thank you! |
Generally I'd say the formatting guidance in the prompts is not strong enough, since some of the prompts get too long. I reckon it is possible to get a version running with smaller models like Vicuna 7B (albeit very slowly), if some of the prompts are reduced in length, so this would be a great place to start if you are looking to go that route. |
Thanks for the information. Can you tell how you did it? It's asking for API from open ai. I have little experience with python. I overall use oobabooga mostly. |
It would be fantastic to have the possibility of running it on local Language Model implementations such as Vicuna or other open-source alternatives.
The text was updated successfully, but these errors were encountered: