-
Notifications
You must be signed in to change notification settings - Fork 134
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running llama-cpp-python OpenAI compatible server #140
Comments
Not sure if it's the same, but I'm using koboldcpp - perhaps try using |
Hi, it's indeed the |
This issue is stale because it has been open for 30 days with no activity. |
I am going to give this a try tomorrow and report back. And sorry I didn't get back sooner. I just saw a note from github notification. Thank you for the reply. I was using |
They might yes, the extension doesn't support chat models atm. The model you use must be compatible with code completion, either with fill in the middle or not (but I strongly advise to use FIM as it generates more relevant completions). |
This issue is stale because it has been open for 30 days with no activity. |
bump |
This issue is stale because it has been open for 30 days with no activity. |
Requesting a little help here. Trying to test out copilot functionality with
llama-cpp-python
with this extension. Below is my configuration setting.I am seeing there is inference going on the server:
So I am not entirely sure what I am missing. Additionally I am trying to see the extension logs.. for the worker calls. But I don't see anything. Would you be able to give any guidance or some step by step explanation on how this can be done.
Thank you so much
The text was updated successfully, but these errors were encountered: