-
Notifications
You must be signed in to change notification settings - Fork 7.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature] Add support for StarCoder2 model architecture #2100
Comments
"Besides: I can't select or copy any of those strings in the UI or error messages which makes writing those issues much more tedious as it should be." As an aside, this is a good point. We should make it easy to copy error messages or other info messages from the UI for ease of use in bug reports and issues and the like. Thanks for the tip! |
In order not to mix up issues and capture this bug report, I created #2101 for this. |
I do not know of any way to identify for certain the llama.cpp model architecture of a GGUF on HuggingFace without downloading it. I don't think GPT4All can do anything about this, people would have to standardize in some metadata on HF in the README or a JSON file in the repo. We could possibly do HTTP range requests to partially download the GGUF, but I think in practice that would be a fragile hack. StarCoder and StarCoder2 are not models you can chat with or give instructions to - they are solely code completion models, which the chat UI is not designed for. What is your use case? |
I had no idea. I'm somebody with no LLM knowledge at all. I'm just curiously testing AI potential for my personal workflows. I assumed that I could use StarCoder2 to ask in the GT4all interface things like "write me a Python function that ..." or "find the bug in ..." instead of mistral which doesn't result in very reliable code so to say. It's unfortunate that you'd have to know deeper information so that you don't have to download many GBs just to find out that this isn't working with GPT4All. Maybe a group of people can maintain a whitelist and watch newer "models" to deal with that issue with an organizational process? |
same problem here I wish to use: no idea how. |
This is not the same issue - this is an original StarCoder model, not StarCoder 2. As described in this thread, StarCoder is not a chat model that can usefully be used with the Chat UI. But you could open a new issue to explain your situation if you'd like. |
Bug Report
Steps to Reproduce
Besides: I can't select or copy any of those strings in the UI or error messages which makes writing those issues much more tedious as it should be.
The model download was complete and had the same md5/sha1 sum as an independent download within my shell.
Models like mistral-7b-openorca.Q4_0.gguf (downloaded weeks or even months ago) or wizardcoder-python-13b-v1.0.Q4_0.gguf (downloaded yesterday) do work.
Expected Behavior
I would expect that a completely and error-free download of a model resulted in a working chat when selected from the dropdown of models.
If the model is somehow not compatible with GPT4All, it should not be available in the list of potential downloads.
Environment
The text was updated successfully, but these errors were encountered: