-
Notifications
You must be signed in to change notification settings - Fork 11.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Investigate supporting starcode #1326
Comments
It looks like gpt-2. I'd imagine it's like codegen where you can use the normal gpt-2/gpt-j conversion/quantization scripts/main binary. You just need enough memory to convert/run it.
|
If it is like Saleforce's CodeGen model, checkout this repo: https://github.com/ravenscroftj/turbopilot |
It’s an independent model. BTW it uses tokenizer from gpt-2. |
I'm very much interested in this too. |
Working on it! Should be coming soon |
PR to support StarCoder/SantaCoder is ready! And results looking fine 🔥 |
I have seen it is supported in ggml-org/ggml#146 but how I can this be used in llama.cpp? |
It cannot be used in llama.cpp as it's not the same model architecture. but you can find a similar repo to llama.cpp for starcoder in https://github.com/bigcode-project/starcoder.cpp @christianwengert |
Bigcode just released starcoder. This is a 15B model trained on 1T Github tokens. This seems like it could be an amazing replacement for gpt-3.5 and maybe gpt-4 for local coding assistance and IDE tooling!
More info: https://huggingface.co/bigcode
The text was updated successfully, but these errors were encountered: