-
Notifications
You must be signed in to change notification settings - Fork 11.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
4bit version of gpt4all-alpaca-oa-codealpaca-Lora-13b? #1037
Comments
llama.cpp can now load LoRA adapters, you need to convert the LoRA model to ggml using convert-lora-to-ggml.py, then load the original LLaMA 13b as the model and your LoRA model on top of it when launching |
Is there a way to work around this with existing llama.cpp options or would it require a PR? |
You're right |
Can anyone convert this to load this model? I'm particularly interested in this topic of using this models to write and work with code |
I created this script to merge the models: https://gist.github.com/SlyEcho/477554916bfc1a9e338240eee6396fbd It creates a HF checkpoint that can be converted using convert.py to ggml f16 format and then later to q4_0 with quantize. However, I'm not sure that the extra tokens are being used for tokenization. EDIT: It seems to work even with the text versions of |
So by converting the files with the ggml python script we can use gpt4all-alpaca-oa-codealpaca-Lora-13b but not as one file. But your script @SlyEcho can do that? Edit: |
The script should download 13b from huggingface.co/decapoda-research/llama-13b-hf automatically. I also tried the |
Thank you, |
|
Can anyone upoad the bin file of this model for using it on llama.cpp? |
I could but I there is no point because it doesn't work well. |
Tks anyway :) |
Where can i find this? I can only find the conversion scripts for gpt4all etc.? |
This one: convert.py edit: if you are seeing gpt4all conversion scripts, then you may need to do a |
Thank you, don't know how I didn't see that. |
I have gotten a vocab size mismatch, how can I fix that? |
You need to use the vocab files from the jordiclive/gpt4all-alpaca-oa-codealpaca-lora-13b repo. But there are some weird things going on, there are embeddings for 16 new tokens in there, but the JSON only specifies 5. My script also cuts it down to 5 but you may want to hack on this because I don't understand how it's supposed to work. |
I have forgotten to put the added_tokens.json in the directory. |
If you change // prefix & suffix for instruct mode
const auto inp_pfx = std::vector<llama_token> { 32002 }; // <|prompter|>
const auto inp_sfx = std::vector<llama_token> { 32004 }; // <|assistant|> edit: I think the |
The output I get is also a bit weird, it doesn't want to write code. It wanted me to visit a GitHub repo which doesn't exist. |
I can recommend other good models that are not LoRA:
These can be converted directly with |
I believe this has been answered! |
* OSX attempt 1 * OSX Pyinstaller * Update kcpp-build-release-osx.yaml * Update kcpp-build-release-osx.yaml * Update kcpp-build-release-osx.yaml * Add .metal file * Update kcpp-build-release-osx.yaml * Polish Mac
* OSX attempt 1 * OSX Pyinstaller * Update kcpp-build-release-osx.yaml * Update kcpp-build-release-osx.yaml * Update kcpp-build-release-osx.yaml * Add .metal file * Update kcpp-build-release-osx.yaml * Polish Mac (cherry picked from commit 52cc0da)
Hello,
to reduce my brain usage even more I thought i'd be nice to run AI which is specifically trained to code and thus hopefully make better code than other language models which are trained for e.g. natural language.
So I found this: https://huggingface.co/jordiclive/gpt4all-alpaca-oa-codealpaca-lora-13b
I of course wanted to try and run it but there's a problem, there aren't even any pytorch_model files or any 4bit variants listed here: https://github.com/underlines/awesome-marketing-datascience/blob/master/awesome-ai.md
Thank your for your support!
The text was updated successfully, but these errors were encountered: