-
Notifications
You must be signed in to change notification settings - Fork 140
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llama.cpp: tensor 'layers.0.attention.wk.weight' has wrong shape; expected 8192 x 8192, got 8192 x 1024 llama_init_from_file: failed to load model #62
Comments
...actually fixing this is well beyond my skill level, but it is indeed related to the newness of the 70B model handling. Quantization changed: |
Did you use the |
with ctransformers? no... OK I did not think that it would be passed. I will try it now. |
No sorry, my mistake. You will need the newest master from llama.cpp and it needs the |
...and no, cannot set up the LLM instance, invalid argument. I have llama.cpp working, but that does not help as I need python bindings. I am waiting for llama-cpp-python to update, otherwise I can build the required .so by pulling the working llama.cpp and building from source. Cannot do the same with ctransformers as it is written in c++ and I do not speak that language |
Added support for LLaMA 2 70B models in the latest version 0.2.15 Since @TheBloke models should work out of the box without any additional configuration: llm = AutoModelForCausalLM.from_pretrained("TheBloke/Llama-2-70B-GGML") |
Good to see ctransformers support. Looking for '70b' wouldn't have worked with https://huggingface.co/TheBloke/StableBeluga2-GGML so I have renamed it StableBeluga2-70B-GGML |
The ggml loader relies on this horrible hack:
There seem to be no way to set Any solution for GGML other than waiting for GGUF support? Workaround: Add |
GGUF support is added in 0.2.24
|
This error occurs with quantized 70B model that works with the latest current master branch of llama.cpp
I am guessing that you would just need to update the PyPi package. Will try to build from source in the meantime.
The text was updated successfully, but these errors were encountered: