-
Notifications
You must be signed in to change notification settings - Fork 11.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
error loading model: missing tok_embeddings.weight #1381
Comments
no compatible ggml model |
I followed this : quantize the model to 4-bits (using q4_0 method)./quantize ./models/7B/ggml-model-f16.bin ./models/7B/ggml-model-q4_0.bin q4_0 run the inference./main -m ./models/7B/ggml-model-q4_0.bin -n 128 can not run it succesfully |
I am getting such error when I try to load not compatible model like neon x instead ggjt . |
the same problem, how to fix it? 12:05AM DBG Loading model 'starchat-beta.ggmlv3.q4_0.bin' greedly |
Same problem, anyone can help? ./main -m ./models/7B/ggml-model-q4_0.bin -n 128 |
I see at least 2 different models, probably corresponding to different branches in @realcarlos : |
i have the same problem, anyone can help?and What is the solution to this one? main: build = 992 (0919a0f) |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
$ ./examples/chat-gpt2.sh
main: build = 480 (f4cef87)
main: seed = 1683650863
llama.cpp: loading model from ./models/ggml-model-gpt2-q4_0.bin
error loading model: missing tok_embeddings.weight
llama_init_from_file: failed to load model
main: error: failed to load model './models/ggml-model-gpt2-q4_0.bin'
The text was updated successfully, but these errors were encountered: