Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Cannot load 2 bit quantized ggml model on Windows #1018

Closed
dillfrescott opened this issue Apr 17, 2023 · 1 comment
Closed

Cannot load 2 bit quantized ggml model on Windows #1018

dillfrescott opened this issue Apr 17, 2023 · 1 comment

Comments

@dillfrescott
Copy link

C:\Users\micro\Downloads>main -m ggml-model-q2_0.bin
main: seed = 1681700481
llama.cpp: loading model from ggml-model-q2_0.bin
error loading model: unrecognized tensor type 5

llama_init_from_file: failed to load model
main: error: failed to load model 'ggml-model-q2_0.bin'
@rabidcopy
Copy link
Contributor

#1004 Not merged yet and still a WIP. You will need to use this PR to load models quantized to 2-bit.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants