-
Notifications
You must be signed in to change notification settings - Fork 10.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Misc. bug: Unable to load model #10696
Comments
Could it be related to #10208? |
No. The problem here is trying to run the Metal backend on Intel device which is not supported. |
Thank you @ggerganov. My setup is:
Is this supported? As I'm experiencing the same problem. |
It is supported. Show the commands and the full log that you get. |
Command:
Error:
However, it works fine when I re-build everything without Metal using |
@ggerganov and @eugeniosegala Thanks for the support! Re-building everything without the metal backend worked for me. I'm closing this issue as it is not a bug. |
@VJHack the fact that the Metal build does not work is okay though? Is it a working in progress feature? |
Provide more logs: |
@ggerganov this is what I got back: I got the same with other models too, like llama 3 8b. |
This should fix it: rm -rf build
cmake -B build
cmake --build build --config Release
./build/bin/llama-cli -m ./FastLlama-3.2-1B-Instruct.i1-IQ3_XS.gguf -p "I believe the meaning of life is" |
@ggerganov I get the same error. I have tried to re-build it several times. P.S: I'm building from latest master I will try to build it from some previous commits as well. |
Clone the repo from scratch. You likely have some local modifications. |
@ggerganov you are correct, I had to re-clone everything and re-build again, then it worked! I did not have local changes or anything but maybe something was cached or similar (I did delete the build folder though). Thank you! |
Name and Version
./build/bin/llama-cli --version
register_backend: registered backend Metal (1 devices)
register_device: registered device Metal (Intel(R) Iris(TM) Plus Graphics 655)
register_backend: registered backend BLAS (1 devices)
register_device: registered device BLAS (Accelerate)
register_backend: registered backend CPU (1 devices)
register_device: registered device CPU (Intel(R) Core(TM) i5-8279U CPU @ 2.40GHz)
version: 4233 (f68367d2)
built with Apple clang version 14.0.0 (clang-1400.0.29.202) for x86_64-apple-darwin21.6.0
Operating systems
Mac
Which llama.cpp modules do you know to be affected?
Other (Please specify in the next section)
Problem description & steps to reproduce
I am unable to load and run inference on any model. Previously with the make build system, I didn't have any issues. But now that our build system has switched to cmake I've been having this issue shown below. I'm not able to use either llama-cli or llama-server.
I'm running on MacOS with 16 GB of memory and Intel Core i5 processor.
How I'm building:
cmake -B build
cmake --build build --config Release
command:
./build/bin/llama-server -m models/tinyllama-1.1b-chat-v1.0.Q2_K.gguf
Any help would be appreciated!
First Bad Commit
No response
Relevant log output
The text was updated successfully, but these errors were encountered: