-
Notifications
You must be signed in to change notification settings - Fork 11.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to use .safetensors model ? #688
Comments
My conversion script (#545) will support this soon. |
I thought I'd give it a spin on some safetensors models:
If I didn't think I'd probably cause even more trouble with clumsy efforts, I'd have a stab at fixing it. |
I’ll take a look. |
try the new |
Deadsg
pushed a commit
to Deadsg/llama.cpp
that referenced
this issue
Dec 19, 2023
ggml-org#688) * Examples from ggml to gguf * Use gguf file extension Update examples to use filenames with gguf extension (e.g. llama-model.gguf). --------- Co-authored-by: Andrei <abetlen@gmail.com>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I downloaded a model
alpaca-30b-lora-int4
from https://huggingface.co/elinas/alpaca-30b-lora-int4/tree/mainThe model is a
.safetensors
in GPTQ format I thinkI need to convert it to
GGML .bin
so I used the script provided inllama.cpp
with the commandpython convert-gptq-to-ggml.py models/30B/alpaca-30b-4bit.safetensors models/30B//tokenizer.model models/30B/alpaca-30b-4bit.bin
But I get the following error
How to use
.safetensors
models withllama.cpp
?The text was updated successfully, but these errors were encountered: