-
Notifications
You must be signed in to change notification settings - Fork 11.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Quantize python script fails. #431
Comments
I'm wondering what's the value of
|
It was a list with two items, each item was the paths concatenated into the same string. I will post. ['models/llama/13B/ggml-model-f16.bin/models/llama/13B/ggml-model-f16.bin.1', 'models/llama/13B/ggml-model-f16.bin/models/llama/13B/ggml-model-f16.bin'] |
I can confirm this bug. Add to quantize.py, line 81: for v in f16_model_parts_paths:
print(v) Run python3 quantize.py --models-path models 7B Output:
|
Can you tru again with the version of the script in #428? That should fix the issue. |
Run as expected, great! |
Bump numpy from 1.24.3 to 1.24.4
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
I have my llama models stored in models/llama/{7B,13B,30B,65B}.
I expect that when I run the following command that the model will be converted
$ python3 quantize.py --models-path models/llama 30B
Current Behavior
When attempting to quantize the model by running
$ python3 quantize.py --models-path models/llama 30B
I get the following error:
The f16 model ggml-model-f16.bin was not found in models/llama/30B. If you want to use it from another location, set the --models-path argument from the command line.
modifying lines 76-79
To
Makes it work.
Failure Information (for bugs)
The text was updated successfully, but these errors were encountered: