-
Notifications
You must be signed in to change notification settings - Fork 11.4k
Quantize python script fails. #431
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Comments
I'm wondering what's the value of
|
It was a list with two items, each item was the paths concatenated into the same string. I will post. ['models/llama/13B/ggml-model-f16.bin/models/llama/13B/ggml-model-f16.bin.1', 'models/llama/13B/ggml-model-f16.bin/models/llama/13B/ggml-model-f16.bin'] |
I can confirm this bug. Add to quantize.py, line 81: for v in f16_model_parts_paths:
print(v) Run python3 quantize.py --models-path models 7B Output:
|
Can you tru again with the version of the script in #428? That should fix the issue. |
Run as expected, great! |
Bump numpy from 1.24.3 to 1.24.4
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
I have my llama models stored in models/llama/{7B,13B,30B,65B}.
I expect that when I run the following command that the model will be converted
$ python3 quantize.py --models-path models/llama 30B
Current Behavior
When attempting to quantize the model by running
$ python3 quantize.py --models-path models/llama 30B
I get the following error:
The f16 model ggml-model-f16.bin was not found in models/llama/30B. If you want to use it from another location, set the --models-path argument from the command line.
modifying lines 76-79
To
Makes it work.
Failure Information (for bugs)
The text was updated successfully, but these errors were encountered: