-
Notifications
You must be signed in to change notification settings - Fork 11.4k
How to use .safetensors model ? #688
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Comments
My conversion script (#545) will support this soon. |
I thought I'd give it a spin on some safetensors models:
If I didn't think I'd probably cause even more trouble with clumsy efforts, I'd have a stab at fixing it. |
I’ll take a look. |
try the new |
Deadsg
pushed a commit
to Deadsg/llama.cpp
that referenced
this issue
Dec 19, 2023
ggml-org#688) * Examples from ggml to gguf * Use gguf file extension Update examples to use filenames with gguf extension (e.g. llama-model.gguf). --------- Co-authored-by: Andrei <abetlen@gmail.com>
# for free
to join this conversation on GitHub.
Already have an account?
# to comment
I downloaded a model
alpaca-30b-lora-int4
from https://huggingface.co/elinas/alpaca-30b-lora-int4/tree/mainThe model is a
.safetensors
in GPTQ format I thinkI need to convert it to
GGML .bin
so I used the script provided inllama.cpp
with the commandpython convert-gptq-to-ggml.py models/30B/alpaca-30b-4bit.safetensors models/30B//tokenizer.model models/30B/alpaca-30b-4bit.bin
But I get the following error
How to use
.safetensors
models withllama.cpp
?The text was updated successfully, but these errors were encountered: