Converting the model into gguf extenstion #9625
spratap123
started this conversation in
General
Replies: 0 comments
# for free
to join this conversation on GitHub.
Already have an account?
# to comment
-
Hello Can Anyone help me in this, I am getting on error.
I am doing some work in colab and used llama 3.1 with 8 B paratameter model and perform quantization in 4bit after this I save that model. Now I want to covert this in gguf format so that I downloaded the llama_cpp repo and I am using one script for this which is already defined in this repo then I am getting error. Please anyone can help.
Traceback (most recent call last):
File "/content/llama.cpp/convert_hf_to_gguf.py", line 4309, in
main()
File "/content/llama.cpp/convert_hf_to_gguf.py", line 4303, in main
model_instance.write()
File "/content/llama.cpp/convert_hf_to_gguf.py", line 425, in write
self.prepare_tensors()
File "/content/llama.cpp/convert_hf_to_gguf.py", line 1618, in prepare_tensors
super().prepare_tensors()
File "/content/llama.cpp/convert_hf_to_gguf.py", line 294, in prepare_tensors
for new_name, data in ((n, d.squeeze().numpy()) for n, d in self.modify_tensors(data_torch, name, bid)):
File "/content/llama.cpp/convert_hf_to_gguf.py", line 1586, in modify_tensors
return [(self.map_tensor_name(name), data_torch)]
File "/content/llama.cpp/convert_hf_to_gguf.py", line 214, in map_tensor_name
raise ValueError(f"Can not map tensor {name!r}")
ValueError: Can not map tensor 'model.layers.0.mlp.down_proj.weight.absmax'
I am using this srcipt --> convert_hf_to_gguf.py
Beta Was this translation helpful? Give feedback.
All reactions