Converting the model into gguf extenstion #9625
spratap123
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello Can Anyone help me in this, I am getting on error.
I am doing some work in colab and used llama 3.1 with 8 B paratameter model and perform quantization in 4bit after this I save that model. Now I want to covert this in gguf format so that I downloaded the llama_cpp repo and I am using one script for this which is already defined in this repo then I am getting error. Please anyone can help.
Traceback (most recent call last):
File "/content/llama.cpp/convert_hf_to_gguf.py", line 4309, in
main()
File "/content/llama.cpp/convert_hf_to_gguf.py", line 4303, in main
model_instance.write()
File "/content/llama.cpp/convert_hf_to_gguf.py", line 425, in write
self.prepare_tensors()
File "/content/llama.cpp/convert_hf_to_gguf.py", line 1618, in prepare_tensors
super().prepare_tensors()
File "/content/llama.cpp/convert_hf_to_gguf.py", line 294, in prepare_tensors
for new_name, data in ((n, d.squeeze().numpy()) for n, d in self.modify_tensors(data_torch, name, bid)):
File "/content/llama.cpp/convert_hf_to_gguf.py", line 1586, in modify_tensors
return [(self.map_tensor_name(name), data_torch)]
File "/content/llama.cpp/convert_hf_to_gguf.py", line 214, in map_tensor_name
raise ValueError(f"Can not map tensor {name!r}")
ValueError: Can not map tensor 'model.layers.0.mlp.down_proj.weight.absmax'
I am using this srcipt --> convert_hf_to_gguf.py
Beta Was this translation helpful? Give feedback.
All reactions