-
Notifications
You must be signed in to change notification settings - Fork 10.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
py : add Gemma conversion from HF models #5647
Conversation
ggerganov
commented
Feb 21, 2024
I successfully created a 2b GGUF and loaded model with the server in master. Thanks! |
Co-authored-by: Aarni Koskela <akx@iki.fi>
Co-authored-by: Aarni Koskela <akx@iki.fi>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Works well for me
I notice that the HF |
I noticed that as well, but I think the actual tensor shape in the |
I just checked. The original internal checkpoint uses [256128, 3072] for 7B. |
Co-authored-by: Jared Van Bortel <jared@nomic.ai>
|
* py : add gemma conversion from HF models * Update convert-hf-to-gguf.py Co-authored-by: Aarni Koskela <akx@iki.fi> * Update convert-hf-to-gguf.py Co-authored-by: Aarni Koskela <akx@iki.fi> * Update convert-hf-to-gguf.py Co-authored-by: Jared Van Bortel <jared@nomic.ai> --------- Co-authored-by: Aarni Koskela <akx@iki.fi> Co-authored-by: Jared Van Bortel <jared@nomic.ai>
* py : add gemma conversion from HF models * Update convert-hf-to-gguf.py Co-authored-by: Aarni Koskela <akx@iki.fi> * Update convert-hf-to-gguf.py Co-authored-by: Aarni Koskela <akx@iki.fi> * Update convert-hf-to-gguf.py Co-authored-by: Jared Van Bortel <jared@nomic.ai> --------- Co-authored-by: Aarni Koskela <akx@iki.fi> Co-authored-by: Jared Van Bortel <jared@nomic.ai>
I recently downloaded the Meta-Llama-3-8b model from huggingface and attempted to convert it to GGUF format using the following command:
However, I encountered the same error. The specific error message I received was:
I also tried modifying the function map_tensor_name in the script to change the suffixes from .weight to .weight_map, which resolved the previous error, but now I am getting a new error:
The reason I made this change is that the structure of model.safetensors.index.json indicates that the weights are mapped with suffixes like .weight_map. Here is an excerpt from the JSON file for reference:
Could anyone please provide guidance on how to properly map these tensors, or if there's a different approach needed for this conversion? Thank you! |