-
Notifications
You must be signed in to change notification settings - Fork 524
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Convert to Gguf format to work with Llama.cpp? #32
Comments
Hi, thank you for your suggestion. I will add compatibility with community tools to my to-do list. |
gguf format is good for ollama users, Any Update? |
It will be nice to have this model in gguf format in ollama. |
any updates on this? the 4b intern model is killer for its size! would love to see it supported with llama.cpp |
Would love internvl-chat-v1-5 in a gguf format! |
I second this |
@ErfeiCui why did you close this as completed? |
Any update on this? InternVL2-Llama3-76B on Ollama/llama.cpp would be amazing! |
If someone gives me a tutorial I will write my own code to tranform this for pytorch to gguf for llama.cpp myself |
It's more involved. you have to implement the model architecture and image preprocessing logic to llama.cpp which uses C++. |
Llava has various quantized models in gguf format, so it can be used with Llama.cpp.
ggerganov/llama.cpp#3436
Is this possible?
The text was updated successfully, but these errors were encountered: