You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I came across a model on Huggingface that supports Llama3 multimodal Bunny-Llama-3-8B-V: bunny-llama, and I'd like to be able to deploy it using llama-cpp-python!
But I found that the existing chat_format:llama-3 doesn't seem to support running it.
I converted it to gguf format via llama.cpp and ran it with the following configuration
Check out #1147 it should be merged soon. The only caveat here is that you'll need use the llava example in llama.cpp to extract the image encoder as well when you quantize the models.
I came across a model on Huggingface that supports Llama3 multimodal Bunny-Llama-3-8B-V: bunny-llama, and I'd like to be able to deploy it using llama-cpp-python!
But I found that the existing chat_format:llama-3 doesn't seem to support running it.
I converted it to gguf format via llama.cpp and ran it with the following configuration
The text was updated successfully, but these errors were encountered: