-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add minicpm-o and qwen2-vl to the list of supported multimodal models. #1904
Comments
+1 |
hmm, just tested again. maybe was me or i did pull an outdated llama or what last time. minicpm-o seems to work with the "minicpm-v-2.6" chat handler. |
Yes, minicpm-o-2.6 works with the minicpm-v-2.6 chat handler. But Qwen2-VL seems does not work with any existing chat handler. I try to use the example chat template from llama.cpp but it still generate random characters... |
This is interesting. Could you give us the GGUF model urls you are using? |
@samkoesnadi I downloaded them from HuggingFace. Hope you have some good news. |
@samkoesnadi i tried my luck with Qwen2-VL-7B-Instruct-GGUF and tried almost every registered chat handler that includes a <|im_start|> and <|im_end|> token in the template and got the same results as @la1ty with random words in random languages as reply. i also tried to implment the chat template myself but unfortunately did fail since i didnt realy understand the jinja template:
the template expects <|vision_start|><|image_pad|><|vision_end|> (which is verry unique for this model and not registered in any chat handler so far) and i didnt realy see where the base64 encoded string / image_url should go in this template to be true. |
@la1ty could you guys try the 2B and see if it works? That's the one I tested... |
@samkoesnadi which chat handler did you use if i may ask? the exact url to the model you used there would be usefull aswell. |
@kseyhan Yes, that's what I exactly experienced. And I don't know if I make errors in compiling, but I found that text responses generating by Qwen2-VL-7b with llama-cpp-python v0.3.7 are mostly nonsense, which is not identical to the behavior in @samkoesnadi Yes it works with |
Support for the Qwen2-VL and MiniCPM-o models would be nice. They already have have been merged into the llava subproject of llama.cpp.
The text was updated successfully, but these errors were encountered: