Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for InternVL #6803

Closed
chigkim opened this issue Apr 21, 2024 · 33 comments
Closed

Support for InternVL #6803

chigkim opened this issue Apr 21, 2024 · 33 comments
Labels
enhancement New feature or request stale

Comments

@chigkim
Copy link

chigkim commented Apr 21, 2024

New InternVL-Chat-V1.5 just came out, and the quality is really great, and the benchmark score is pretty high too. Possibly best open source vision language model yet?

Can we have llama.cpp to support it? @cmp-nct, @cjpais, @danbev, @monatis, has any of you tried it?

Demo: https://internvl.opengvlab.com/

@chigkim chigkim added the enhancement New feature or request label Apr 21, 2024
@paryska99
Copy link

Would be great

@cjpais
Copy link
Contributor

cjpais commented Apr 23, 2024

I am working on a few projects right now, but if I get a chance I will try to get support in (assuming it doesn't already work). I would also like to get moondream support in

@2132660698
Copy link

+1

@cjpais
Copy link
Contributor

cjpais commented Apr 27, 2024

fwiw moondream support was merged in #6899, haven't had a chance to look at/try internvl

@sapere-aude-incipe
Copy link

I would really like to get InternVL support in llama.cpp.

I have tested the demo extensively and it is really good, so much so that I feel like it is a game changer in many ways. But running it on consumer hardware is not possible right now.

As noted here: InternLM/lmdeploy#1501 (comment)

architecture: InternViT-6B-448px-V1-5 + MLP + InternLM2-Chat-20B
I am afraid it cannot fit into A10 (24G) even though LLM weights are quantized into 4 bits.

Is it possible to GGUF the weights to allow for multi GPU splitting or splitting layers between CPU RAM and VRAM? Adding support for InternVL 1.5 would also (probably) make it easier to support future versions when they come out.

@Single430
Copy link

@cjpais Hello, may I ask what is the progress of internvl support now? We are looking forward to using it on llama.cpp.

@cjpais
Copy link
Contributor

cjpais commented May 22, 2024

Hey I am quite busy with a few projects, it's on my list but just not very high priority at the moment. It's really only something I can do in my spare/free time

@Single430
Copy link

Hey I am quite busy with a few projects, it's on my list but just not very high priority at the moment. It's really only something I can do in my spare/free time

Thank you for your reply. Thank you for your hard work. Looking forward to your future work.

@chigkim
Copy link
Author

chigkim commented Jun 5, 2024

Which one would be better to focus: CogVLM or InternVL?

I wish there is more resource/interest for language vision models among the llama.cpp community. Llama.cpp is the only hope to run newer language vision models on Apple Silicon. Especially since flash attention python library is not available for Apple Sillicon, you can't even run inference using Torch with MPS support. :(

@opisaac9001
Copy link

Which one would be better to focus: CogVLM or InternVL?

I wish there is more resource/interest for language vision models among the llama.cpp community. Llama.cpp is the only hope to run newer language vision models on Apple Silicon. Especially since flash attention python library is not available for Apple Sillicon, you can't even run inference using Torch with MPS support. :(

Please internVL,. In my tests it works better than CogVLM. Especially for stuff like receipts and documents.

@fzzylogic
Copy link

fzzylogic commented Jun 9, 2024

InternVL is quite good. Benchmarks, HF, Demo.

@DoiiarX
Copy link

DoiiarX commented Jun 17, 2024

how about now? any update?

@James4Ever0
Copy link

upvote for this

@fzzylogic
Copy link

fzzylogic commented Jul 6, 2024

InternLM-XComposer-2.5-7b is out now out and having only tested the image capabilities, it seems great. HF, Demo.

@KOG-Nisse
Copy link

This would be great!

@v3ss0n
Copy link

v3ss0n commented Jul 9, 2024

Any status on this. this is currently highest performing Vision LLM from user's tests on LocalLLama reddit.

@suncloudsmoon
Copy link

Any updates?

@CNEA-lw
Copy link

CNEA-lw commented Jul 25, 2024

嘿,我有几个项目很忙,它在我的清单上,但目前优先级并不多。这真的只是我可以在业余时间做的事情

I tested the now available InternVL2 model and it is indeed a great choice, I hope to give it a higher priority, thank you for your hard work.

@goto-loop
Copy link

InternVL2 would be great to have! Seems to be SOTA in open source vision LLMs

@v3ss0n
Copy link

v3ss0n commented Jul 29, 2024

Any thoughts on this?
Since Vision models varies alot , compare to LLM models do Maintainers thinks LLamacpp should be focusing on supporting it? Since there are already a lot of LLM models coming out and the core team is doing tremendous work on those already. Do core team feels VLMs should be supported outside of llamacpp project?
May be addon/extention architecture viable?

@Backendmagier
Copy link

This would be a gamechanger! @cjpais

@cjpais
Copy link
Contributor

cjpais commented Aug 5, 2024

I'm sorry I don't know when I can do this, I have a huge backlog of projects I'm currently working on! I am very curious to try it but unfortunately it's not very high priority for me right now

@nogifeet
Copy link

InternVL2 would be great to have! Seems to be SOTA in open source vision LLMs

+1

@v3ss0n
Copy link

v3ss0n commented Aug 20, 2024

I think model builder should contribute their vision model works in here.

@felixslu
Copy link

I think model builder should contribute their vision model works in here.

In an ideal situation,it's model builder's work!
but sadly, maybe their work not focus on device,or they have self-deploy server framework,such as LMDeploy.

So, I really hope llama.cpp contributor can support this model, it is really good!

@ZhongQiyu
Copy link

I think the devs can add their own branches to the llama.cpp repo or huggingface.co?
The 2.5 version of InternVL also got released..can take it a try for transfer as a helper if needed.

@v3ss0n
Copy link

v3ss0n commented Sep 11, 2024

I think model builder should contribute their vision model works in here.

In an ideal situation,it's model builder's work! but sadly, maybe their work not focus on device,or they have self-deploy server framework,such as LMDeploy.

If they want to be popular and used by many , that would be the case.

LMDeploy is full of bufferoverflow crashes , not recommended for any secure deployment.

@James4Ever0
Copy link

James4Ever0 commented Sep 25, 2024

They have closed my issue for now. Guess this is never on their roadmap.

OpenGVLab/InternVL#522

@rampageservices
Copy link

They have closed my issue for now. Guess this is never on their roadmap.

OpenGVLab/InternVL#522

It was reopened and they stated they are actively working towards "progressing the work."

@github-actions github-actions bot added the stale label Nov 2, 2024
Copy link
Contributor

This issue was closed because it has been inactive for 14 days since being marked as stale.

@v3ss0n
Copy link

v3ss0n commented Nov 20, 2024

So , LLamacpp is not gonna focus on VLMs i guess.

@James4Ever0
Copy link

James4Ever0 commented Jan 14, 2025

Now I have made some progress.

#9403 qlylangyu#1

Please consider reopen this issue.

@rccrdmr
Copy link

rccrdmr commented Feb 5, 2025

Anyway we can extract confidence score?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request stale
Projects
None yet
Development

No branches or pull requests