Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Local run instructions please #14

Open
ramkumarkoppu opened this issue Dec 2, 2024 · 2 comments
Open

Local run instructions please #14

ramkumarkoppu opened this issue Dec 2, 2024 · 2 comments

Comments

@ramkumarkoppu
Copy link

Hi,

Can you please provide instructions to run this model locally Linux/Mac computer?

@XuGW-Kevin
Copy link
Collaborator

Hi, to run this model locally on Linux/Mac computer, a good way is to convert the model to GGUF and use ollama.
However, llama.cpp does not support this feature.
It would be very hard for me to implement this, but it seems that the community has been working on this.

@XuGW-Kevin
Copy link
Collaborator

XuGW-Kevin commented Dec 2, 2024

However, you can try running a quantized version of the model locally.
I found that the repository at unslothai supports quick quantization for using models in the llama-3.2-vision series, though I haven’t tried it myself. I find a demo here where you can run the model on a single T4 GPU.
If you’re willing to give it a try and encounter any issues, I’d be happy to work together to solve them. Thank you for your interest!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants