Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

bug: Missing NGL Setting for Vision Models #1763

Closed
3 tasks
imtuyethan opened this issue Dec 3, 2024 · 2 comments
Closed
3 tasks

bug: Missing NGL Setting for Vision Models #1763

imtuyethan opened this issue Dec 3, 2024 · 2 comments
Assignees
Labels
category: tools RAG, function calling, etc type: bug Something isn't working

Comments

@imtuyethan
Copy link
Contributor

Jan version

0.5.10

Describe the Bug

https://discord.com/channels/1107178041848909847/1307915350406467664/1313079041821249556
The vision model configuration in model.json lacks an NGL (number of GPU layers) setting, which is needed to properly manage GPU resource allocation when running multiple models simultaneously (text model + vision/projector model). This causes uncertainty about proper GPU layer allocation between models.

Steps to Reproduce

  1. Load a vision model with the following settings:
"settings": {
    "vision_model": true,
    "text_model": false,
    "ctx_len": 4096,
    "prompt_template": "\n### Instruction:\n{prompt}\n### Response:\n",
    "llama_model_path": "llava-v1.6-mistral-7b.Q4_K_M.gguf",
    "mmproj": "mmproj-model-f16.gguf"
}
  1. Notice that NGL setting is missing, making it unclear how to allocate GPU layers when running multiple models

Screenshots / Logs

Screenshot 2024-12-03 at 9 48 08 AM Screenshot 2024-12-03 at 9 48 13 AM Screenshot 2024-12-03 at 9 48 18 AM Screenshot 2024-12-03 at 9 48 24 AM

What is your OS?

  • MacOS
  • Windows
  • Linux
@imtuyethan imtuyethan added the type: bug Something isn't working label Dec 3, 2024
@imtuyethan imtuyethan added the category: tools RAG, function calling, etc label Dec 3, 2024
@imtuyethan imtuyethan transferred this issue from janhq/jan Dec 3, 2024
@louis-jan
Copy link
Contributor

Can check after supporting vision models pull.

@louis-jan
Copy link
Contributor

On this one, it's more about supporting vision models which actually here #1493. There will be an action item to pull correct NGL from Chat model (not projector model). I tried it, and it really sped up. For now we will close this one and refer to the original epic as an action item, since now we temporarily fixed the listed models on the model hub.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
category: tools RAG, function calling, etc type: bug Something isn't working
Projects
Status: QA
Development

No branches or pull requests

4 participants