Skip to content

Conversation

@lhez
Copy link
Collaborator

@lhez lhez commented Nov 1, 2025

This PR adds support for imrope.

@github-actions github-actions bot added ggml changes relating to the ggml tensor library for machine learning OpenCL Issues specific to the OpenCL backend labels Nov 1, 2025
@lhez lhez marked this pull request as ready for review November 3, 2025 16:49
@lhez lhez requested a review from max-krasnyansky as a code owner November 3, 2025 16:49
@max-krasnyansky
Copy link
Collaborator

Looks good.
I did a quick test with Qwen3-VL-2B on the Gen5 device.

./scripts/snapdragon/adb/run-tool.sh llama-mtmd-cli -v -m ../gguf/Qwen3-VL-2B-Instruct-Q8_0.gguf --mmproj ../gguf/mmproj-Qwen3-VL-2B-Instruct-Q8_0.gguf -t 6 --device GPUOpenCL --no-mmap -fa 1 --image bmw-ix3.jpg -p \"what do you see on the picture?\".
...
Based on the image provided, here is a detailed analysis of the vehicle:

The vehicle shown is the **BMW iX3**, ...

@max-krasnyansky max-krasnyansky merged commit c5023da into ggml-org:master Nov 3, 2025
124 of 130 checks passed
gabe-l-hart added a commit to gabe-l-hart/llama.cpp that referenced this pull request Nov 3, 2025
* origin/master: (169 commits)
opencl: support imrope (ggml-org#16914)
fix: Viewing multiple PDF attachments (ggml-org#16974)
model-conversion : pass config to from_pretrained (ggml-org#16963)
server : add props.model_alias (ggml-org#16943)
ggml: CUDA: add head size 72 for flash-attn (ggml-org#16962)
mtmd: add --image-min/max-tokens (ggml-org#16921)
mtmd: pad mask for qwen2.5vl (ggml-org#16954)
ggml : LoongArch fixes (ggml-org#16958)
sync: minja (glm 4.6 & minmax m2 templates) (ggml-org#16949)
SYCL: optimized repeat_back kernel (3× fewer asm instructions, 2× faster)Feature/sycl repeat back opt (ggml-org#16869)
feat(webui): improve LaTeX rendering with currency detection (ggml-org#16508)
test-backend-ops : fix segfault in moe-expert-reduce test in support mode and coverage (ggml-org#16936)
ci : disable failing riscv cross build (ggml-org#16952)
model: add Janus Pro for image understanding (ggml-org#16906)
clip : use FA (ggml-org#16837)
server : support unified cache across slots (ggml-org#16736)
common : move gpt-oss reasoning processing to init params (ggml-org#16937)
docs: remove llama_sampler_accept reference in sampling sample usage (ggml-org#16920)
CUDA: add FLOOR, CEIL, ROUND, TRUNC unary ops (ggml-org#16917)
devops: fix failing s390x docker build (ggml-org#16918)
...
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ggml changes relating to the ggml tensor library for machine learning OpenCL Issues specific to the OpenCL backend

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants