diff --git a/llama_bringup/models/Qwen2-VL.yaml b/llama_bringup/models/Qwen2-VL.yaml index 253ffe0..fde8791 100644 --- a/llama_bringup/models/Qwen2-VL.yaml +++ b/llama_bringup/models/Qwen2-VL.yaml @@ -6,13 +6,13 @@ n_gpu_layers: 29 n_threads: -1 n_predict: 8192 +image_prefix: "<|vision_start|>" +image_suffix: "<|vision_end|>" + model_repo: "bartowski/Qwen2-VL-2B-Instruct-GGUF" model_filename: "Qwen2-VL-2B-Instruct-Q4_K_M.gguf" mmproj_repo: "bartowski/Qwen2-VL-2B-Instruct-GGUF" mmproj_filename: "mmproj-Qwen2-VL-2B-Instruct-f16.gguf" -image_prefix: "<|vision_start|>" -image_suffix: "<|vision_end|>" - system_prompt_type: "ChatML" diff --git a/llama_ros/src/llava_ros/llava.cpp b/llama_ros/src/llava_ros/llava.cpp index b3277bf..b32580b 100644 --- a/llama_ros/src/llava_ros/llava.cpp +++ b/llama_ros/src/llava_ros/llava.cpp @@ -245,7 +245,6 @@ bool Llava::eval_prompt() { } // eval the image - LLAMA_LOG_INFO("Checking the image"); if (this->image_embed != nullptr) { LLAMA_LOG_INFO("Evaluating the image");