Hi,
Just tried to quantize Mistral-Small-3.1-24B with GPTQ using your example script.
Everything went smoothly but when I tried to load the model with vLLM, I got the following error message
ValueError: There is no module or parameter named 'multi_modal_projector.patch_merger.merging_layer.weight_packed' in Mistral3ForConditionalGeneration
I'm using one of the latest commits on the main branch of vLLM (0.9.2rc2.dev269+gf29fd8a7f) and same for llm-compressor (llmcompressor-0.6.1.dev35+g53240c63).
Thanks!