-
-
Notifications
You must be signed in to change notification settings - Fork 11.2k
Closed
Labels
feature requestNew feature or requestNew feature or request
Description
🚀 The feature, motivation and pitch
I am using vLLM for hosting of LLMs/SLMs and with the recent release of Gemma 3, I would love to have it supported in vLLM. Google has stated Gemma 3 has day 1 support from HF Transformers, so it should (hopefully) be relatively simple to integrate into vLLM.
Currently, when attempting to load google/gemma-3-12b-it, the following error is given:
ERROR 03-12 18:19:00 engine.py:400] ValueError: The checkpoint you are trying to load has model type `gemma3` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
ERROR 03-12 18:19:00 engine.py:400]
ERROR 03-12 18:19:00 engine.py:400] You can update Transformers with the command `pip install --upgrade transformers`. If this does not work, and the checkpoint is very new, then there may not be a release version that supports this model yet. In this case, you can get the most up-to-date code by installing Transformers from source with the command `pip install git+https://github.com/huggingface/transformers.git`
Alternatives
No response
Additional context
https://blog.google/technology/developers/gemma-3/
https://developers.googleblog.com/en/introducing-gemma3/
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
stelterlab, bd-estorm, galatolofederico, scheckley, hanyangbot and 10 more
Metadata
Metadata
Assignees
Labels
feature requestNew feature or requestNew feature or request