-
-
Notifications
You must be signed in to change notification settings - Fork 5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Misc]: Support for Shieldgemma model #7084
Comments
The error means that you don't have FlashInfer installed. Please follow the steps shared here. |
Thanks, there was an error with the flashinfer installed, so I tried the flashinfer installation using this
Now, I have a different error:
UPDATE: Looks similar to this error #7070 This works. Please use Thanks @DarkLight1337 |
Thanks @DarkLight1337 |
Is there more context on the change the OP made in regards to the "hidden_act" versus "hidden_activation" reference? I am seeing the following error as well:
|
@JerryGamble1 When the weights are downloaded, please change "hidden_activation" to "hidden_act" in the file "config.json". Usually the weights are present in the huggingface cache directory. https://huggingface.co/google/shieldgemma-2b/blob/main/config.json If you use the command |
We've moved on from trying to get this to work on VLLM for now so no need to respond, but just FYI... Modifying the config file I was able to load the model into VLLM, but every requests generates a bad request error with this log message...
|
Trying to run the Shieldgemma model.
The architecture is Gemma2ForCausalLM which should be already supported. The config file specifies the transformers version to be 4.42.4.
I have the following installed:
I have also the Transformers 4.43.3.
After checking the config file, it appears that the config specifies
hidden_activation
instead ofhidden_act
. After changing it manually in the config.json file, I get an error which specifies that I should use flashinfer backend.VLLM_ATTENTION_BACKEND=FLASHINFER
After which, the following error is occurring:
The text was updated successfully, but these errors were encountered: