You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
With VLLM 0.7.0 if I don't set VLLM_USE_V1=1 then it loads and runs fine but when setting it to 1 I get this error Model architectures ["VoxelPhiForCausalLM"] are not supported for now.
Is it possible to use V1 architecture with custom models?
Before submitting a new issue...
Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
The text was updated successfully, but these errors were encountered:
Can you share the full stacktrace? It is likely the real error is in the middle of the trace, and that message you shared is reporting the caught failure
VLMs don't support V1 by default. Please update your model by walking through the model development docs again. If that still doesn't work, can you share your model implementation?
@mgoin@DarkLight1337 Looks like it was crashing because I was running it in a notebook, when running it from the command line it works fine. There's a very small difference in behavior though: When running generation without V1 enabled output.outputs[0].text does not contain the EOS token but with V1 enabled it does. I'm guessing this is not intentional?
Your current environment
Model Input Dumps
No response
🐛 Describe the bug
I have a custom VisionLanguage model that I'm running using VLLM.
Before loading it I register it like this:
ModelRegistry.register_model("VoxelPhiForCausalLM", VoxelForCausalLM)
and then load it like so:
With VLLM 0.7.0 if I don't set VLLM_USE_V1=1 then it loads and runs fine but when setting it to 1 I get this error
Model architectures ["VoxelPhiForCausalLM"] are not supported for now.
Is it possible to use V1 architecture with custom models?
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: