-
Notifications
You must be signed in to change notification settings - Fork 677
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Decreasing inference time on cpu #31
Comments
It is a good question, but we have not explored it yet. The straightest way is to deploy the model like ONNX. |
@SlongLiu should the operations in the model already support ONNX export? e.g. code vaguely similar to the below should work?
I'll look in to optimising it with tools like OpenVINO if it does. |
i try to use torch.onnx.export to transfer grounding-dino to onnx, it seems like some problem , such as several logical operators are not supported by onnx. for example: |
Have you solve it please |
am also interested in onx export code of GDINO |
Thanks for this awesome model, it does evaluate good with pretrained model
Right now i am getting 15s average inference time, any way to reduce it to 2-3s
The text was updated successfully, but these errors were encountered: