-
Notifications
You must be signed in to change notification settings - Fork 548
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Improving inference time #321
Comments
Thanks for raising the question. But that does not apply to the code in this repo. If you can read CUDA code, you'd see that I don't do NMS in the yolo_layer plugin. tensorrt_demos/plugins/yolo_layer.cu Lines 198 to 249 in 8b81bbc
Instead, I do NMS with python as shown below. The NMS code is written in python and indeed could be slow. You might improve FPS by optimizing this part (for example, replace it with C++ code). tensorrt_demos/utils/yolo_with_plugins.py Lines 139 to 146 in 8b81bbc
|
Thanks for the detailed explanation. I'll think about optimizing the NMS code part with C++ code. |
Hi,
I came across this trick for improving inference time. opencv/opencv#14827 (comment)
While converting yolov4.cfg to tensort engine file, The cfg file that Used did not have nms-threshold=0 set in all yolo layers.
I do see some code in this repo for nms boxes etc.
Will i get any speedup in inference time of yolov4 if I add the values in cfg file? Or is this already taken care of while building trt engine?
Thanks
The text was updated successfully, but these errors were encountered: