-
Notifications
You must be signed in to change notification settings - Fork 109
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
trt conversion #110
Comments
改一下tensorrt的版本 |
Hi, I've already tried tensorrt versions 8.6 and 10.7
however, it produced a segmentation fault I produce the model engine using the command
|
The version 10.5.0 of TensorRT works fine for me. By the way, You need to ensure that the version of trtexec you use for model conversion matches the version of the Python TensorRT API you are using. |
Hi @SebastianJanampa , were you able to solve this issue by using a different TensorRT version? |
Hi @migsdigs |
Hi again, thanks for letting me know. Strangely enough, I tried 10.5 and it seems to improve the accuracy of the inference a bit, but the inference speed seems very slow - at least slower than real time. Although, I am using CUDA 12.4, so not sure if there might be some conflict there. I see in the repo they recommend Tensorrt 10.4, so I will try that. I was using Tensorrt 10.7 before and the inference was fast but very low accuracy on fp16, and not much improvement on fp32 |
Hello,
Thanks for the incredible work you've done. I tried to convert the model to tensorrt fp 16, but I encountered a problem with segmentation. However, if I remove --fp16, it works (I know it computes a model with fp32).
I just wanted to see how much faster the model became after using trt (I am new to this, so I was curious), and I faced another issue.
Could anyone help me, please?
The text was updated successfully, but these errors were encountered: