You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to convert my torchscript module to tensorrt using torch_tensorrt.compile is there any argument to skip the layers which shows warnings while converting.
This is the warning its giving.
WARNING: [Torch-TensorRT TorchScript Conversion Context] - Running layernorm after self-attention in FP16 may cause overflow. Exporting the model to the latest available ONNX opset (later than opset 17) to use the INormalizationLayer, or forcing layernorm layers to run in FP32 precision can help with preserving accuracy. Although I am not converting to onnx at my end as I am using torch_tensorrt.
The text was updated successfully, but these errors were encountered:
I am trying to convert my torchscript module to tensorrt using torch_tensorrt.compile is there any argument to skip the layers which shows warnings while converting.
This is the warning its giving.
WARNING: [Torch-TensorRT TorchScript Conversion Context] - Running layernorm after self-attention in FP16 may cause overflow. Exporting the model to the latest available ONNX opset (later than opset 17) to use the INormalizationLayer, or forcing layernorm layers to run in FP32 precision can help with preserving accuracy. Although I am not converting to onnx at my end as I am using torch_tensorrt.
The text was updated successfully, but these errors were encountered: