Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add tensorrt quantization #172

Merged
merged 13 commits into from
May 9, 2022
Merged

add tensorrt quantization #172

merged 13 commits into from
May 9, 2022

Conversation

Data-Iab
Copy link
Collaborator

@Data-Iab Data-Iab commented Apr 22, 2022

  • New 💥 :

    • TensorRt engines can now be built with int8 precision using Post Training Quantization.
    • 4 calibrators are available for quantization : MinMaxCalibrator, LegacyCalibrator, EntropyCalibrator and EntropyCalibrator2.
    • Added a QuantizedModel interface to convert model to quantized model for Training Aware Quantization.
  • Fixed 🔧 :

    • Adapt graph option is removed, we just adapt graph once it's exported from torch to ONNX.

@Data-Iab Data-Iab self-assigned this Apr 22, 2022
@thibo73800 thibo73800 merged commit 4caa0d2 into master May 9, 2022
@thibo73800 thibo73800 deleted the dev branch May 9, 2022 15:48
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants