diff --git a/.circleci/config.yml b/.circleci/config.yml index f1143f25b0..d1e36447d3 100644 --- a/.circleci/config.yml +++ b/.circleci/config.yml @@ -802,7 +802,7 @@ commands: - store_artifacts: path: /tmp/testlogs - test-dynamo-models_torch_export: + test-dynamo-models_export: description: "Test the Dynamo models via torch_export path" steps: - run: @@ -818,6 +818,20 @@ commands: - store_artifacts: path: /tmp/testlogs + test-dynamo-export_serde: + description: "Test the export serialize/deserialize functionality for Dynamo models" + steps: + - run: + name: Run Dynamo models and test export serde with TRT compiled modules + command: | + cd tests/py/dynamo/models + pytest test_export_serde.py --junitxml=/tmp/artifacts/test_results/dynamo/backend/test_results.xml --ir dynamo + + - store_test_results: + path: /tmp/artifacts + - store_artifacts: + path: /tmp/testlogs + test-dynamo-converters: description: "Test the Dynamo aten converters" steps: @@ -1122,7 +1136,8 @@ jobs: - test-dynamo-backend - test-dynamo-shared_utilities - test-dynamo-models_torch_compile - - test-dynamo-models_torch_export + - test-dynamo-models_export + - test-dynamo-export_serde package-x86_64-linux: parameters: diff --git a/.github/workflows/build-test.yml b/.github/workflows/build-test.yml index 1c9f376659..bfc19cce45 100644 --- a/.github/workflows/build-test.yml +++ b/.github/workflows/build-test.yml @@ -54,39 +54,40 @@ jobs: AWS_PYTORCH_UPLOADER_ACCESS_KEY_ID: ${{ secrets.AWS_PYTORCH_UPLOADER_ACCESS_KEY_ID }} AWS_PYTORCH_UPLOADER_SECRET_ACCESS_KEY: ${{ secrets.AWS_PYTORCH_UPLOADER_SECRET_ACCESS_KEY }} - # tests-py-torchscript-fe: - # name: Test torchscript frontend [Python] - # needs: [generate-matrix, build] - # strategy: - # fail-fast: false - # matrix: - # include: - # - repository: pytorch/tensorrt - # package-name: torch_tensorrt - # pre-script: packaging/pre_build_script.sh - # uses: pytorch/tensorrt/.github/workflows/linux-test.yml@main - # with: - # job-name: tests-py-torchscript-fe - # repository: "pytorch/tensorrt" - # ref: "" - # test-infra-repository: pytorch/test-infra - # test-infra-ref: main - # build-matrix: ${{ needs.generate-matrix.outputs.matrix }} - # pre-script: ${{ matrix.pre-script }} - # script: | - # export USE_HOST_DEPS=1 - # pushd . - # cd tests/modules - # ${CONDA_RUN} python -m pip install -r requirements.txt - # ${CONDA_RUN} python hub.py - # popd - # pushd . - # cd tests/py/ts - # ${CONDA_RUN} python -m pip install --pre pytest timm transformers parameterized expecttest --use-deprecated=legacy-resolver - # ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/ts_api_test_results.xml api/ - # ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/ts_models_test_results.xml models/ - # ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/ts_integrations_test_results.xml integrations/ - # popd + tests-py-torchscript-fe: + name: Test torchscript frontend [Python] + needs: [generate-matrix, build] + strategy: + fail-fast: false + matrix: + include: + - repository: pytorch/tensorrt + package-name: torch_tensorrt + pre-script: packaging/pre_build_script.sh + uses: pytorch/tensorrt/.github/workflows/linux-test.yml@main + with: + job-name: tests-py-torchscript-fe + repository: "pytorch/tensorrt" + ref: "" + test-infra-repository: pytorch/test-infra + test-infra-ref: main + build-matrix: ${{ needs.generate-matrix.outputs.matrix }} + pre-script: ${{ matrix.pre-script }} + script: | + export USE_HOST_DEPS=1 + export LD_LIBRARY_PATH=/usr/lib64:$LD_LIBRARY_PATH + pushd . + cd tests/modules + ${CONDA_RUN} python -m pip install --pre -r requirements.txt --use-deprecated=legacy-resolver + ${CONDA_RUN} python hub.py + popd + pushd . + cd tests/py/ts + ${CONDA_RUN} python -m pip install --pre pytest timm transformers parameterized expecttest --use-deprecated=legacy-resolver + ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/ts_api_test_results.xml api/ + ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/ts_models_test_results.xml models/ + ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/ts_integrations_test_results.xml integrations/ + popd tests-py-dynamo-converters: name: Test dynamo converters [Python] @@ -140,6 +141,8 @@ jobs: cd tests/py/dynamo ${CONDA_RUN} python -m pip install --pre pytest timm transformers parameterized expecttest --use-deprecated=legacy-resolver ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/dynamo_fe_test_results.xml --ir dynamo models/test_models_export.py + ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/export_serde_test_results.xml --ir dynamo models/test_export_serde.py + ${CONDA_RUN} python -m pytest --junitxml=${RUNNER_TEST_RESULTS_DIR}/dyn_models_export.xml --ir dynamo models/test_dyn_models.py popd tests-py-torch-compile-be: diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index ac24623eef..4738ea80be 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -40,7 +40,7 @@ repos: rev: 'v1.4.1' hooks: - id: mypy - exclude: "^py/torch_tensorrt/fx|^examples|^tests|^tools|^docs|noxfile.py|setup.py|versions.py" + exclude: "^py/torch_tensorrt/fx|^examples|^tests|^py/torch_tensorrt/dynamo/_experimental|^tools|^docs|noxfile.py|setup.py|versions.py" - repo: https://github.com/astral-sh/ruff-pre-commit # Ruff version. rev: v0.0.278 diff --git a/core/conversion/converters/impl/shuffle.cpp b/core/conversion/converters/impl/shuffle.cpp index 5a8e992d90..8d25525a6e 100644 --- a/core/conversion/converters/impl/shuffle.cpp +++ b/core/conversion/converters/impl/shuffle.cpp @@ -20,7 +20,12 @@ static auto shuffle_registrations TORCHTRT_UNUSED = auto in_shape = util::toVec(in->getDimensions()); std::vector out_shape; if (ctx->input_is_dynamic) { - end_dim = (end_dim == -1) ? in_shape.size() - 1 : end_dim; + if (start_dim < 0) { + start_dim = start_dim + in_shape.size(); + } + if (end_dim < 0) { + end_dim = end_dim + in_shape.size(); + } int nbDynamicFlattenedDims = 0; int nbDynamicUnflattenedDims = 0; for (int i = 0; i < (int)in_shape.size(); i++) { diff --git a/core/runtime/execute_engine.cpp b/core/runtime/execute_engine.cpp index c4b34cb218..2a7fe884da 100644 --- a/core/runtime/execute_engine.cpp +++ b/core/runtime/execute_engine.cpp @@ -43,8 +43,8 @@ bool is_switch_required(const RTDevice& curr_device, const RTDevice& engine_devi return false; } -RTDevice select_rt_device(const RTDevice& engine_device) { - auto new_target_device_opt = get_most_compatible_device(engine_device); +RTDevice select_rt_device(const RTDevice& engine_device, const RTDevice& curr_device) { + auto new_target_device_opt = get_most_compatible_device(engine_device, curr_device); // REVIEW: THIS DOES NOT LIST DLA PROBABLY, WHICH WE SHOULD // TODO: I think this logic could be way simpler at execution time since if the tensors arent on the right @@ -89,7 +89,7 @@ std::vector execute_engine(std::vector inputs, c10::intr if (is_switch_required(curr_device, compiled_engine->device_info)) { // Scan through available CUDA devices and set the CUDA device context correctly - RTDevice device = select_rt_device(compiled_engine->device_info); + RTDevice device = select_rt_device(compiled_engine->device_info, curr_device); set_rt_device(device); // Target device is new device diff --git a/core/runtime/runtime.cpp b/core/runtime/runtime.cpp index 0c054d8a3c..0372258919 100644 --- a/core/runtime/runtime.cpp +++ b/core/runtime/runtime.cpp @@ -7,9 +7,16 @@ namespace torch_tensorrt { namespace core { namespace runtime { -c10::optional get_most_compatible_device(const RTDevice& target_device) { +c10::optional get_most_compatible_device(const RTDevice& target_device, const RTDevice& curr_device) { LOG_DEBUG("Target Device: " << target_device); auto device_options = find_compatible_devices(target_device); + RTDevice current_device; + if (current_device.id == -1) { + current_device = get_current_device(); + } else { + current_device = curr_device; + } + if (device_options.size() == 0) { return {}; } else if (device_options.size() == 1) { @@ -21,10 +28,20 @@ c10::optional get_most_compatible_device(const RTDevice& target_device dev_list << "[" << std::endl; for (auto device : device_options) { dev_list << " " << device << ',' << std::endl; - if (device.device_name == target_device.device_name && best_match.device_name != target_device.device_name) { - best_match = device; - } else if (device.device_name == target_device.device_name && best_match.device_name == target_device.device_name) { - if (device.id == target_device.id && best_match.id != target_device.id) { + if (device.device_name == target_device.device_name) { + // First priority is selecting a candidate which agrees with the current device ID + // If such a device is found, we can select it and break out of the loop + if (device.id == current_device.id && best_match.id != current_device.id) { + best_match = device; + break; + } + // Second priority is selecting a candidate which agrees with the target device ID + // At deserialization time, the current device and target device may not agree + else if (device.id == target_device.id && best_match.id != target_device.id) { + best_match = device; + } + // If no such GPU ID is found, select the first available candidate GPU + else if (best_match.device_name != target_device.device_name) { best_match = device; } } diff --git a/core/runtime/runtime.h b/core/runtime/runtime.h index 4c7565c9fc..05d97a30b8 100644 --- a/core/runtime/runtime.h +++ b/core/runtime/runtime.h @@ -26,7 +26,9 @@ typedef enum { SERIALIZATION_LEN, // NEVER USED FOR DATA, USED TO DETERMINE LENGTH OF SERIALIZED INFO } SerializedInfoIndex; -c10::optional get_most_compatible_device(const RTDevice& target_device); +c10::optional get_most_compatible_device( + const RTDevice& target_device, + const RTDevice& curr_device = RTDevice()); std::vector find_compatible_devices(const RTDevice& target_device); std::vector execute_engine(std::vector inputs, c10::intrusive_ptr compiled_engine); diff --git a/core/util/trt_util.cpp b/core/util/trt_util.cpp index 77c88b465d..50b58a0bdb 100644 --- a/core/util/trt_util.cpp +++ b/core/util/trt_util.cpp @@ -216,7 +216,7 @@ nvinfer1::Dims squeezeDims(const nvinfer1::Dims& d, int pos, bool use_zeros, boo // Replace all instances of -1, indicating dynamic dimension // with 0, indicating copy the dimension from another tensor // (Generally used for reshape operations) - if (use_zeros && d.d[i] == -1) { + if (use_zeros && d.d[i] == -1 && i < pos) { dims.d[j] = 0; // If zeros already exist in the dimensions (empty tensor), // Replace all instances of 0, indicating empty dimension diff --git a/cpp/include/torch_tensorrt/torch_tensorrt.h b/cpp/include/torch_tensorrt/torch_tensorrt.h index 29f860c8b3..adac75d984 100644 --- a/cpp/include/torch_tensorrt/torch_tensorrt.h +++ b/cpp/include/torch_tensorrt/torch_tensorrt.h @@ -60,6 +60,8 @@ class DataType { enum Value : int8_t { /// INT64 kLong, + /// FP64 + kDouble, /// FP32 kFloat, /// FP16 diff --git a/cpp/src/types.cpp b/cpp/src/types.cpp index 2be7fea338..69b956a162 100644 --- a/cpp/src/types.cpp +++ b/cpp/src/types.cpp @@ -97,6 +97,8 @@ at::ScalarType toAtenDataType(DataType value) { return at::kInt; case DataType::kLong: return at::kLong; + case DataType::kDouble: + return at::kDouble; case DataType::kBool: return at::kBool; case DataType::kFloat: @@ -119,7 +121,8 @@ nvinfer1::TensorFormat toTRTTensorFormat(TensorFormat value) { DataType::DataType(c10::ScalarType t) { TORCHTRT_CHECK( - t == at::kHalf || t == at::kFloat || t == at::kChar || t == at::kLong || t == at::kInt || t == at::kBool, + t == at::kHalf || t == at::kFloat || t == at::kChar || t == at::kLong || t == at::kDouble || t == at::kInt || + t == at::kBool, "Data type is unsupported (" << t << ")"); switch (t) { case at::kHalf: @@ -134,6 +137,9 @@ DataType::DataType(c10::ScalarType t) { case at::kLong: value = DataType::kLong; break; + case at::kDouble: + value = DataType::kDouble; + break; case at::kBool: value = DataType::kBool; break; diff --git a/docker/WORKSPACE.ngc b/docker/WORKSPACE.ngc index a01bbed32b..c3d9bea0fc 100755 --- a/docker/WORKSPACE.ngc +++ b/docker/WORKSPACE.ngc @@ -9,24 +9,28 @@ http_archive( sha256 = "778197e26c5fbeb07ac2a2c5ae405b30f6cb7ad1f5510ea6fdac03bded96cc6f", ) -load("@rules_python//python:pip.bzl", "pip_install") +load("@rules_python//python:repositories.bzl", "py_repositories") + +py_repositories() http_archive( name = "rules_pkg", + sha256 = "8f9ee2dc10c1ae514ee599a8b42ed99fa262b757058f65ad3c384289ff70c4b8", urls = [ - "https://mirror.bazel.build/github.com/bazelbuild/rules_pkg/releases/download/0.4.0/rules_pkg-0.4.0.tar.gz", - "https://github.com/bazelbuild/rules_pkg/releases/download/0.4.0/rules_pkg-0.4.0.tar.gz", + "https://mirror.bazel.build/github.com/bazelbuild/rules_pkg/releases/download/0.9.1/rules_pkg-0.9.1.tar.gz", + "https://github.com/bazelbuild/rules_pkg/releases/download/0.9.1/rules_pkg-0.9.1.tar.gz", ], - sha256 = "038f1caa773a7e35b3663865ffb003169c6a71dc995e39bf4815792f385d837d", ) + load("@rules_pkg//:deps.bzl", "rules_pkg_dependencies") + rules_pkg_dependencies() -git_repository( +http_archive( name = "googletest", - remote = "https://github.com/google/googletest", - commit = "703bd9caab50b139428cea1aaff9974ebee5742e", - shallow_since = "1570114335 -0400" + sha256 = "755f9a39bc7205f5a0c428e920ddad092c33c8a1b46997def3f1d4a82aded6e1", + strip_prefix = "googletest-5ab508a01f9eb089207ee87fd547d290da39d015", + urls = ["https://github.com/google/googletest/archive/5ab508a01f9eb089207ee87fd547d290da39d015.zip"], ) # External dependency for torch_tensorrt if you already have precompiled binaries. @@ -80,17 +84,13 @@ new_local_repository( ######################################################################### # Testing Dependencies (optional - comment out on aarch64) ######################################################################### -pip_install( - name = "torch_tensorrt_py_deps", - requirements = "//py:requirements.txt", -) +load("@rules_python//python:pip.bzl", "pip_parse") -pip_install( - name = "py_test_deps", - requirements = "//tests/py:requirements.txt", +pip_parse( + name = "devtools_deps", + requirements_lock = "//:requirements-dev.txt", ) -pip_install( - name = "pylinter_deps", - requirements = "//tools/linter:requirements.txt", -) +load("@devtools_deps//:requirements.bzl", "install_deps") + +install_deps() diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1DataType.html b/docs/_cpp_api/classtorch__tensorrt_1_1DataType.html index 46ff961782..017a76cab6 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1DataType.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1DataType.html @@ -10,7 +10,7 @@ - Class DataType — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Class DataType — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
- v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
@@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    @@ -414,6 +417,12 @@

    Class Documentation +
    +enumerator kDouble
    +

    FP64.

    +
    +
    enumerator kFloat
    diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.html b/docs/_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.html index 1c02e8fcaa..3174050c75 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.html @@ -10,7 +10,7 @@ - Class Device::DeviceType — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Class Device::DeviceType — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1TensorFormat.html b/docs/_cpp_api/classtorch__tensorrt_1_1TensorFormat.html index 368de94a2c..f430de084f 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1TensorFormat.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1TensorFormat.html @@ -10,7 +10,7 @@ - Class TensorFormat — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Class TensorFormat — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.html b/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.html index 6237f58c12..f40acf541f 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.html @@ -10,7 +10,7 @@ - Template Class Int8CacheCalibrator — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Template Class Int8CacheCalibrator — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.html b/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.html index b73bcd333f..8bc47d8863 100644 --- a/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.html +++ b/docs/_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.html @@ -10,7 +10,7 @@ - Template Class Int8Calibrator — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Template Class Int8Calibrator — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.html b/docs/_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.html index 9a82426bd4..36172e8d88 100644 --- a/docs/_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.html +++ b/docs/_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.html @@ -10,7 +10,7 @@ - Define STR — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Define STR — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.html b/docs/_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.html index f3e0818af1..64dca5bae5 100644 --- a/docs/_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.html +++ b/docs/_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.html @@ -10,7 +10,7 @@ - Define TORCH_TENSORRT_PATCH_VERSION — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Define TORCH_TENSORRT_PATCH_VERSION — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.html b/docs/_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.html index ac3a9b0a33..bbc2e3b510 100644 --- a/docs/_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.html +++ b/docs/_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.html @@ -10,7 +10,7 @@ - Define TORCH_TENSORRT_MAJOR_VERSION — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Define TORCH_TENSORRT_MAJOR_VERSION — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.html b/docs/_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.html index 25f5857465..11ed445cf3 100644 --- a/docs/_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.html +++ b/docs/_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.html @@ -10,7 +10,7 @@ - Define TORCH_TENSORRT_MINOR_VERSION — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Define TORCH_TENSORRT_MINOR_VERSION — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.html b/docs/_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.html index 2072390a7c..adc573b1a1 100644 --- a/docs/_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.html +++ b/docs/_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.html @@ -10,7 +10,7 @@ - Define TORCHTRT_API — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Define TORCHTRT_API — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.html b/docs/_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.html index 3e2639f77c..a0d280a396 100644 --- a/docs/_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.html +++ b/docs/_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.html @@ -10,7 +10,7 @@ - Define XSTR — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Define XSTR — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.html b/docs/_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.html index 1a42e110cb..39eb42060f 100644 --- a/docs/_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.html +++ b/docs/_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.html @@ -10,7 +10,7 @@ - Define TORCHTRT_HIDDEN — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Define TORCHTRT_HIDDEN — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.html b/docs/_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.html index 53c2eb4fc9..7223937f96 100644 --- a/docs/_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.html +++ b/docs/_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.html @@ -10,7 +10,7 @@ - Define TORCH_TENSORRT_VERSION — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Define TORCH_TENSORRT_VERSION — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/dir_cpp.html b/docs/_cpp_api/dir_cpp.html index 2d1b3b4956..1aa914ea2c 100644 --- a/docs/_cpp_api/dir_cpp.html +++ b/docs/_cpp_api/dir_cpp.html @@ -10,7 +10,7 @@ - Directory cpp — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Directory cpp — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/dir_cpp_include.html b/docs/_cpp_api/dir_cpp_include.html index 8a54330c6a..b339b3302b 100644 --- a/docs/_cpp_api/dir_cpp_include.html +++ b/docs/_cpp_api/dir_cpp_include.html @@ -10,7 +10,7 @@ - Directory include — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Directory include — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/dir_cpp_include_torch_tensorrt.html b/docs/_cpp_api/dir_cpp_include_torch_tensorrt.html index 14b47a4304..e3e39d74fb 100644 --- a/docs/_cpp_api/dir_cpp_include_torch_tensorrt.html +++ b/docs/_cpp_api/dir_cpp_include_torch_tensorrt.html @@ -10,7 +10,7 @@ - Directory torch_tensorrt — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Directory torch_tensorrt — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/enum_namespacetorch__tensorrt_1_1logging_1a130f65408ad8cbaee060f05e8db69558.html b/docs/_cpp_api/enum_namespacetorch__tensorrt_1_1logging_1a130f65408ad8cbaee060f05e8db69558.html index 572b580d34..bb4cc0a0e6 100644 --- a/docs/_cpp_api/enum_namespacetorch__tensorrt_1_1logging_1a130f65408ad8cbaee060f05e8db69558.html +++ b/docs/_cpp_api/enum_namespacetorch__tensorrt_1_1logging_1a130f65408ad8cbaee060f05e8db69558.html @@ -10,7 +10,7 @@ - Enum Level — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Enum Level — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/enum_namespacetorch__tensorrt_1a3fbe5d72e4fc624dbd038853079620eb.html b/docs/_cpp_api/enum_namespacetorch__tensorrt_1a3fbe5d72e4fc624dbd038853079620eb.html index f89b74848f..edb4e1c7ec 100644 --- a/docs/_cpp_api/enum_namespacetorch__tensorrt_1a3fbe5d72e4fc624dbd038853079620eb.html +++ b/docs/_cpp_api/enum_namespacetorch__tensorrt_1a3fbe5d72e4fc624dbd038853079620eb.html @@ -10,7 +10,7 @@ - Enum EngineCapability — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Enum EngineCapability — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/file_cpp_include_torch_tensorrt_logging.h.html b/docs/_cpp_api/file_cpp_include_torch_tensorrt_logging.h.html index d2bb42d66a..7a3f718225 100644 --- a/docs/_cpp_api/file_cpp_include_torch_tensorrt_logging.h.html +++ b/docs/_cpp_api/file_cpp_include_torch_tensorrt_logging.h.html @@ -10,7 +10,7 @@ - File logging.h — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + File logging.h — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/file_cpp_include_torch_tensorrt_macros.h.html b/docs/_cpp_api/file_cpp_include_torch_tensorrt_macros.h.html index 498777838c..2c4bab9358 100644 --- a/docs/_cpp_api/file_cpp_include_torch_tensorrt_macros.h.html +++ b/docs/_cpp_api/file_cpp_include_torch_tensorrt_macros.h.html @@ -10,7 +10,7 @@ - File macros.h — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + File macros.h — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.html b/docs/_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.html index baaaaa884b..4564e497f4 100644 --- a/docs/_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.html +++ b/docs/_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.html @@ -10,7 +10,7 @@ - File ptq.h — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + File ptq.h — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.html b/docs/_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.html index 345adcdf1c..cc14021d58 100644 --- a/docs/_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.html +++ b/docs/_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.html @@ -10,7 +10,7 @@ - File torch_tensorrt.h — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + File torch_tensorrt.h — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0593f776f469c20469e2f729fc7861a3.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0593f776f469c20469e2f729fc7861a3.html index fe0c8e5276..577c5acacc 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0593f776f469c20469e2f729fc7861a3.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0593f776f469c20469e2f729fc7861a3.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::get_logging_prefix — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::logging::get_logging_prefix — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0c012cb374addd90eb1f42eaec570650.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0c012cb374addd90eb1f42eaec570650.html index 387d40881e..f5ab6c0438 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0c012cb374addd90eb1f42eaec570650.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0c012cb374addd90eb1f42eaec570650.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::get_reportable_log_level — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::logging::get_reportable_log_level — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a56e110feaaba2c3fd44bd201fd21a76a.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a56e110feaaba2c3fd44bd201fd21a76a.html index 3a56bc5664..4b0a94006b 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a56e110feaaba2c3fd44bd201fd21a76a.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a56e110feaaba2c3fd44bd201fd21a76a.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::get_is_colored_output_on — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::logging::get_is_colored_output_on — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a7cb50492421ea9de4e3db895819df6f2.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a7cb50492421ea9de4e3db895819df6f2.html index b2e66394e1..a91529483a 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a7cb50492421ea9de4e3db895819df6f2.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a7cb50492421ea9de4e3db895819df6f2.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::set_reportable_log_level — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::logging::set_reportable_log_level — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ac46ac0901cb97e3ae6e93b45f24e90b8.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ac46ac0901cb97e3ae6e93b45f24e90b8.html index 21d4f925aa..c77f40ca56 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ac46ac0901cb97e3ae6e93b45f24e90b8.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ac46ac0901cb97e3ae6e93b45f24e90b8.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::log — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::logging::log — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ad2efd47b6c3689e58ccc595680579ae5.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ad2efd47b6c3689e58ccc595680579ae5.html index e95140d8a7..2e4ee3207a 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ad2efd47b6c3689e58ccc595680579ae5.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ad2efd47b6c3689e58ccc595680579ae5.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::set_is_colored_output_on — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::logging::set_is_colored_output_on — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1af8f3443813315af7901903d25dd495cc.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1af8f3443813315af7901903d25dd495cc.html index d16b24a657..4f46deb11c 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1af8f3443813315af7901903d25dd495cc.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1logging_1af8f3443813315af7901903d25dd495cc.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::logging::set_logging_prefix — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::logging::set_logging_prefix — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a226e3c83379d1012cde8578c1c86b16c.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a226e3c83379d1012cde8578c1c86b16c.html index e95e802158..b1060aebcb 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a226e3c83379d1012cde8578c1c86b16c.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a226e3c83379d1012cde8578c1c86b16c.html @@ -10,7 +10,7 @@ - Template Function torch_tensorrt::ptq::make_int8_cache_calibrator — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Template Function torch_tensorrt::ptq::make_int8_cache_calibrator — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a6186e305f47c1d94b6130ef6c7f7e178.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a6186e305f47c1d94b6130ef6c7f7e178.html index d6f90d9866..eb0341d3d7 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a6186e305f47c1d94b6130ef6c7f7e178.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a6186e305f47c1d94b6130ef6c7f7e178.html @@ -10,7 +10,7 @@ - Template Function torch_tensorrt::ptq::make_int8_calibrator — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Template Function torch_tensorrt::ptq::make_int8_calibrator — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a5b405fd3bf3c8fc2e2a54cbbab979797.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a5b405fd3bf3c8fc2e2a54cbbab979797.html index 88dada564e..c6135eceb9 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a5b405fd3bf3c8fc2e2a54cbbab979797.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a5b405fd3bf3c8fc2e2a54cbbab979797.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::torchscript::check_method_operator_support — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::torchscript::check_method_operator_support — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a6e19490a08fb1553c9dd347a5ae79db9.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a6e19490a08fb1553c9dd347a5ae79db9.html index 4abac617d0..56e9465fff 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a6e19490a08fb1553c9dd347a5ae79db9.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a6e19490a08fb1553c9dd347a5ae79db9.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::torchscript::compile — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::torchscript::compile — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a81f9783517335dda877d8cfcf38987c9.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a81f9783517335dda877d8cfcf38987c9.html index 8e79971d6e..93adf2ebde 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a81f9783517335dda877d8cfcf38987c9.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a81f9783517335dda877d8cfcf38987c9.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::torchscript::embed_engine_in_new_module — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::torchscript::embed_engine_in_new_module — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1ae8d56472106eeef37fbe51ff7f40c9b2.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1ae8d56472106eeef37fbe51ff7f40c9b2.html index bcec891235..8b1762199e 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1ae8d56472106eeef37fbe51ff7f40c9b2.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1ae8d56472106eeef37fbe51ff7f40c9b2.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::torchscript::convert_method_to_trt_engine — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::torchscript::convert_method_to_trt_engine — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1ac4ab8313ae72c2c899ea31548b528528.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1ac4ab8313ae72c2c899ea31548b528528.html index edcab427dd..6bbc7fed1f 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1ac4ab8313ae72c2c899ea31548b528528.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1ac4ab8313ae72c2c899ea31548b528528.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::get_build_info — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::get_build_info — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1ad1acd06eaeaffbbcf6e7ebf426891384.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1ad1acd06eaeaffbbcf6e7ebf426891384.html index f8e6ca9a23..5081ccb229 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1ad1acd06eaeaffbbcf6e7ebf426891384.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1ad1acd06eaeaffbbcf6e7ebf426891384.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::set_device — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::set_device — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/function_namespacetorch__tensorrt_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.html b/docs/_cpp_api/function_namespacetorch__tensorrt_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.html index a5978c2aa8..24f5d9a009 100644 --- a/docs/_cpp_api/function_namespacetorch__tensorrt_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.html +++ b/docs/_cpp_api/function_namespacetorch__tensorrt_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.html @@ -10,7 +10,7 @@ - Function torch_tensorrt::dump_build_info — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Function torch_tensorrt::dump_build_info — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/namespace_torch_tensorrt.html b/docs/_cpp_api/namespace_torch_tensorrt.html index 7250793cce..fee2067766 100644 --- a/docs/_cpp_api/namespace_torch_tensorrt.html +++ b/docs/_cpp_api/namespace_torch_tensorrt.html @@ -10,7 +10,7 @@ - Namespace torch_tensorrt — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Namespace torch_tensorrt — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/namespace_torch_tensorrt__logging.html b/docs/_cpp_api/namespace_torch_tensorrt__logging.html index 96433fdcc7..19d2ceabc7 100644 --- a/docs/_cpp_api/namespace_torch_tensorrt__logging.html +++ b/docs/_cpp_api/namespace_torch_tensorrt__logging.html @@ -10,7 +10,7 @@ - Namespace torch_tensorrt::logging — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Namespace torch_tensorrt::logging — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/namespace_torch_tensorrt__ptq.html b/docs/_cpp_api/namespace_torch_tensorrt__ptq.html index 21c223a6b9..ebc5c2961a 100644 --- a/docs/_cpp_api/namespace_torch_tensorrt__ptq.html +++ b/docs/_cpp_api/namespace_torch_tensorrt__ptq.html @@ -10,7 +10,7 @@ - Namespace torch_tensorrt::ptq — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Namespace torch_tensorrt::ptq — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/namespace_torch_tensorrt__torchscript.html b/docs/_cpp_api/namespace_torch_tensorrt__torchscript.html index 950883c6c8..8c0b9c967c 100644 --- a/docs/_cpp_api/namespace_torch_tensorrt__torchscript.html +++ b/docs/_cpp_api/namespace_torch_tensorrt__torchscript.html @@ -10,7 +10,7 @@ - Namespace torch_tensorrt::torchscript — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Namespace torch_tensorrt::torchscript — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.html b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.html index e229db6f41..4ca839627c 100644 --- a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.html +++ b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.html @@ -10,7 +10,7 @@ - Program Listing for File logging.h — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Program Listing for File logging.h — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.html b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.html index b2f2ea666e..996392e26c 100644 --- a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.html +++ b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.html @@ -10,7 +10,7 @@ - Program Listing for File macros.h — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Program Listing for File macros.h — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    @@ -412,7 +415,7 @@ #define STR(x) XSTR(x) #define TORCH_TENSORRT_MAJOR_VERSION 2 -#define TORCH_TENSORRT_MINOR_VERSION 0 +#define TORCH_TENSORRT_MINOR_VERSION 2 #define TORCH_TENSORRT_PATCH_VERSION 0 #define TORCH_TENSORRT_VERSION \ STR(TORCH_TENSORRT_MAJOR_VERSION) \ diff --git a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.html b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.html index 7d558374b6..2d70f7b088 100644 --- a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.html +++ b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.html @@ -10,7 +10,7 @@ - Program Listing for File ptq.h — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Program Listing for File ptq.h — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.html b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.html index f3e5e43f03..0d012b487b 100644 --- a/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.html +++ b/docs/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.html @@ -10,7 +10,7 @@ - Program Listing for File torch_tensorrt.h — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Program Listing for File torch_tensorrt.h — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    @@ -432,6 +435,7 @@ public: enum Value : int8_t { kLong, + kDouble, kFloat, kHalf, kChar, diff --git a/docs/_cpp_api/structtorch__tensorrt_1_1Device.html b/docs/_cpp_api/structtorch__tensorrt_1_1Device.html index 982416d9ac..badde58f21 100644 --- a/docs/_cpp_api/structtorch__tensorrt_1_1Device.html +++ b/docs/_cpp_api/structtorch__tensorrt_1_1Device.html @@ -10,7 +10,7 @@ - Struct Device — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Struct Device — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/structtorch__tensorrt_1_1GraphInputs.html b/docs/_cpp_api/structtorch__tensorrt_1_1GraphInputs.html index 5fe44040f9..560fe3d00d 100644 --- a/docs/_cpp_api/structtorch__tensorrt_1_1GraphInputs.html +++ b/docs/_cpp_api/structtorch__tensorrt_1_1GraphInputs.html @@ -10,7 +10,7 @@ - Struct GraphInputs — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Struct GraphInputs — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/structtorch__tensorrt_1_1Input.html b/docs/_cpp_api/structtorch__tensorrt_1_1Input.html index 7cc8fc5b04..f1bdbfbcb7 100644 --- a/docs/_cpp_api/structtorch__tensorrt_1_1Input.html +++ b/docs/_cpp_api/structtorch__tensorrt_1_1Input.html @@ -10,7 +10,7 @@ - Struct Input — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Struct Input — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.html b/docs/_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.html index 970aac6819..09a250a2ba 100644 --- a/docs/_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.html +++ b/docs/_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.html @@ -10,7 +10,7 @@ - Struct CompileSpec — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Struct CompileSpec — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/torch_tensort_cpp.html b/docs/_cpp_api/torch_tensort_cpp.html index e9f8b0b0da..1078204785 100644 --- a/docs/_cpp_api/torch_tensort_cpp.html +++ b/docs/_cpp_api/torch_tensort_cpp.html @@ -10,7 +10,7 @@ - Torch-TensorRT C++ API — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Torch-TensorRT C++ API — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/_cpp_api/unabridged_orphan.html b/docs/_cpp_api/unabridged_orphan.html index 9631362f6f..93ce5dfa53 100644 --- a/docs/_cpp_api/unabridged_orphan.html +++ b/docs/_cpp_api/unabridged_orphan.html @@ -10,7 +10,7 @@ - Full API — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Full API — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/_downloads/6a6052d9668b2cb8332d349d328e21c1/_rendered_examples_jupyter.zip b/docs/_downloads/6a6052d9668b2cb8332d349d328e21c1/_rendered_examples_jupyter.zip index 20d7ce98af..f8d27df258 100644 Binary files a/docs/_downloads/6a6052d9668b2cb8332d349d328e21c1/_rendered_examples_jupyter.zip and b/docs/_downloads/6a6052d9668b2cb8332d349d328e21c1/_rendered_examples_jupyter.zip differ diff --git a/docs/_downloads/798cda8f83bd9f5e2cc93f329a04332c/_rendered_examples_python.zip b/docs/_downloads/798cda8f83bd9f5e2cc93f329a04332c/_rendered_examples_python.zip index d01d40773f..7b7e91f289 100644 Binary files a/docs/_downloads/798cda8f83bd9f5e2cc93f329a04332c/_rendered_examples_python.zip and b/docs/_downloads/798cda8f83bd9f5e2cc93f329a04332c/_rendered_examples_python.zip differ diff --git a/docs/_modules/index.html b/docs/_modules/index.html index 5ab7471258..fee6094ca5 100644 --- a/docs/_modules/index.html +++ b/docs/_modules/index.html @@ -9,7 +9,7 @@ - Overview: module code — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Overview: module code — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/_Device.html b/docs/_modules/torch_tensorrt/_Device.html index 90fdc83d59..e1ed0b7edc 100644 --- a/docs/_modules/torch_tensorrt/_Device.html +++ b/docs/_modules/torch_tensorrt/_Device.html @@ -9,7 +9,7 @@ - torch_tensorrt._Device — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt._Device — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/_Input.html b/docs/_modules/torch_tensorrt/_Input.html index 5b14baa96a..33aeb1d897 100644 --- a/docs/_modules/torch_tensorrt/_Input.html +++ b/docs/_modules/torch_tensorrt/_Input.html @@ -9,7 +9,7 @@ - torch_tensorrt._Input — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt._Input — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    @@ -429,6 +432,7 @@

    Source code for torch_tensorrt._Input

         low_tensor_domain_incl: float = 0.0
         high_tensor_domain_excl: float = low_tensor_domain_incl + DOMAIN_OFFSET
         torch_dtype: torch.dtype = torch.float32
    +    torch_tensor: torch.Tensor = None
     
     
    [docs] def __init__(self, *args: Any, **kwargs: Any) -> None: """__init__ Method for torch_tensorrt.Input @@ -552,7 +556,15 @@

    Source code for torch_tensorrt._Input

             else:
                 domain = None
     
    -        self.tensor_domain = Input._parse_tensor_domain(domain)
    + self.tensor_domain = Input._parse_tensor_domain(domain) + + if "torch_tensor" in kwargs: + self.torch_tensor = kwargs["torch_tensor"] + else: + if self.shape_mode == Input._ShapeMode.DYNAMIC: + self.torch_tensor = self.example_tensor("opt_shape") + else: + self.torch_tensor = self.example_tensor()
    def __str__(self) -> str: if self.shape_mode == Input._ShapeMode.STATIC: @@ -603,6 +615,8 @@

    Source code for torch_tensorrt._Input

                     return _enums.dtype.half
                 elif dtype == torch.float:
                     return _enums.dtype.float
    +            elif dtype == torch.float64:
    +                return _enums.dtype.double
                 elif dtype == torch.bool:
                     return _enums.dtype.bool
                 else:
    @@ -632,6 +646,8 @@ 

    Source code for torch_tensorrt._Input

                 return torch.float
             elif dtype == _enums.dtype.bool:
                 return torch.bool
    +        elif dtype == _enums.dtype.double:
    +            return torch.float64
             else:
                 # Default torch_dtype used in FX path
                 return torch.float32
    @@ -648,7 +664,7 @@ 

    Source code for torch_tensorrt._Input

                     return _enums.TensorFormat.channels_last
                 else:
                     raise ValueError(
    -                    "Provided an unsupported tensor format (support: NHCW/contiguous_format, NHWC/channel_last)"
    +                    "Provided an unsupported tensor format (support: NCHW/contiguous_format, NHWC/channel_last)"
                     )
     
             elif isinstance(format, _enums.TensorFormat):
    @@ -737,7 +753,7 @@ 

    Source code for torch_tensorrt._Input

                 )
                 else torch.channels_last
             )
    -        return cls(shape=t.shape, dtype=t.dtype, format=frmt)
    + return cls(shape=t.shape, dtype=t.dtype, format=frmt, torch_tensor=t)
    [docs] @classmethod def from_tensors( diff --git a/docs/_modules/torch_tensorrt/_compile.html b/docs/_modules/torch_tensorrt/_compile.html index c739cfbdf3..362c9be621 100644 --- a/docs/_modules/torch_tensorrt/_compile.html +++ b/docs/_modules/torch_tensorrt/_compile.html @@ -9,7 +9,7 @@ - torch_tensorrt._compile — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt._compile — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    @@ -597,23 +600,24 @@

    Source code for torch_tensorrt._compile

             )
             return compiled_fx_module
         elif target_ir == _IRType.dynamo:
    +        # Prepare torch and torchtrt inputs
             import collections.abc
     
    -        from torch_tensorrt import Device
    -        from torch_tensorrt.dynamo.utils import prepare_inputs, to_torch_device
    +        from torch_tensorrt.dynamo.utils import prepare_inputs
     
    -        if not isinstance(inputs, collections.abc.Sequence):
    -            inputs = [inputs]
    -        device = kwargs.get("device", Device._current_device())
    -        torchtrt_inputs, torch_inputs = prepare_inputs(inputs, to_torch_device(device))
    -        module = torch_tensorrt.dynamo.trace(module, torch_inputs, **kwargs)
    -        compiled_aten_module: torch.fx.GraphModule = dynamo_compile(
    -            module,
    -            inputs=input_list,
    +        if not isinstance(input_list, collections.abc.Sequence):
    +            input_list = [input_list]
    +
    +        # Export the module
    +        torchtrt_inputs = prepare_inputs(input_list)
    +        exp_program = torch_tensorrt.dynamo.trace(module, torchtrt_inputs, **kwargs)
    +        trt_graph_module = dynamo_compile(
    +            exp_program,
    +            inputs=torchtrt_inputs,
                 enabled_precisions=enabled_precisions_set,
                 **kwargs,
             )
    -        return compiled_aten_module
    +        return trt_graph_module
         elif target_ir == _IRType.torch_compile:
             return torch_compile(
                 module, enabled_precisions=enabled_precisions_set, **kwargs
    diff --git a/docs/_modules/torch_tensorrt/_utils.html b/docs/_modules/torch_tensorrt/_utils.html
    index c20fb36725..d37ff3c688 100644
    --- a/docs/_modules/torch_tensorrt/_utils.html
    +++ b/docs/_modules/torch_tensorrt/_utils.html
    @@ -9,7 +9,7 @@
       
       
       
    -  torch_tensorrt._utils — Torch-TensorRT v2.0.0.dev0+1fec519 documentation
    +  torch_tensorrt._utils — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation
       
     
       
    @@ -222,7 +222,7 @@
                   
                   
                     
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/fx/fx2trt.html b/docs/_modules/torch_tensorrt/fx/fx2trt.html index 4a6b10fd03..43afa4cd11 100644 --- a/docs/_modules/torch_tensorrt/fx/fx2trt.html +++ b/docs/_modules/torch_tensorrt/fx/fx2trt.html @@ -9,7 +9,7 @@ - torch_tensorrt.fx.fx2trt — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.fx.fx2trt — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/fx/input_tensor_spec.html b/docs/_modules/torch_tensorrt/fx/input_tensor_spec.html index 7edbbc5d1b..673a4de080 100644 --- a/docs/_modules/torch_tensorrt/fx/input_tensor_spec.html +++ b/docs/_modules/torch_tensorrt/fx/input_tensor_spec.html @@ -9,7 +9,7 @@ - torch_tensorrt.fx.input_tensor_spec — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.fx.input_tensor_spec — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/fx/lower.html b/docs/_modules/torch_tensorrt/fx/lower.html index 602ed7088b..fdea1c2e7a 100644 --- a/docs/_modules/torch_tensorrt/fx/lower.html +++ b/docs/_modules/torch_tensorrt/fx/lower.html @@ -9,7 +9,7 @@ - torch_tensorrt.fx.lower — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.fx.lower — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/fx/trt_module.html b/docs/_modules/torch_tensorrt/fx/trt_module.html index 9d65f97f81..93a6719c1a 100644 --- a/docs/_modules/torch_tensorrt/fx/trt_module.html +++ b/docs/_modules/torch_tensorrt/fx/trt_module.html @@ -9,7 +9,7 @@ - torch_tensorrt.fx.trt_module — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.fx.trt_module — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/logging.html b/docs/_modules/torch_tensorrt/logging.html index 224ad52aa8..d4b0a2cb6b 100644 --- a/docs/_modules/torch_tensorrt/logging.html +++ b/docs/_modules/torch_tensorrt/logging.html @@ -9,7 +9,7 @@ - torch_tensorrt.logging — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.logging — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/ptq.html b/docs/_modules/torch_tensorrt/ptq.html index 1cd417f54c..945d2fc917 100644 --- a/docs/_modules/torch_tensorrt/ptq.html +++ b/docs/_modules/torch_tensorrt/ptq.html @@ -9,7 +9,7 @@ - torch_tensorrt.ptq — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.ptq — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/ts/_compile_spec.html b/docs/_modules/torch_tensorrt/ts/_compile_spec.html index ab8429676e..d0cc766652 100644 --- a/docs/_modules/torch_tensorrt/ts/_compile_spec.html +++ b/docs/_modules/torch_tensorrt/ts/_compile_spec.html @@ -9,7 +9,7 @@ - torch_tensorrt.ts._compile_spec — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.ts._compile_spec — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_modules/torch_tensorrt/ts/_compiler.html b/docs/_modules/torch_tensorrt/ts/_compiler.html index bdceb39310..f3396794c2 100644 --- a/docs/_modules/torch_tensorrt/ts/_compiler.html +++ b/docs/_modules/torch_tensorrt/ts/_compiler.html @@ -9,7 +9,7 @@ - torch_tensorrt.ts._compiler — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.ts._compiler — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/_sources/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.rst.txt b/docs/_sources/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.rst.txt index b2875cba3c..e8c23214e9 100644 --- a/docs/_sources/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.rst.txt +++ b/docs/_sources/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.rst.txt @@ -36,7 +36,7 @@ Program Listing for File macros.h #define STR(x) XSTR(x) #define TORCH_TENSORRT_MAJOR_VERSION 2 - #define TORCH_TENSORRT_MINOR_VERSION 0 + #define TORCH_TENSORRT_MINOR_VERSION 2 #define TORCH_TENSORRT_PATCH_VERSION 0 #define TORCH_TENSORRT_VERSION \ STR(TORCH_TENSORRT_MAJOR_VERSION) \ diff --git a/docs/_sources/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst.txt b/docs/_sources/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst.txt index 07718fc152..67848a40a0 100644 --- a/docs/_sources/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst.txt +++ b/docs/_sources/_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst.txt @@ -56,6 +56,7 @@ Program Listing for File torch_tensorrt.h public: enum Value : int8_t { kLong, + kDouble, kFloat, kHalf, kChar, diff --git a/docs/_sources/contributors/fx_converters.rst.txt b/docs/_sources/contributors/fx_converters.rst.txt new file mode 100644 index 0000000000..75ee1c6341 --- /dev/null +++ b/docs/_sources/contributors/fx_converters.rst.txt @@ -0,0 +1,211 @@ +.. _dynamo_conversion: + +Dynamo Converters +================== +The dynamo converter library in Torch-TensorRT is located in ``TensorRT/py/torch_tensorrt/dynamo/conversion``. + + + +Steps +================== + +Operation Set +------------------- +The converters in dynamo are produced by ``aten_trace`` and falls under ``aten_ops_converters`` ( FX earlier had ``acc_ops_converters``, ``aten_ops_converters`` or ``nn_ops_converters`` depending on the trace through which it was produced). The converters are registered using ``dynamo_tensorrt_converter`` for dynamo. The function decorated +has the arguments - ``network, target, args, kwargs, name``, which is common across all the operators schema. +These functions are mapped in the ``aten`` converter registry dictionary (at present a compilation of FX and dynamo converters, FX will be deprecated soon), with key as the function target name. + + * aten_trace is produced by ``torch_tensorrt.dynamo.trace(..)`` for the export path and ``torch_tensorrt.compile(ir=dynamo)`` for the compile path. + The export path makes use of ``aten_tracer`` whereas the alternate trace in compile is produced by the AOT Autograd library. + Both these simplify the torch operators to reduced set of Aten operations. + + +As mentioned above, if you would like to add a new converter, its implementation will be included in ``TensorRT/py/torch_tensorrt/dynamo/conversion/impl`` +Although there is a corresponding implementation of the converters included in the common implementation library present in ``TensorRT/py/torch_tensorrt/fx/impl`` for FX converters, this documentation focuses on the implementation of the ``aten_ops`` converters in dynamo. + + +Converter implementation +------------------------ +In this section, we illustrate the steps to be implemented for writing a converter. We divide them according to activation, operator, lowering pass implementation or evaluator. +Each of them is detailed with the help of an example + + * Registration + + The converter needs to be registered with the appropriate op code in the ``dynamo_tensorrt_converter``. + + * Activation type + + Example: ``leaky_relu`` + + + * aten_ops_converters: Dynamo_converters + + Define in ``py/torch_tensorrt/dynamo/conversion/aten_ops_converters``. One needs to register the opcode generated in the trace with ``dynamo_tensorrt_converter`` decorator. Op code to be used for the registration or the converter registry key in this case is ``torch.ops.aten.leaky_relu.default`` + + .. code-block:: python + + @dynamo_tensorrt_converter(torch.ops.aten.leaky_relu.default) + def aten_ops_leaky_relu( + network: TRTNetwork, + target: Target, + args: Tuple[Argument, ...], + kwargs: Dict[str, Argument], + name: str, + ) -> Union[TRTTensor, Sequence[TRTTensor]]: + return activation.leaky_relu(network, target, SourceIR.ATEN, name, args[0], args[1]) + + The ``tensorrt_converter`` (used for FX registration) and ``dynamo_tensorrt_converter`` are similar decorator functions with some differences. + + #. Both register the converters in the registeries (python dictionaries) - ``CONVERTERS`` and ``DYNAMO_CONVERTERS`` respectively. These are two dictioneries which are concatenated to form the overall converter registry + #. The dictionary is keyed on the ``OpOverLoad`` which is mentioned in more detail below with examples + #. Both return the decorated converter implementation + #. The ``CONVERTERS`` directly registers the decorated ``converter_implementation`` function, while ``DYNAMO_CONVERTERS`` has additionational arguments and registers the ``ConverterSupport`` object + #. The additional arguments are: + + .. code-block:: python + def dynamo_tensorrt_converter( + key: Target, + enabled: bool = True, + capability_validator: Optional[Callable[[Node], bool]] = None, + priority: ConverterPriority = ConverterPriority.STANDARD, + ) -> Callable[[Any], Union[TRTTensor, Sequence[TRTTensor]]]: + + #. key: Node target for which the converter is implemented for (for example, torch.ops.aten.leaky_relu.Tensor) + #. enabled: Whether the converter should be enabled/cached or not + #. capability_validator: Function which evaluates whether a node is valid for conversion by the decorated converter. It defaults to None, implying the capability_validator function is always true. This means all nodes of "key" kind can be supported by this converter by default. See ``embedding`` example for more details + #. priority: Converter's level of priority relative to other converters with the same target + + #. The ``ConverterSupport`` is a compilation of ``converter_implementation`` and ``capability_validator``. + + + The function decorated by ``tensorrt_converter`` and ``dynamo_tensorrt_converter`` has the following arguments which are automatically generated by the trace functions mentioned above. + + #. network : Node in the form of ``call_module`` or ``call_function`` having the target as the key + #. target: Target key in the ``call_module`` or ``call_function`` above. eg: ``torch.ops.aten_.leaky_relu.default``. Note that ``torch.ops.aten._leaky_relu`` is the ``OpOverloadPacket`` while ``torch.ops.aten_.leaky_relu.default`` is ``OpOverload``. + #. args: The arguments passed in the ``call_module`` or ``call_function`` above + #. kwargs: The kwargs passed in the ``call_module`` or ``call_function`` above + #. name: String containing the name of the target + + As a user writing new converters, one just needs to take care that the approriate arguments are extracted from the trace generated to the implementation function in the implementation lib function ``activation.leaky_relu`` (which we will discuss below in detail). + + * Operation type + + Example: ``fmod`` + + It follows the same steps as the above converter. In this case the opcode is ``torch.ops.aten.fmod.Scalar`` or ``torch.ops.aten.fmod.Tensor``. + Hence both the opcodes are registered in ``py/torch_tensorrt/dynamo/conversion/aten_ops_converters``. + Note that ``torch.ops.aten.fmod`` is the ``OpOverLoadPacket`` while the registry is keyed on ``torch.ops.aten.fmod.Scalar`` or ``torch.ops.aten.fmod.Tensor``, which is ``OpOverLoad`` + + Example: ``embedding`` + + It follows the same steps as the above converter. In this case the opcode is ``torch.ops.aten.embedding.default``. + There are some converters which have special cases to be accounted for. In those cases, one should use ``capability_validators`` to register the converter using ``@dynamo_tensorrt_converter`` + We illustrate this through ``torch.ops.aten.embedding.default``. It has parameters - ``scale_grad_by_freq`` and ``sparse`` which are not currently supported by the implementation. + In such cases we can write validator ``embedding_param_validator`` which implements that given those paramters the converter is not supported and register the converter by + + .. code-block:: python + @dynamo_tensorrt_converter( + torch.ops.aten.embedding.default, capability_validator=embedding_param_validator + ) + + So if there is a new converter in which certain special cases are not to be supported then they can be specified in the ``capability_validator``. + + * Evaluator type + + Example: ``operator.getitem`` + + Evaluators are categorized as so since they do not make any modification to the graph. This is implemented in ``py/torch_tensorrt/dynamo/conversion/op_evaluators.py``, with the corresponding ``capbility_validator``. + The opcode is ``operator.getitem``. + + + * Implementation Library + + The dynamo converters would be located in ``py/torch_tensorrt/dynamo/conversion/impl`` + + * Activation + + Example: ``leaky_relu`` + + The implementation is to be placed in present in ``py/torch_tensorrt/dynamo/conversion/impl/activation.py``. This is where all the activation functions are defined and implemented. + + .. code-block:: python + + def leaky_relu( + network: TRTNetwork, + target: Target, + source_ir: Optional[SourceIR], + name: str, + input_val: TRTTensor, + alpha: Optional[Any], + ): + #implementation + + The implementation function has the following arguments. + + #. network : ``network`` passed from the decorated function registration + #. target: ``target`` passed from the decorated function registration + #. source_ir: Enum attribute. ``SourceIR`` enum is defined in ``py/torch_tensorrt/dynamo/conversion/impl/converter_utils`` + #. name: ``name`` passed from the decorated function registration + #. input_val: Approriate arguments extracted from the decorated function registration from args or kwargs + #. alpha: Approriate arguments extracted from the decorated function registration from args or kwargs. If not None, it will set the alpha attribute of the created TensorRT activation layer eg: Used in leaky_relu, elu, hardtanh + #. beta: Approriate arguments extracted from the decorated function registration from args or kwargs. If not None, it will set the beta attribute of the created TensorRT activation layer eg: Used in hardtanh + #. dyn_range_fn: A optional function which takes the dynamic range of a TensorRT Tensor and returns the output dynamic range + + The implementation functions call the ``convert_activation`` function in ``py/torch_tensorrt/dynamo/conversion/impl/activation.py``. This function will add the approriate activation layer via ``network.add_activation``. + + * Operator + + The implementation is to be placed in ``py/torch_tensorrt/dynamo/conversion/impl/elementwise/ops.py`` for dynamo. This is where all the elementwise functions are defined and implemented. + For a new operator, one should identify the category to which it belongs. Following are some examples + + #. Elementwise operators like ``fmod`` is present in ``py/torch_tensorrt/dynamo/conversion/impl/elementwise``. The ``py/torch_tensorrt/dynamo/conversion/impl/elementwise/base`` contains base functions for elementwise operator. + #. Unary operators like ``sqrt`` will be present in ``py/torch_tensorrt/dynamo/conversion/impl/unary``. The ``py/torch_tensorrt/dynamo/conversion/impl/unary/base`` contains base functions for unary operator. + #. Normalization operators like ``softmax``, ``layer_norm``, ``batch_norm`` will be present in ``py/torch_tensorrt/dynamo/conversion/impl/normalization``. Since there are no base operations common to all, there is no base file. But one can choose to implement a base file, if there are common functions across all normalization operations + #. Individual operators like ``slice``, ``select``, ``where``, ``embedding`` will be present in ``py/torch_tensorrt/dynamo/conversion/impl/*.py``. They will have individual operator implementation with the same API structure as above but with different individual arguments + + Please note that the above operators would have common functions to be implemented which should be placed in + ``py/torch_tensorrt/dynamo/conversion/impl/converter_utils.py`` + + + * Lowering type + + There are some converters which can be decomposed into suboperations and need not have seperate converter registration. + Such converters can be implemented via ``lowering passes`` + + Example: ``addmm`` + + The decompositions are registered via ``register_decomposition`` in ``py/torch_tensorrt/dynamo/backend/lowering/_decompositions.py`` + We define ``addmm_replacement`` and replace it with the torch ops, which will have their corresponding converters called. + + .. code-block:: python + + @register_decomposition(torch.ops.aten.addmm, registry=DECOMPOSITIONS) + def addmm_replacement( + input_: torch.Tensor, mat1: torch.Tensor, mat2: torch.Tensor, *, beta=1, alpha=1 + ) -> torch.Tensor: + return torch.add( + torch.mul(input_, beta), torch.mul(torch.matmul(mat1, mat2), alpha) + ) + + Note that there are some pre-existing dynamo decompositions in torch directory, in which case they should be used, + In that case please enable the decompositions in ``py/torch_tensorrt/dynamo/lowering/_decomposition_groups.py`` in ``torch_enabled_decompositions``. + Similarly you can choose to disable any in ``torch_disabled_decompositions``. Please note that the ones already defined in the lowering will take precedence over torch lowering ops. + + + + +Tests +----- + +* Dynamo testing: + + Dynamo tests are present for the lowering ops in ``tests/py/dynamo/lowering/test_decompositions.py``. The above converters will soon be ported to dynamo tests + + #. Compare the results for ``fx.symbolic_trace`` and ``torch_tensorrt.dynamo.compile``. + #. Test for the ``expected_op`` and the ``unexpected_op``. + + #. ``expected_op``: Operations the operations are lowered to. eg: ``mul`` and ``add`` for ``addmm`` + #. Note that specify that ``disable_passes= True`` for cases where you would not want lowering passes (which should be the default when testing converters) + #. ``unexpected_op``: Original operation. eg: ``addmm`` for ``addmm`` + +The tests should fail if any of the above two conditions fail diff --git a/docs/_sources/contributors/writing_dynamo_aten_lowering_passes.rst.txt b/docs/_sources/contributors/writing_dynamo_aten_lowering_passes.rst.txt new file mode 100644 index 0000000000..4c29bc9b75 --- /dev/null +++ b/docs/_sources/contributors/writing_dynamo_aten_lowering_passes.rst.txt @@ -0,0 +1,109 @@ +.. _writing_dynamo_aten_lowering_passes: + +Writing Dynamo ATen Lowering Passes +=================== + +Basics of a Lowering Pass +------------ + +ATen lowering passes are Python functions which take as input a graph of ATen operators, apply some desired modification such as operator coalescing/fusion, operator replacement, subgraph rewriting, custom operator insertion, or other operation on a `torch.fx.GraphModule`, then return the modified graph to the caller. These lowering passes generally modify the graph in-place and return the same input object. + +Lowering Pass Requirements +------------ + +An ATen lowering pass function in Torch-TRT must satisfy two requirements: +- The function must take as input a `torch.fx.GraphModule` and a sequence of torch Tensors, `Sequence[torch.Tensor]`, and return the lowered `torch.fx.GraphModule` +- The function must leave the graph in a valid and invoke-able state, including performing any necessary linting and recompilation + +See this link for information on `Graph Manipulations `_ in FX. See below for an example of a lowering pass which repairs graphs that have inputs which are also outputs, a disallowed configuration for TRT Engines. + +Example Lowering Pass +------------ + +.. code-block:: python + + def repair_input_as_output(gm: torch.fx.GraphModule, sample_inputs: Sequence[torch.Tensor]) -> torch.fx.GraphModule: + """Repair scenarios where inputs are also outputs of the graph + + TRT does not allow such cases, so we insert a clone (identity) layer + """ + modified_graph = False + + # Extract graph placeholder Tensors + placeholders = [ + node + for node in gm.graph.nodes + if ( + node.op == "placeholder" + and isinstance(node.type, type) + and issubclass(node.type, torch.Tensor) + ) + ] + + for placeholder in placeholders: + # If any placeholder has any users which are direct graph outputs + if len(placeholder.users) >= 1 and any( + user.op == "output" for user in placeholder.users + ): + modified_graph = True + + # Get direct graph outputs which are direct uses of placeholders + direct_outputs = [user for user in placeholder.users if user.op == "output"] + + # Insert clone node for placeholder to ensure + # placeholder is not a direct output + with gm.graph.inserting_after(placeholder): + cloned_placeholder = gm.graph.call_function( + torch.ops.aten.clone.default, + args=(placeholder,), + ) + + # Replace placeholder as output with cloned version + for output in direct_outputs: + output.replace_input_with(placeholder, cloned_placeholder) + + # If the graph was modified, clean up the graph and ensure it is up-to-date + if modified_graph: + gm.graph.eliminate_dead_code() + gm.graph.lint() + gm.recompile() + logger.debug(f"Graph after repair_input_as_output:\n{gm.graph}") + + return gm + + +Registering Lowering Passes +---------------------- + +Lowering passes are currently registered in `py/torch_tensorrt/dynamo/lowering/passes/__init__.py`, using the `torch.fx.passes.pass_manager.PassManager` utility to assemble the list of passes in a desired order. New passes added directly to that list will be applied to graphs in the Torch-TensorRT `torch.compile` backend. Currently, we offer an ATen lowering pass registration decorator for convenience, which can be invoked either directly, or with the optional `index` keyword argument which controls where in the pass list the lowering pass will be inserted. + +For instance, to insert the pass at the default location (end of the list), the following code can be used: + +.. code-block:: python + + @_aten_lowering_pass + def my_custom_pass(gm: torch.fx.GraphModule, sample_inputs: Sequence[torch.Tensor]) -> torch.fx.GraphModule: + ... + +Alternatively, to insert the pass at a custom index (such as the front of the list) in the passlist, the following code can be used: + +.. code-block:: python + + @_aten_lowering_pass(index=0) + def my_custom_pass(gm: torch.fx.GraphModule, sample_inputs: Sequence[torch.Tensor]) -> torch.fx.GraphModule: + ... + +There are also provided utilities in `torch_tensorrt.dynamo.lowering.passes` for displaying the currently-available lowering pass list, applying those passes to an arbitrary `torch.fx.GraphModule`, and removing the lowering pass at a specific index. + +.. code-block:: python + + # Print all lowering passes in the list + print(dump_lowering_passes()) + + # Apply lowering passes to a GraphModule + apply_lowering_passes(graph_module, sample_inputs) + + # Remove the lowering pass at index 1 + _remove_lowering_pass(index=1) + +**Note:** The above APIs are subject to change, as the lowering pass system evolves. diff --git a/docs/_sources/index.rst.txt b/docs/_sources/index.rst.txt index eee62bc2f7..9e98c7a63d 100644 --- a/docs/_sources/index.rst.txt +++ b/docs/_sources/index.rst.txt @@ -42,6 +42,8 @@ User Guide * :ref:`getting_started_with_fx` * :ref:`ptq` * :ref:`runtime` +* :ref:`saving_models` +* :ref:`dynamic_shapes` * :ref:`use_from_pytorch` * :ref:`using_dla` @@ -54,6 +56,8 @@ User Guide user_guide/getting_started_with_fx_path user_guide/ptq user_guide/runtime + user_guide/saving_models + user_guide/dynamic_shapes user_guide/use_from_pytorch user_guide/using_dla @@ -128,6 +132,7 @@ Contributor Documentation -------------------------------- * :ref:`system_overview` * :ref:`writing_converters` +* :ref:`writing_dynamo_aten_lowering_passes` * :ref:`useful_links` .. toctree:: @@ -137,6 +142,7 @@ Contributor Documentation contributors/system_overview contributors/writing_converters + contributors/writing_dynamo_aten_lowering_passes contributors/useful_links Indices diff --git a/docs/_sources/user_guide/dynamic_shapes.rst.txt b/docs/_sources/user_guide/dynamic_shapes.rst.txt new file mode 100644 index 0000000000..28320956c4 --- /dev/null +++ b/docs/_sources/user_guide/dynamic_shapes.rst.txt @@ -0,0 +1,218 @@ +.. _runtime: + +Dynamic shapes with Torch-TensorRT +==================================== + +By default, you can run a pytorch model with varied input shapes and the output shapes are determined eagerly. +However, Torch-TensorRT is an AOT compiler which requires some prior information about the input shapes to compile and optimize the model. +In the case of dynamic input shapes, we must provide the (min_shape, opt_shape, max_shape) arguments so that the model can be optimized for +these range of input shapes. An example usage of static and dynamic shapes is as follows. + +NOTE: The following code uses dynamo IR. Incase of Torchscript IR, please swap out ``ir=dynamo`` with ``ir=ts`` and the behavior is exactly the same. + +.. code-block:: python + + import torch + import torch_tensorrt + + model = MyModel().eval().cuda() + # Compile with static shapes + inputs = torch_tensorrt.Input(shape=[1, 3, 224, 224], dtype=torch.float32) + # or compile with dynamic shapes + inputs = torch_tensorrt.Input(min_shape=[1, 3, 224, 224], + opt_shape=[4, 3, 224, 224], + max_shape=[8, 3, 224, 224], + dtype=torch.float32) + trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) + +Under the hood +-------------- + +There are two phases of compilation when we use ``torch_tensorrt.compile`` API with ``ir=dynamo`` (default). + +- aten_tracer.trace (which uses torch.export to trace the graph with the given inputs) + +In the tracing phase, we use torch.export along with the constraints. In the case of +dynamic shaped inputs, the range can be provided to the tracing via constraints. Please +refer to this `docstring `_ +for detailed information on how to set constraints. In short, we create new inputs for +torch.export tracing and provide constraints on the min and max values(provided by the user), a particular dimension can take. +Please take a look at ``aten_tracer.py`` file to understand how this works under the hood. + +- dynamo.compile (which compiles a torch.fx.GraphModule object using TensorRT) + +In the conversion to TensorRT, we use the user provided dynamic shape inputs. +We perform shape analysis using dummy inputs (across min, opt and max shapes) and store the +intermediate output shapes which can be used in case the graph has a mix of Pytorch +and TensorRT submodules. + +Custom Constraints +------------------ + +Given an input ``x = torch_tensorrt.Input(min_shape, opt_shape, max_shape, dtype)``, +Torch-TensorRT automatically sets the constraints during ``torch.export`` tracing as follows + +.. code-block:: python + + for dim in constraint_dims: + if min_shape[dim] > 1: + constraints.append(min_shape[dim] <= dynamic_dim(trace_input, dim)) + if max_shape[dim] > 1: + constraints.append(dynamic_dim(trace_input, dim) <= max_shape[dim]) + +Sometimes, we might need to set additional constraints and Torchdynamo errors out if we don't specify them. +For example, in the case of BERT model compilation, there are two inputs and a constraint has to be set involving the sequence length size of these two inputs. + +.. code-block:: python + + constraints.append(dynamic_dim(trace_inputs[0], 0) == dynamic_dim(trace_inputs[1], 0)) + + +If you have to provide any custom constraints to your model, the overall workflow for model compilation using ``ir=dynamo`` would involve a few steps. + +.. code-block:: python + + import torch + import torch_tensorrt + from torch_tensorrt.dynamo.lowering import apply_lowering_passes, get_decompositions + # Assume the model has two inputs + model = MyModel() + torch_input_1 = torch.randn((1, 14), dtype=torch.int32).cuda() + torch_input_2 = torch.randn((1, 14), dtype=torch.int32).cuda() + + dynamic_inputs = [torch_tensorrt.Input(min_shape=[1, 14], + opt_shape=[4, 14], + max_shape=[8, 14], + dtype=torch.int32), + torch_tensorrt.Input(min_shape=[1, 14], + opt_shape=[4, 14], + max_shape=[8, 14], + dtype=torch.int32)] + + # Export the model with additional constraints + constraints = [] + # The following constraints are automatically added by Torch-TensorRT in the + # general case when you call torch_tensorrt.compile directly on MyModel() + constraints.append(dynamic_dim(torch_input_1, 0) < 8) + constraints.append(dynamic_dim(torch_input_2, 0) < 8) + # This is an additional constraint as instructed by Torchdynamo + constraints.append(dynamic_dim(torch_input_1, 0) == dynamic_dim(torch_input_2, 0)) + with unittest.mock.patch( + "torch._export.DECOMP_TABLE", get_decompositions(experimental_decompositions) + ): + graph_module = export( + model, (torch_input_1, torch_input_2), constraints=constraints + ).module() + + # Use the dynamo.compile API + trt_mod = torch_tensorrt.dynamo.compile(graph_module, inputs=dynamic_inputs, **compile_spec) + +Limitations +----------- + +If there are operations in the graph that use the dynamic dimension of the input, Pytorch +introduces ``torch.ops.aten.sym_size.int`` ops in the graph. Currently, we cannot handle these operators and +the compilation results in undefined behavior. We plan to add support for these operators and implement +robust support for shape tensors in the next release. Here is an example of the limitation described above + +.. code-block:: python + + import torch + import torch_tensorrt + + class MyModule(torch.nn.Module): + def __init__(self): + super().__init__() + self.avgpool = torch.nn.AdaptiveAvgPool2d((1, 1)) + + def forward(self, x): + x = self.avgpool(x) + out = torch.flatten(x, 1) + return out + + model = MyModel().eval().cuda() + # Compile with dynamic shapes + inputs = torch_tensorrt.Input(min_shape=(1, 512, 1, 1), + opt_shape=(4, 512, 1, 1), + max_shape=(8, 512, 1, 1), + dtype=torch.float32) + trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) + + +The traced graph of `MyModule()` looks as follows + +.. code-block:: python + + Post export graph: graph(): + %arg0_1 : [num_users=2] = placeholder[target=arg0_1] + %mean : [num_users=1] = call_function[target=torch.ops.aten.mean.dim](args = (%arg0_1, [-1, -2], True), kwargs = {}) + %sym_size : [num_users=1] = call_function[target=torch.ops.aten.sym_size.int](args = (%arg0_1, 0), kwargs = {}) + %view : [num_users=1] = call_function[target=torch.ops.aten.view.default](args = (%mean, [%sym_size, 512]), kwargs = {}) + return (view,) + + +Here the ``%sym_size`` node captures the dynamic batch and uses it in the ``aten.view`` layer. This requires shape tensors support +which would be a part of our next release. + +Workaround (BERT static compilation example) +------------------------------------------ + +In the case where you encounter the issues mentioned in the **Limitations** section, +you can compile the model (static mode) with max input size that can be provided. In the cases of smaller inputs, +we can pad them accordingly. This is only a workaround until we address the limitations. + +.. code-block:: python + + import torch + import torch_tensorrt + from transformers.utils.fx import symbolic_trace as transformers_trace + + model = BertModel.from_pretrained("bert-base-uncased").cuda().eval() + + # Input sequence length is 20. + input1 = torch.randint(0, 5, (1, 20), dtype=torch.int32).to("cuda") + input2 = torch.randint(0, 5, (1, 20), dtype=torch.int32).to("cuda") + + model = transformers_trace(model, input_names=["input_ids", "attention_mask"]).eval().cuda() + trt_mod = torch_tensorrt.compile(model, inputs=[input1, input2], **compile_spec) + model_outputs = model(input, input2) + + # If you have a sequence of length 14, pad 6 zero tokens and run inference + # or recompile for sequence length of 14. + input1 = torch.randint(0, 5, (1, 14), dtype=torch.int32).to("cuda") + input2 = torch.randint(0, 5, (1, 14), dtype=torch.int32).to("cuda") + trt_mod = torch_tensorrt.compile(model, inputs=[input1, input2], **compile_spec) + model_outputs = model(input, input2) + + +Dynamic shapes with ir=torch_compile +------------------------------------ + +``torch_tensorrt.compile(model, inputs, ir="torch_compile")`` returns a torch.compile boxed function with the backend +configured to Tensorrt. In the case of ``ir=torch_compile``, users have to recompile for different input shapes. +In the future, we plan to explore the option of compiling with dynamic shapes in the first execution of the model. + +.. code-block:: python + + import torch + import torch_tensorrt + + model = MyModel().eval().cuda() + inputs = torch.randn((1, 3, 224, 224), dtype=float32) + trt_gm = torch_tensorrt.compile(model, ir="torch_compile", inputs) + # Compilation happens when you call the model + trt_gm(inputs) + + # Recompilation happens with modified batch size + inputs_bs2 = torch.randn((2, 3, 224, 224), dtype=torch.float32) + trt_gm = torch_tensorrt.compile(model, ir="torch_compile", inputs_bs2) + + + + + + + + + + diff --git a/docs/_sources/user_guide/saving_models.rst.txt b/docs/_sources/user_guide/saving_models.rst.txt new file mode 100644 index 0000000000..46fadcb905 --- /dev/null +++ b/docs/_sources/user_guide/saving_models.rst.txt @@ -0,0 +1,77 @@ +.. _runtime: + +Saving models compiled with Torch-TensorRT +==================================== + +Saving models compiled with Torch-TensorRT varies slightly with the `ir` that has been used for compilation. + +1) Dynamo IR + +Starting with 2.1 release of Torch-TensorRT, we are switching the default compilation to be dynamo based. +The output of `ir=dynamo` compilation is a `torch.fx.GraphModule` object. There are two ways to save these objects + +a) Converting to Torchscript +`torch.fx.GraphModule` objects cannot be serialized directly. Hence we use `torch.jit.trace` to convert this into a `ScriptModule` object which can be saved to disk. +The following code illustrates this approach. + +.. code-block:: python + + import torch + import torch_tensorrt + + model = MyModel().eval().cuda() + inputs = torch.randn((1, 3, 224, 224)).cuda() + trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) # Output is a torch.fx.GraphModule + trt_script_model = torch.jit.trace(trt_gm, inputs) + torch.jit.save(trt_script_model, "trt_model.ts") + + # Later, you can load it and run inference + model = torch.jit.load("trt_model.ts").cuda() + model(inputs) + +b) ExportedProgram +`torch.export.ExportedProgram` is a new format introduced in Pytorch 2.1. After we compile a Pytorch module using Torch-TensorRT, the resultant +`torch.fx.GraphModule` along with additional metadata can be used to create `ExportedProgram` which can be saved and loaded from disk. + +.. code-block:: python + + import torch + import torch_tensorrt + from torch_tensorrt.dynamo.export import transform, create_exported_program + + model = MyModel().eval().cuda() + inputs = torch.randn((1, 3, 224, 224)).cuda() + trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) # Output is a torch.fx.GraphModule + # Transform and create an exported program + trt_gm = transform(trt_gm, inputs) + trt_exp_program = create_exported_program(trt_gm, call_spec, trt_gm.state_dict()) + torch._export.save(trt_exp_program, "trt_model.ep") + + # Later, you can load it and run inference + model = torch._export.load("trt_model.ep") + model(inputs) + +`torch_tensorrt.dynamo.export.transform` inlines the submodules within a GraphModule to their corresponding nodes and stiches all the nodes together. +This is needed as `torch._export` serialization cannot handle serializing and deserializing of submodules (`call_module` nodes). + +NOTE: This way of saving the models using `ExportedProgram` is experimental. Here is a known issue : https://github.com/pytorch/TensorRT/issues/2341 + +2) Torchscript IR + + In Torch-TensorRT 1.X versions, the primary way to compile and run inference with Torch-TensorRT is using Torchscript IR. + This behavior stays the same in 2.X versions as well. + + .. code-block:: python + + import torch + import torch_tensorrt + + model = MyModel().eval().cuda() + inputs = torch.randn((1, 3, 224, 224)).cuda() + trt_ts = torch_tensorrt.compile(model, ir="ts", inputs) # Output is a ScriptModule object + torch.jit.save(trt_ts, "trt_model.ts") + + # Later, you can load it and run inference + model = torch.jit.load("trt_model.ts").cuda() + model(inputs) + diff --git a/docs/_static/documentation_options.js b/docs/_static/documentation_options.js index eee6308e52..b97c8ee26e 100644 --- a/docs/_static/documentation_options.js +++ b/docs/_static/documentation_options.js @@ -1,6 +1,6 @@ var DOCUMENTATION_OPTIONS = { URL_ROOT: document.getElementById("documentation_options").getAttribute('data-url_root'), - VERSION: 'v2.0.0.dev0+1fec519', + VERSION: 'v2.2.0.dev0+50ab2c1', LANGUAGE: 'None', COLLAPSE_INDEX: false, BUILDER: 'html', diff --git a/docs/cli/torchtrtc.html b/docs/cli/torchtrtc.html index d1347d3394..5d8cdfde3f 100644 --- a/docs/cli/torchtrtc.html +++ b/docs/cli/torchtrtc.html @@ -10,7 +10,7 @@ - torchtrtc — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torchtrtc — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/contributors/conversion.html b/docs/contributors/conversion.html index 98e23c883f..acbf20dbf5 100644 --- a/docs/contributors/conversion.html +++ b/docs/contributors/conversion.html @@ -10,7 +10,7 @@ - Conversion Phase — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Conversion Phase — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/contributors/fx_converters.html b/docs/contributors/fx_converters.html new file mode 100644 index 0000000000..0dbeca96bf --- /dev/null +++ b/docs/contributors/fx_converters.html @@ -0,0 +1,928 @@ + + + + + + + + + + + + + Dynamo Converters — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    +
    +
    + + + + + + + + + + + + + + + + +
    + +
      + +
    • + + + Docs + + > +
    • + + +
    • Dynamo Converters
    • + + +
    • + + + + + +
    • + +
    + + +
    +
    + +
    + Shortcuts +
    +
    + +
    +
    + + + + + + +
    + +
    +
    + +
    +

    Dynamo Converters

    +

    The dynamo converter library in Torch-TensorRT is located in TensorRT/py/torch_tensorrt/dynamo/conversion.

    +
    +
    +

    Steps

    +
    +

    Operation Set

    +

    The converters in dynamo are produced by aten_trace and falls under aten_ops_converters ( FX earlier had acc_ops_converters, aten_ops_converters or nn_ops_converters depending on the trace through which it was produced). The converters are registered using dynamo_tensorrt_converter for dynamo. The function decorated +has the arguments - network, target, args, kwargs, name, which is common across all the operators schema. +These functions are mapped in the aten converter registry dictionary (at present a compilation of FX and dynamo converters, FX will be deprecated soon), with key as the function target name.

    +
    +
      +
    • aten_trace is produced by torch_tensorrt.dynamo.trace(..) for the export path and torch_tensorrt.compile(ir=dynamo) for the compile path.

    • +
    +

    The export path makes use of aten_tracer whereas the alternate trace in compile is produced by the AOT Autograd library. +Both these simplify the torch operators to reduced set of Aten operations.

    +
    +

    As mentioned above, if you would like to add a new converter, its implementation will be included in TensorRT/py/torch_tensorrt/dynamo/conversion/impl +Although there is a corresponding implementation of the converters included in the common implementation library present in TensorRT/py/torch_tensorrt/fx/impl for FX converters, this documentation focuses on the implementation of the aten_ops converters in dynamo.

    +
    +
    +

    Converter implementation

    +

    In this section, we illustrate the steps to be implemented for writing a converter. We divide them according to activation, operator, lowering pass implementation or evaluator. +Each of them is detailed with the help of an example

    +
    +
      +
    • Registration

      +
      +

      The converter needs to be registered with the appropriate op code in the dynamo_tensorrt_converter.

      +
        +
      • Activation type

        +
        +

        Example: leaky_relu

        +
          +
        • aten_ops_converters: Dynamo_converters

          +
          +

          Define in py/torch_tensorrt/dynamo/conversion/aten_ops_converters. One needs to register the opcode generated in the trace with dynamo_tensorrt_converter decorator. Op code to be used for the registration or the converter registry key in this case is torch.ops.aten.leaky_relu.default

          +
          +
          @dynamo_tensorrt_converter(torch.ops.aten.leaky_relu.default)
          +    def aten_ops_leaky_relu(
          +        network: TRTNetwork,
          +        target: Target,
          +        args: Tuple[Argument, ...],
          +        kwargs: Dict[str, Argument],
          +        name: str,
          +    ) -> Union[TRTTensor, Sequence[TRTTensor]]:
          +            return activation.leaky_relu(network, target, SourceIR.ATEN, name, args[0], args[1])
          +
          +
          +
          +
          +
        • +
        +

        The tensorrt_converter (used for FX registration) and dynamo_tensorrt_converter are similar decorator functions with some differences.

        +
          +
        1. Both register the converters in the registeries (python dictionaries) - CONVERTERS and DYNAMO_CONVERTERS respectively. These are two dictioneries which are concatenated to form the overall converter registry

        2. +
        3. The dictionary is keyed on the OpOverLoad which is mentioned in more detail below with examples

        4. +
        5. Both return the decorated converter implementation

        6. +
        7. The CONVERTERS directly registers the decorated converter_implementation function, while DYNAMO_CONVERTERS has additionational arguments and registers the ConverterSupport object

        8. +
        9. The additional arguments are:

          +
          +
            +
          1. key: Node target for which the converter is implemented for (for example, torch.ops.aten.leaky_relu.Tensor)

          2. +
          3. enabled: Whether the converter should be enabled/cached or not

          4. +
          5. capability_validator: Function which evaluates whether a node is valid for conversion by the decorated converter. It defaults to None, implying the capability_validator function is always true. This means all nodes of “key” kind can be supported by this converter by default. See embedding example for more details

          6. +
          7. priority: Converter’s level of priority relative to other converters with the same target

          8. +
          +
          +
        10. +
        11. The ConverterSupport is a compilation of converter_implementation and capability_validator.

        12. +
        +

        The function decorated by tensorrt_converter and dynamo_tensorrt_converter has the following arguments which are automatically generated by the trace functions mentioned above.

        +
          +
        1. network : Node in the form of call_module or call_function having the target as the key

        2. +
        3. target: Target key in the call_module or call_function above. eg: torch.ops.aten_.leaky_relu.default. Note that torch.ops.aten._leaky_relu is the OpOverloadPacket while torch.ops.aten_.leaky_relu.default is OpOverload.

        4. +
        5. args: The arguments passed in the call_module or call_function above

        6. +
        7. kwargs: The kwargs passed in the call_module or call_function above

        8. +
        9. name: String containing the name of the target

        10. +
        +

        As a user writing new converters, one just needs to take care that the approriate arguments are extracted from the trace generated to the implementation function in the implementation lib function activation.leaky_relu (which we will discuss below in detail).

        +
        +
      • +
      • Operation type

        +
        +

        Example: fmod

        +

        It follows the same steps as the above converter. In this case the opcode is torch.ops.aten.fmod.Scalar or torch.ops.aten.fmod.Tensor. +Hence both the opcodes are registered in py/torch_tensorrt/dynamo/conversion/aten_ops_converters. +Note that torch.ops.aten.fmod is the OpOverLoadPacket while the registry is keyed on torch.ops.aten.fmod.Scalar or torch.ops.aten.fmod.Tensor, which is OpOverLoad

        +

        Example: embedding

        +

        It follows the same steps as the above converter. In this case the opcode is torch.ops.aten.embedding.default. +There are some converters which have special cases to be accounted for. In those cases, one should use capability_validators to register the converter using @dynamo_tensorrt_converter +We illustrate this through torch.ops.aten.embedding.default. It has parameters - scale_grad_by_freq and sparse which are not currently supported by the implementation. +In such cases we can write validator embedding_param_validator which implements that given those paramters the converter is not supported and register the converter by

        +
        +
        +

        So if there is a new converter in which certain special cases are not to be supported then they can be specified in the capability_validator.

        +
        +
      • +
      • Evaluator type

        +
        +

        Example: operator.getitem

        +

        Evaluators are categorized as so since they do not make any modification to the graph. This is implemented in py/torch_tensorrt/dynamo/conversion/op_evaluators.py, with the corresponding capbility_validator. +The opcode is operator.getitem.

        +
        +
      • +
      +
      +
    • +
    • Implementation Library

      +
      +

      The dynamo converters would be located in py/torch_tensorrt/dynamo/conversion/impl

      +
        +
      • Activation

        +
        +

        Example: leaky_relu

        +

        The implementation is to be placed in present in py/torch_tensorrt/dynamo/conversion/impl/activation.py. This is where all the activation functions are defined and implemented.

        +
        def leaky_relu(
        +    network: TRTNetwork,
        +    target: Target,
        +    source_ir: Optional[SourceIR],
        +    name: str,
        +    input_val: TRTTensor,
        +    alpha: Optional[Any],
        +):
        +    #implementation
        +
        +
        +

        The implementation function has the following arguments.

        +
          +
        1. network : network passed from the decorated function registration

        2. +
        3. target: target passed from the decorated function registration

        4. +
        5. source_ir: Enum attribute. SourceIR enum is defined in py/torch_tensorrt/dynamo/conversion/impl/converter_utils

        6. +
        7. name: name passed from the decorated function registration

        8. +
        9. input_val: Approriate arguments extracted from the decorated function registration from args or kwargs

        10. +
        11. alpha: Approriate arguments extracted from the decorated function registration from args or kwargs. If not None, it will set the alpha attribute of the created TensorRT activation layer eg: Used in leaky_relu, elu, hardtanh

        12. +
        13. beta: Approriate arguments extracted from the decorated function registration from args or kwargs. If not None, it will set the beta attribute of the created TensorRT activation layer eg: Used in hardtanh

        14. +
        15. dyn_range_fn: A optional function which takes the dynamic range of a TensorRT Tensor and returns the output dynamic range

        16. +
        +

        The implementation functions call the convert_activation function in py/torch_tensorrt/dynamo/conversion/impl/activation.py. This function will add the approriate activation layer via network.add_activation.

        +
        +
      • +
      • Operator

        +
        +

        The implementation is to be placed in py/torch_tensorrt/dynamo/conversion/impl/elementwise/ops.py for dynamo. This is where all the elementwise functions are defined and implemented. +For a new operator, one should identify the category to which it belongs. Following are some examples

        +
          +
        1. Elementwise operators like fmod is present in py/torch_tensorrt/dynamo/conversion/impl/elementwise. The py/torch_tensorrt/dynamo/conversion/impl/elementwise/base contains base functions for elementwise operator.

        2. +
        3. Unary operators like sqrt will be present in py/torch_tensorrt/dynamo/conversion/impl/unary. The py/torch_tensorrt/dynamo/conversion/impl/unary/base contains base functions for unary operator.

        4. +
        5. Normalization operators like softmax, layer_norm, batch_norm will be present in py/torch_tensorrt/dynamo/conversion/impl/normalization. Since there are no base operations common to all, there is no base file. But one can choose to implement a base file, if there are common functions across all normalization operations

        6. +
        7. Individual operators like slice, select, where, embedding will be present in py/torch_tensorrt/dynamo/conversion/impl/*.py. They will have individual operator implementation with the same API structure as above but with different individual arguments

        8. +
        +

        Please note that the above operators would have common functions to be implemented which should be placed in +py/torch_tensorrt/dynamo/conversion/impl/converter_utils.py

        +
        +
      • +
      +
      +
    • +
    • Lowering type

      +
      +

      There are some converters which can be decomposed into suboperations and need not have seperate converter registration. +Such converters can be implemented via lowering passes

      +

      Example: addmm

      +

      The decompositions are registered via register_decomposition in py/torch_tensorrt/dynamo/backend/lowering/_decompositions.py +We define addmm_replacement and replace it with the torch ops, which will have their corresponding converters called.

      +
      @register_decomposition(torch.ops.aten.addmm, registry=DECOMPOSITIONS)
      +def addmm_replacement(
      +    input_: torch.Tensor, mat1: torch.Tensor, mat2: torch.Tensor, *, beta=1, alpha=1
      +) -> torch.Tensor:
      +    return torch.add(
      +        torch.mul(input_, beta), torch.mul(torch.matmul(mat1, mat2), alpha)
      +    )
      +
      +
      +

      Note that there are some pre-existing dynamo decompositions in torch directory, in which case they should be used, +In that case please enable the decompositions in py/torch_tensorrt/dynamo/lowering/_decomposition_groups.py in torch_enabled_decompositions. +Similarly you can choose to disable any in torch_disabled_decompositions. Please note that the ones already defined in the lowering will take precedence over torch lowering ops.

      +
      +
    • +
    +
    +
    +
    +

    Tests

    +
      +
    • Dynamo testing:

      +
      +

      Dynamo tests are present for the lowering ops in tests/py/dynamo/lowering/test_decompositions.py. The above converters will soon be ported to dynamo tests

      +
        +
      1. Compare the results for fx.symbolic_trace and torch_tensorrt.dynamo.compile.

      2. +
      3. Test for the expected_op and the unexpected_op.

        +
        +
          +
        1. expected_op: Operations the operations are lowered to. eg: mul and add for addmm

        2. +
        3. Note that specify that disable_passes= True for cases where you would not want lowering passes (which should be the default when testing converters)

        4. +
        5. unexpected_op: Original operation. eg: addmm for addmm

        6. +
        +
        +
      4. +
      +
      +
    • +
    +

    The tests should fail if any of the above two conditions fail

    +
    +
    + + +
    + +
    +
    + + + + +
    + + + +
    +

    + © Copyright 2022, NVIDIA Corporation. + +

    +
    + +
    + Built with Sphinx using a theme provided by Read the Docs. +
    + + +
    + +
    +
    + + +
    +
    + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    +
    +
    +
    +

    Docs

    +

    Access comprehensive developer documentation for PyTorch

    + View Docs +
    + +
    +

    Tutorials

    +

    Get in-depth tutorials for beginners and advanced developers

    + View Tutorials +
    + +
    +

    Resources

    +

    Find development resources and get your questions answered

    + View Resources +
    +
    +
    +
    + + + + + + + + + +
    +
    +
    +
    + + +
    +
    +
    + + +
    + + + + + + + + \ No newline at end of file diff --git a/docs/contributors/lowering.html b/docs/contributors/lowering.html index 6f6c71f6f6..7440afeadf 100644 --- a/docs/contributors/lowering.html +++ b/docs/contributors/lowering.html @@ -10,7 +10,7 @@ - Lowering Phase — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Lowering Phase — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/contributors/partitioning.html b/docs/contributors/partitioning.html index 681b57902e..4b25892c59 100644 --- a/docs/contributors/partitioning.html +++ b/docs/contributors/partitioning.html @@ -10,7 +10,7 @@ - Partitioning Phase — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Partitioning Phase — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/contributors/phases.html b/docs/contributors/phases.html index f7bcbfb28f..cf874de2ce 100644 --- a/docs/contributors/phases.html +++ b/docs/contributors/phases.html @@ -10,7 +10,7 @@ - Compiler Phases — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Compiler Phases — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    @@ -411,7 +414,7 @@

    Conversion

    Compilation and Runtime

    -

    Deploying Torch-TensorRT Programs

    +

    Saving models compiled with Torch-TensorRT

    The final compilation phase constructs a TorchScript program to run the converted TensorRT engine. It takes a serialized engine and instantiates it within a engine manager, then the compiler will build out a JIT graph that references this engine and wraps it in a module to return to the user. diff --git a/docs/contributors/runtime.html b/docs/contributors/runtime.html index a5697ba53a..844df07d7a 100644 --- a/docs/contributors/runtime.html +++ b/docs/contributors/runtime.html @@ -10,7 +10,7 @@ - Runtime Phase — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Runtime Phase — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@

    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/contributors/system_overview.html b/docs/contributors/system_overview.html index 7798c7309a..b4fa7d9ed6 100644 --- a/docs/contributors/system_overview.html +++ b/docs/contributors/system_overview.html @@ -10,7 +10,7 @@ - System Overview — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + System Overview — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    @@ -434,7 +437,7 @@

    Conversion

    Compilation and Runtime

    -

    Deploying Torch-TensorRT Programs

    +

    Saving models compiled with Torch-TensorRT

    The final compilation phase constructs a TorchScript program to run the converted TensorRT engine. It takes a serialized engine and instantiates it within a engine manager, then the compiler will build out a JIT graph that references this engine and wraps it in a module to return to the user. diff --git a/docs/contributors/useful_links.html b/docs/contributors/useful_links.html index 72e593593a..006caf967e 100644 --- a/docs/contributors/useful_links.html +++ b/docs/contributors/useful_links.html @@ -10,7 +10,7 @@ - Useful Links for Torch-TensorRT Development — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Useful Links for Torch-TensorRT Development — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -40,7 +40,7 @@ - + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +

    + + + + + + + + + + + +
    +
    +
    + + + + + + + + + + + + + + + + +
    + +
      + +
    • + + + Docs + + > +
    • + + +
    • Writing Dynamo ATen Lowering Passes
    • + + +
    • + + + + + +
    • + +
    + + +
    +
    + +
    + Shortcuts +
    +
    + +
    +
    + + + + + + +
    + +
    +
    + +
    +

    Writing Dynamo ATen Lowering Passes

    +
    +

    Basics of a Lowering Pass

    +

    ATen lowering passes are Python functions which take as input a graph of ATen operators, apply some desired modification such as operator coalescing/fusion, operator replacement, subgraph rewriting, custom operator insertion, or other operation on a torch.fx.GraphModule, then return the modified graph to the caller. These lowering passes generally modify the graph in-place and return the same input object.

    +
    +
    +

    Lowering Pass Requirements

    +

    An ATen lowering pass function in Torch-TRT must satisfy two requirements: +- The function must take as input a torch.fx.GraphModule and a sequence of torch Tensors, Sequence[torch.Tensor], and return the lowered torch.fx.GraphModule +- The function must leave the graph in a valid and invoke-able state, including performing any necessary linting and recompilation

    +

    See this link for information on Graph Manipulations in FX. See below for an example of a lowering pass which repairs graphs that have inputs which are also outputs, a disallowed configuration for TRT Engines.

    +
    +
    +

    Example Lowering Pass

    +
    def repair_input_as_output(gm: torch.fx.GraphModule, sample_inputs: Sequence[torch.Tensor]) -> torch.fx.GraphModule:
    +    """Repair scenarios where inputs are also outputs of the graph
    +
    +    TRT does not allow such cases, so we insert a clone (identity) layer
    +    """
    +    modified_graph = False
    +
    +    # Extract graph placeholder Tensors
    +    placeholders = [
    +        node
    +        for node in gm.graph.nodes
    +        if (
    +            node.op == "placeholder"
    +            and isinstance(node.type, type)
    +            and issubclass(node.type, torch.Tensor)
    +        )
    +    ]
    +
    +    for placeholder in placeholders:
    +        # If any placeholder has any users which are direct graph outputs
    +        if len(placeholder.users) >= 1 and any(
    +            user.op == "output" for user in placeholder.users
    +        ):
    +            modified_graph = True
    +
    +            # Get direct graph outputs which are direct uses of placeholders
    +            direct_outputs = [user for user in placeholder.users if user.op == "output"]
    +
    +            # Insert clone node for placeholder to ensure
    +            # placeholder is not a direct output
    +            with gm.graph.inserting_after(placeholder):
    +                cloned_placeholder = gm.graph.call_function(
    +                    torch.ops.aten.clone.default,
    +                    args=(placeholder,),
    +                )
    +
    +            # Replace placeholder as output with cloned version
    +            for output in direct_outputs:
    +                output.replace_input_with(placeholder, cloned_placeholder)
    +
    +    # If the graph was modified, clean up the graph and ensure it is up-to-date
    +    if modified_graph:
    +        gm.graph.eliminate_dead_code()
    +        gm.graph.lint()
    +        gm.recompile()
    +        logger.debug(f"Graph after repair_input_as_output:\n{gm.graph}")
    +
    +    return gm
    +
    +
    +
    +
    +

    Registering Lowering Passes

    +

    Lowering passes are currently registered in py/torch_tensorrt/dynamo/lowering/passes/__init__.py, using the torch.fx.passes.pass_manager.PassManager utility to assemble the list of passes in a desired order. New passes added directly to that list will be applied to graphs in the Torch-TensorRT torch.compile backend. Currently, we offer an ATen lowering pass registration decorator for convenience, which can be invoked either directly, or with the optional index keyword argument which controls where in the pass list the lowering pass will be inserted.

    +

    For instance, to insert the pass at the default location (end of the list), the following code can be used:

    +
    @_aten_lowering_pass
    +def my_custom_pass(gm: torch.fx.GraphModule, sample_inputs: Sequence[torch.Tensor]) -> torch.fx.GraphModule:
    +    ...
    +
    +
    +

    Alternatively, to insert the pass at a custom index (such as the front of the list) in the passlist, the following code can be used:

    +
    @_aten_lowering_pass(index=0)
    +def my_custom_pass(gm: torch.fx.GraphModule, sample_inputs: Sequence[torch.Tensor]) -> torch.fx.GraphModule:
    +    ...
    +
    +
    +

    There are also provided utilities in torch_tensorrt.dynamo.lowering.passes for displaying the currently-available lowering pass list, applying those passes to an arbitrary torch.fx.GraphModule, and removing the lowering pass at a specific index.

    +
    # Print all lowering passes in the list
    +print(dump_lowering_passes())
    +
    +# Apply lowering passes to a GraphModule
    +apply_lowering_passes(graph_module, sample_inputs)
    +
    +# Remove the lowering pass at index 1
    +_remove_lowering_pass(index=1)
    +
    +
    +

    Note: The above APIs are subject to change, as the lowering pass system evolves.

    +
    +
    + + +
    + +
    + + +
    +
    + + +
    +
    + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    +
    +
    +
    +

    Docs

    +

    Access comprehensive developer documentation for PyTorch

    + View Docs +
    + +
    +

    Tutorials

    +

    Get in-depth tutorials for beginners and advanced developers

    + View Tutorials +
    + +
    +

    Resources

    +

    Find development resources and get your questions answered

    + View Resources +
    +
    +
    +
    + + + + + + + + + +
    +
    +
    +
    + + +
    +
    +
    + + +
    + + + + + + + + \ No newline at end of file diff --git a/docs/genindex.html b/docs/genindex.html index d1a67ea893..25ef16bb39 100644 --- a/docs/genindex.html +++ b/docs/genindex.html @@ -9,7 +9,7 @@ - Index — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Index — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    @@ -720,6 +723,8 @@

    T

  • torch_tensorrt::DataType::Value::kBool (C++ enumerator)
  • torch_tensorrt::DataType::Value::kChar (C++ enumerator) +
  • +
  • torch_tensorrt::DataType::Value::kDouble (C++ enumerator)
  • torch_tensorrt::DataType::Value::kFloat (C++ enumerator)
  • diff --git a/docs/getting_started/getting_started_with_cpp_api.html b/docs/getting_started/getting_started_with_cpp_api.html index bc90526366..8edb42dffc 100644 --- a/docs/getting_started/getting_started_with_cpp_api.html +++ b/docs/getting_started/getting_started_with_cpp_api.html @@ -10,7 +10,7 @@ - Using Torch-TensorRT in C++ — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Using Torch-TensorRT in C++ — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/getting_started/getting_started_with_python_api.html b/docs/getting_started/getting_started_with_python_api.html index d9228c3c91..bf009991ba 100644 --- a/docs/getting_started/getting_started_with_python_api.html +++ b/docs/getting_started/getting_started_with_python_api.html @@ -10,7 +10,7 @@ - Using Torch-TensorRT in Python — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Using Torch-TensorRT in Python — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/getting_started/getting_started_with_windows.html b/docs/getting_started/getting_started_with_windows.html index f7ef856916..887622e1af 100644 --- a/docs/getting_started/getting_started_with_windows.html +++ b/docs/getting_started/getting_started_with_windows.html @@ -10,7 +10,7 @@ - Building Torch-TensorRT on Windows — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Building Torch-TensorRT on Windows — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/getting_started/installation.html b/docs/getting_started/installation.html index 7f7e7451a7..59575783c7 100644 --- a/docs/getting_started/installation.html +++ b/docs/getting_started/installation.html @@ -10,7 +10,7 @@ - Installation — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Installation — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/index.html b/docs/index.html index 2b39807853..b35a38801f 100644 --- a/docs/index.html +++ b/docs/index.html @@ -10,7 +10,7 @@ - Torch-TensorRT — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Torch-TensorRT — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -224,7 +224,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -268,6 +268,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -303,6 +305,7 @@

    Indices

    @@ -416,7 +419,9 @@

    User GuideCreating a TorchScript Module

  • Torch-TensorRT (FX Frontend) User Guide

  • Post Training Quantization (PTQ)

  • -
  • Deploying Torch-TensorRT Programs

  • +
  • Saving models compiled with Torch-TensorRT

  • +
  • saving_models

  • +
  • dynamic_shapes

  • Using Torch-TensorRT Directly From PyTorch

  • DLA

  • @@ -469,6 +474,7 @@

    Contributor Documentation
  • System Overview

  • Writing Converters

  • +
  • Writing Dynamo ATen Lowering Passes

  • Useful Links for Torch-TensorRT Development

  • diff --git a/docs/indices/supported_ops.html b/docs/indices/supported_ops.html index 6598896642..66e857341f 100644 --- a/docs/indices/supported_ops.html +++ b/docs/indices/supported_ops.html @@ -10,7 +10,7 @@ - Operators Supported — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Operators Supported — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -224,7 +224,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -268,6 +268,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -303,6 +305,7 @@

    Indices

    diff --git a/docs/objects.inv b/docs/objects.inv index af73fd8b81..4d08d3c9c5 100644 Binary files a/docs/objects.inv and b/docs/objects.inv differ diff --git a/docs/py-modindex.html b/docs/py-modindex.html index c76c976578..9a76e3b420 100644 --- a/docs/py-modindex.html +++ b/docs/py-modindex.html @@ -9,7 +9,7 @@ - Python Module Index — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Python Module Index — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/py_api/fx.html b/docs/py_api/fx.html index 19212dd9a8..b9594b6e3c 100644 --- a/docs/py_api/fx.html +++ b/docs/py_api/fx.html @@ -10,7 +10,7 @@ - torch_tensorrt.fx — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.fx — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/py_api/logging.html b/docs/py_api/logging.html index 1c5e880b68..f1eee93f87 100644 --- a/docs/py_api/logging.html +++ b/docs/py_api/logging.html @@ -10,7 +10,7 @@ - torch_tensorrt.logging — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.logging — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/py_api/ptq.html b/docs/py_api/ptq.html index 7d104bc55d..d927b9b6ff 100644 --- a/docs/py_api/ptq.html +++ b/docs/py_api/ptq.html @@ -10,7 +10,7 @@ - torch_tensorrt.ptq — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.ptq — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/py_api/torch_tensorrt.html b/docs/py_api/torch_tensorrt.html index a17f769268..2a759f4b92 100644 --- a/docs/py_api/torch_tensorrt.html +++ b/docs/py_api/torch_tensorrt.html @@ -10,7 +10,7 @@ - torch_tensorrt — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    @@ -562,7 +565,7 @@

    Classes
    -dtype: _enums.dtype = <dtype.unknown: 6>
    +dtype: _enums.dtype = <dtype.unknown: 7>

    torch_tensorrt.dtype.float32)

    Type
    @@ -730,6 +733,8 @@

    Enums

    int32 : 32 bit integer number

    long : 64 bit integer number

    int64 : 64 bit integer number

    +

    double : 64 bit floating point number

    +

    float64 : 64 bit floating point number

    bool : Boolean value

    unknown : Unknown data type

    diff --git a/docs/py_api/ts.html b/docs/py_api/ts.html index 7381dc340d..6331ec92d1 100644 --- a/docs/py_api/ts.html +++ b/docs/py_api/ts.html @@ -10,7 +10,7 @@ - torch_tensorrt.ts — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + torch_tensorrt.ts — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    @@ -609,7 +612,7 @@

    Functions
    -torch_tensorrt.ts.TensorRTCompileSpec(inputs: typing.Optional[typing.List[torch.Tensor | torch_tensorrt._Input.Input]] = None, input_signature: typing.Optional[typing.Any] = None, device: torch.device | torch_tensorrt._Device.Device = Device(type=DeviceType.GPU, gpu_id=0), disable_tf32: bool = False, sparse_weights: bool = False, enabled_precisions: typing.Optional[typing.Set[torch.dtype | torch_tensorrt._C.dtype]] = None, refit: bool = False, debug: bool = False, capability: torch_tensorrt._C.EngineCapability = <EngineCapability.default: 0>, num_avg_timing_iters: int = 1, workspace_size: int = 0, dla_sram_size: int = 1048576, dla_local_dram_size: int = 1073741824, dla_global_dram_size: int = 536870912, truncate_long_and_double: bool = False, calibrator: object = None, allow_shape_tensors: bool = False) <torch.ScriptClass object at 0x7f6412476370>[source]
    +torch_tensorrt.ts.TensorRTCompileSpec(inputs: typing.Optional[typing.List[torch.Tensor | torch_tensorrt._Input.Input]] = None, input_signature: typing.Optional[typing.Any] = None, device: torch.device | torch_tensorrt._Device.Device = Device(type=DeviceType.GPU, gpu_id=0), disable_tf32: bool = False, sparse_weights: bool = False, enabled_precisions: typing.Optional[typing.Set[torch.dtype | torch_tensorrt._C.dtype]] = None, refit: bool = False, debug: bool = False, capability: torch_tensorrt._C.EngineCapability = <EngineCapability.default: 0>, num_avg_timing_iters: int = 1, workspace_size: int = 0, dla_sram_size: int = 1048576, dla_local_dram_size: int = 1073741824, dla_global_dram_size: int = 536870912, truncate_long_and_double: bool = False, calibrator: object = None, allow_shape_tensors: bool = False) <torch.ScriptClass object at 0x7f6f6edc45b0>[source]

    Utility to create a formated spec dictionary for using the PyTorch TensorRT backend

    Keyword Arguments
    diff --git a/docs/search.html b/docs/search.html index 55f33c22a0..888364a4b1 100644 --- a/docs/search.html +++ b/docs/search.html @@ -9,7 +9,7 @@ - Search — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Search — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -222,7 +222,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -266,6 +266,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -301,6 +303,7 @@

    Indices

    diff --git a/docs/searchindex.js b/docs/searchindex.js index 0f80738941..438b54023a 100644 --- a/docs/searchindex.js +++ b/docs/searchindex.js @@ -1 +1 @@ -Search.setIndex({docnames:["_cpp_api/classtorch__tensorrt_1_1DataType","_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType","_cpp_api/classtorch__tensorrt_1_1TensorFormat","_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator","_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator","_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502","_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268","_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e","_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827","_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b","_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da","_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59","_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883","_cpp_api/dir_cpp","_cpp_api/dir_cpp_include","_cpp_api/dir_cpp_include_torch_tensorrt","_cpp_api/enum_namespacetorch__tensorrt_1_1logging_1a130f65408ad8cbaee060f05e8db69558","_cpp_api/enum_namespacetorch__tensorrt_1a3fbe5d72e4fc624dbd038853079620eb","_cpp_api/file_cpp_include_torch_tensorrt_logging.h","_cpp_api/file_cpp_include_torch_tensorrt_macros.h","_cpp_api/file_cpp_include_torch_tensorrt_ptq.h","_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0593f776f469c20469e2f729fc7861a3","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0c012cb374addd90eb1f42eaec570650","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a56e110feaaba2c3fd44bd201fd21a76a","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a7cb50492421ea9de4e3db895819df6f2","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ac46ac0901cb97e3ae6e93b45f24e90b8","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ad2efd47b6c3689e58ccc595680579ae5","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1af8f3443813315af7901903d25dd495cc","_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a226e3c83379d1012cde8578c1c86b16c","_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a6186e305f47c1d94b6130ef6c7f7e178","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a5b405fd3bf3c8fc2e2a54cbbab979797","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a6e19490a08fb1553c9dd347a5ae79db9","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a81f9783517335dda877d8cfcf38987c9","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1ae8d56472106eeef37fbe51ff7f40c9b2","_cpp_api/function_namespacetorch__tensorrt_1ac4ab8313ae72c2c899ea31548b528528","_cpp_api/function_namespacetorch__tensorrt_1ad1acd06eaeaffbbcf6e7ebf426891384","_cpp_api/function_namespacetorch__tensorrt_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1","_cpp_api/namespace_torch_tensorrt","_cpp_api/namespace_torch_tensorrt__logging","_cpp_api/namespace_torch_tensorrt__ptq","_cpp_api/namespace_torch_tensorrt__torchscript","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h","_cpp_api/structtorch__tensorrt_1_1Device","_cpp_api/structtorch__tensorrt_1_1GraphInputs","_cpp_api/structtorch__tensorrt_1_1Input","_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec","_cpp_api/torch_tensort_cpp","_cpp_api/unabridged_orphan","cli/torchtrtc","contributors/conversion","contributors/lowering","contributors/partitioning","contributors/phases","contributors/runtime","contributors/system_overview","contributors/useful_links","contributors/writing_converters","getting_started/getting_started_with_cpp_api","getting_started/getting_started_with_python_api","getting_started/getting_started_with_windows","getting_started/installation","index","indices/supported_ops","py_api/fx","py_api/logging","py_api/ptq","py_api/torch_tensorrt","py_api/ts","src/pytorch-sphinx-theme/docs/changelog","src/pytorch-sphinx-theme/docs/configuring","src/pytorch-sphinx-theme/docs/demo/api","src/pytorch-sphinx-theme/docs/demo/demo","src/pytorch-sphinx-theme/docs/demo/lists_tables","src/pytorch-sphinx-theme/docs/demo/long","src/pytorch-sphinx-theme/docs/demo/structure","src/pytorch-sphinx-theme/docs/index","src/pytorch-sphinx-theme/docs/installing","tutorials/_rendered_examples/dynamo/index","tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage","tutorials/_rendered_examples/dynamo/torch_compile_resnet_example","tutorials/_rendered_examples/dynamo/torch_compile_transformers_example","tutorials/_rendered_examples/index","tutorials/notebooks","tutorials/serving_torch_tensorrt_with_triton","user_guide/creating_torchscript_module_in_python","user_guide/getting_started_with_fx_path","user_guide/ptq","user_guide/runtime","user_guide/use_from_pytorch","user_guide/using_dla"],envversion:{"sphinx.domains.c":2,"sphinx.domains.changeset":1,"sphinx.domains.citation":1,"sphinx.domains.cpp":5,"sphinx.domains.index":1,"sphinx.domains.javascript":2,"sphinx.domains.math":2,"sphinx.domains.python":3,"sphinx.domains.rst":2,"sphinx.domains.std":2,"sphinx.ext.intersphinx":1,"sphinx.ext.todo":2,"sphinx.ext.viewcode":1,nbsphinx:4,sphinx:56},filenames:["_cpp_api/classtorch__tensorrt_1_1DataType.rst","_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.rst","_cpp_api/classtorch__tensorrt_1_1TensorFormat.rst","_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.rst","_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.rst","_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.rst","_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.rst","_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.rst","_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.rst","_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.rst","_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.rst","_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.rst","_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.rst","_cpp_api/dir_cpp.rst","_cpp_api/dir_cpp_include.rst","_cpp_api/dir_cpp_include_torch_tensorrt.rst","_cpp_api/enum_namespacetorch__tensorrt_1_1logging_1a130f65408ad8cbaee060f05e8db69558.rst","_cpp_api/enum_namespacetorch__tensorrt_1a3fbe5d72e4fc624dbd038853079620eb.rst","_cpp_api/file_cpp_include_torch_tensorrt_logging.h.rst","_cpp_api/file_cpp_include_torch_tensorrt_macros.h.rst","_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.rst","_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0593f776f469c20469e2f729fc7861a3.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0c012cb374addd90eb1f42eaec570650.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a56e110feaaba2c3fd44bd201fd21a76a.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a7cb50492421ea9de4e3db895819df6f2.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ac46ac0901cb97e3ae6e93b45f24e90b8.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ad2efd47b6c3689e58ccc595680579ae5.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1af8f3443813315af7901903d25dd495cc.rst","_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a226e3c83379d1012cde8578c1c86b16c.rst","_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a6186e305f47c1d94b6130ef6c7f7e178.rst","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a5b405fd3bf3c8fc2e2a54cbbab979797.rst","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a6e19490a08fb1553c9dd347a5ae79db9.rst","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a81f9783517335dda877d8cfcf38987c9.rst","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1ae8d56472106eeef37fbe51ff7f40c9b2.rst","_cpp_api/function_namespacetorch__tensorrt_1ac4ab8313ae72c2c899ea31548b528528.rst","_cpp_api/function_namespacetorch__tensorrt_1ad1acd06eaeaffbbcf6e7ebf426891384.rst","_cpp_api/function_namespacetorch__tensorrt_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.rst","_cpp_api/namespace_torch_tensorrt.rst","_cpp_api/namespace_torch_tensorrt__logging.rst","_cpp_api/namespace_torch_tensorrt__ptq.rst","_cpp_api/namespace_torch_tensorrt__torchscript.rst","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.rst","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.rst","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.rst","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst","_cpp_api/structtorch__tensorrt_1_1Device.rst","_cpp_api/structtorch__tensorrt_1_1GraphInputs.rst","_cpp_api/structtorch__tensorrt_1_1Input.rst","_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.rst","_cpp_api/torch_tensort_cpp.rst","_cpp_api/unabridged_orphan.rst","cli/torchtrtc.rst","contributors/conversion.rst","contributors/lowering.rst","contributors/partitioning.rst","contributors/phases.rst","contributors/runtime.rst","contributors/system_overview.rst","contributors/useful_links.rst","contributors/writing_converters.rst","getting_started/getting_started_with_cpp_api.rst","getting_started/getting_started_with_python_api.rst","getting_started/getting_started_with_windows.rst","getting_started/installation.rst","index.rst","indices/supported_ops.rst","py_api/fx.rst","py_api/logging.rst","py_api/ptq.rst","py_api/torch_tensorrt.rst","py_api/ts.rst","src/pytorch-sphinx-theme/docs/changelog.rst","src/pytorch-sphinx-theme/docs/configuring.rst","src/pytorch-sphinx-theme/docs/demo/api.rst","src/pytorch-sphinx-theme/docs/demo/demo.rst","src/pytorch-sphinx-theme/docs/demo/lists_tables.rst","src/pytorch-sphinx-theme/docs/demo/long.rst","src/pytorch-sphinx-theme/docs/demo/structure.rst","src/pytorch-sphinx-theme/docs/index.rst","src/pytorch-sphinx-theme/docs/installing.rst","tutorials/_rendered_examples/dynamo/index.rst","tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.rst","tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.rst","tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.rst","tutorials/_rendered_examples/index.rst","tutorials/notebooks.rst","tutorials/serving_torch_tensorrt_with_triton.rst","user_guide/creating_torchscript_module_in_python.rst","user_guide/getting_started_with_fx_path.rst","user_guide/ptq.rst","user_guide/runtime.rst","user_guide/use_from_pytorch.rst","user_guide/using_dla.rst"],objects:{"":[[5,0,1,"c.STR","STR"],[9,0,1,"c.TORCHTRT_API","TORCHTRT_API"],[11,0,1,"c.TORCHTRT_HIDDEN","TORCHTRT_HIDDEN"],[7,0,1,"c.TORCH_TENSORRT_MAJOR_VERSION","TORCH_TENSORRT_MAJOR_VERSION"],[8,0,1,"c.TORCH_TENSORRT_MINOR_VERSION","TORCH_TENSORRT_MINOR_VERSION"],[6,0,1,"c.TORCH_TENSORRT_PATCH_VERSION","TORCH_TENSORRT_PATCH_VERSION"],[12,0,1,"c.TORCH_TENSORRT_VERSION","TORCH_TENSORRT_VERSION"],[10,0,1,"c.XSTR","XSTR"],[0,1,1,"_CPPv4N14torch_tensorrt8DataTypeE","torch_tensorrt::DataType"],[0,2,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeE5Value","torch_tensorrt::DataType::DataType"],[0,2,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeEN3c1010ScalarTypeE","torch_tensorrt::DataType::DataType"],[0,2,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeEv","torch_tensorrt::DataType::DataType"],[0,3,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeE5Value","torch_tensorrt::DataType::DataType::t"],[0,3,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeEN3c1010ScalarTypeE","torch_tensorrt::DataType::DataType::t"],[0,4,1,"_CPPv4N14torch_tensorrt8DataType5ValueE","torch_tensorrt::DataType::Value"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kBoolE","torch_tensorrt::DataType::Value::kBool"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kCharE","torch_tensorrt::DataType::Value::kChar"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value6kFloatE","torch_tensorrt::DataType::Value::kFloat"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kHalfE","torch_tensorrt::DataType::Value::kHalf"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value4kIntE","torch_tensorrt::DataType::Value::kInt"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kLongE","torch_tensorrt::DataType::Value::kLong"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value8kUnknownE","torch_tensorrt::DataType::Value::kUnknown"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kBoolE","torch_tensorrt::DataType::kBool"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kCharE","torch_tensorrt::DataType::kChar"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value6kFloatE","torch_tensorrt::DataType::kFloat"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kHalfE","torch_tensorrt::DataType::kHalf"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value4kIntE","torch_tensorrt::DataType::kInt"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kLongE","torch_tensorrt::DataType::kLong"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value8kUnknownE","torch_tensorrt::DataType::kUnknown"],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypecv5ValueEv","torch_tensorrt::DataType::operator Value"],[0,2,1,"_CPPv4N14torch_tensorrt8DataTypecvbEv","torch_tensorrt::DataType::operator bool"],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypeneE8DataType","torch_tensorrt::DataType::operator!="],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypeneEN8DataType5ValueE","torch_tensorrt::DataType::operator!="],[0,3,1,"_CPPv4NK14torch_tensorrt8DataTypeneE8DataType","torch_tensorrt::DataType::operator!=::other"],[0,3,1,"_CPPv4NK14torch_tensorrt8DataTypeneEN8DataType5ValueE","torch_tensorrt::DataType::operator!=::other"],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypeeqE8DataType","torch_tensorrt::DataType::operator=="],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypeeqEN8DataType5ValueE","torch_tensorrt::DataType::operator=="],[0,3,1,"_CPPv4NK14torch_tensorrt8DataTypeeqE8DataType","torch_tensorrt::DataType::operator==::other"],[0,3,1,"_CPPv4NK14torch_tensorrt8DataTypeeqEN8DataType5ValueE","torch_tensorrt::DataType::operator==::other"],[46,1,1,"_CPPv4N14torch_tensorrt6DeviceE","torch_tensorrt::Device"],[46,2,1,"_CPPv4N14torch_tensorrt6Device6DeviceEv","torch_tensorrt::Device::Device"],[1,1,1,"_CPPv4N14torch_tensorrt6Device10DeviceTypeE","torch_tensorrt::Device::DeviceType"],[46,1,1,"_CPPv4N14torch_tensorrt6Device10DeviceTypeE","torch_tensorrt::Device::DeviceType"],[1,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value","torch_tensorrt::Device::DeviceType::DeviceType"],[1,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE","torch_tensorrt::Device::DeviceType::DeviceType"],[1,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEv","torch_tensorrt::Device::DeviceType::DeviceType"],[46,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value","torch_tensorrt::Device::DeviceType::DeviceType"],[46,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE","torch_tensorrt::Device::DeviceType::DeviceType"],[46,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEv","torch_tensorrt::Device::DeviceType::DeviceType"],[1,3,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value","torch_tensorrt::Device::DeviceType::DeviceType::t"],[1,3,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE","torch_tensorrt::Device::DeviceType::DeviceType::t"],[46,3,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value","torch_tensorrt::Device::DeviceType::DeviceType::t"],[46,3,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE","torch_tensorrt::Device::DeviceType::DeviceType::t"],[1,4,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5ValueE","torch_tensorrt::Device::DeviceType::Value"],[46,4,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5ValueE","torch_tensorrt::Device::DeviceType::Value"],[1,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE","torch_tensorrt::Device::DeviceType::Value::kDLA"],[46,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE","torch_tensorrt::Device::DeviceType::Value::kDLA"],[1,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE","torch_tensorrt::Device::DeviceType::Value::kGPU"],[46,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE","torch_tensorrt::Device::DeviceType::Value::kGPU"],[1,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE","torch_tensorrt::Device::DeviceType::kDLA"],[1,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE","torch_tensorrt::Device::DeviceType::kGPU"],[1,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypecv5ValueEv","torch_tensorrt::Device::DeviceType::operator Value"],[46,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypecv5ValueEv","torch_tensorrt::Device::DeviceType::operator Value"],[1,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceTypecvbEv","torch_tensorrt::Device::DeviceType::operator bool"],[46,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceTypecvbEv","torch_tensorrt::Device::DeviceType::operator bool"],[1,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType","torch_tensorrt::Device::DeviceType::operator!="],[46,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType","torch_tensorrt::Device::DeviceType::operator!="],[1,3,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType","torch_tensorrt::Device::DeviceType::operator!=::other"],[46,3,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType","torch_tensorrt::Device::DeviceType::operator!=::other"],[1,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType","torch_tensorrt::Device::DeviceType::operator=="],[46,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType","torch_tensorrt::Device::DeviceType::operator=="],[1,3,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType","torch_tensorrt::Device::DeviceType::operator==::other"],[46,3,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType","torch_tensorrt::Device::DeviceType::operator==::other"],[46,6,1,"_CPPv4N14torch_tensorrt6Device18allow_gpu_fallbackE","torch_tensorrt::Device::allow_gpu_fallback"],[46,6,1,"_CPPv4N14torch_tensorrt6Device11device_typeE","torch_tensorrt::Device::device_type"],[46,6,1,"_CPPv4N14torch_tensorrt6Device8dla_coreE","torch_tensorrt::Device::dla_core"],[46,6,1,"_CPPv4N14torch_tensorrt6Device6gpu_idE","torch_tensorrt::Device::gpu_id"],[17,4,1,"_CPPv4N14torch_tensorrt16EngineCapabilityE","torch_tensorrt::EngineCapability"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability15kDLA_STANDALONEE","torch_tensorrt::EngineCapability::kDLA_STANDALONE"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability7kSAFETYE","torch_tensorrt::EngineCapability::kSAFETY"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability9kSTANDARDE","torch_tensorrt::EngineCapability::kSTANDARD"],[47,1,1,"_CPPv4N14torch_tensorrt11GraphInputsE","torch_tensorrt::GraphInputs"],[47,6,1,"_CPPv4N14torch_tensorrt11GraphInputs15input_signatureE","torch_tensorrt::GraphInputs::input_signature"],[47,6,1,"_CPPv4N14torch_tensorrt11GraphInputs6inputsE","torch_tensorrt::GraphInputs::inputs"],[48,1,1,"_CPPv4N14torch_tensorrt5InputE","torch_tensorrt::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN2at6TensorE","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEv","torch_tensorrt::Input::Input"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN2at6TensorE","torch_tensorrt::Input::Input::tensor"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,6,1,"_CPPv4N14torch_tensorrt5Input5dtypeE","torch_tensorrt::Input::dtype"],[48,6,1,"_CPPv4N14torch_tensorrt5Input6formatE","torch_tensorrt::Input::format"],[48,6,1,"_CPPv4N14torch_tensorrt5Input9max_shapeE","torch_tensorrt::Input::max_shape"],[48,6,1,"_CPPv4N14torch_tensorrt5Input9min_shapeE","torch_tensorrt::Input::min_shape"],[48,6,1,"_CPPv4N14torch_tensorrt5Input9opt_shapeE","torch_tensorrt::Input::opt_shape"],[48,6,1,"_CPPv4N14torch_tensorrt5Input5shapeE","torch_tensorrt::Input::shape"],[48,6,1,"_CPPv4N14torch_tensorrt5Input13tensor_domainE","torch_tensorrt::Input::tensor_domain"],[2,1,1,"_CPPv4N14torch_tensorrt12TensorFormatE","torch_tensorrt::TensorFormat"],[2,2,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatE5Value","torch_tensorrt::TensorFormat::TensorFormat"],[2,2,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEN2at12MemoryFormatE","torch_tensorrt::TensorFormat::TensorFormat"],[2,2,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEv","torch_tensorrt::TensorFormat::TensorFormat"],[2,3,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatE5Value","torch_tensorrt::TensorFormat::TensorFormat::t"],[2,3,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEN2at12MemoryFormatE","torch_tensorrt::TensorFormat::TensorFormat::t"],[2,4,1,"_CPPv4N14torch_tensorrt12TensorFormat5ValueE","torch_tensorrt::TensorFormat::Value"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value13kChannelsLastE","torch_tensorrt::TensorFormat::Value::kChannelsLast"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value11kContiguousE","torch_tensorrt::TensorFormat::Value::kContiguous"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value8kUnknownE","torch_tensorrt::TensorFormat::Value::kUnknown"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value13kChannelsLastE","torch_tensorrt::TensorFormat::kChannelsLast"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value11kContiguousE","torch_tensorrt::TensorFormat::kContiguous"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value8kUnknownE","torch_tensorrt::TensorFormat::kUnknown"],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormatcv5ValueEv","torch_tensorrt::TensorFormat::operator Value"],[2,2,1,"_CPPv4N14torch_tensorrt12TensorFormatcvbEv","torch_tensorrt::TensorFormat::operator bool"],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormatneE12TensorFormat","torch_tensorrt::TensorFormat::operator!="],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormatneEN12TensorFormat5ValueE","torch_tensorrt::TensorFormat::operator!="],[2,3,1,"_CPPv4NK14torch_tensorrt12TensorFormatneE12TensorFormat","torch_tensorrt::TensorFormat::operator!=::other"],[2,3,1,"_CPPv4NK14torch_tensorrt12TensorFormatneEN12TensorFormat5ValueE","torch_tensorrt::TensorFormat::operator!=::other"],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormateqE12TensorFormat","torch_tensorrt::TensorFormat::operator=="],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormateqEN12TensorFormat5ValueE","torch_tensorrt::TensorFormat::operator=="],[2,3,1,"_CPPv4NK14torch_tensorrt12TensorFormateqE12TensorFormat","torch_tensorrt::TensorFormat::operator==::other"],[2,3,1,"_CPPv4NK14torch_tensorrt12TensorFormateqEN12TensorFormat5ValueE","torch_tensorrt::TensorFormat::operator==::other"],[37,2,1,"_CPPv4N14torch_tensorrt15dump_build_infoEv","torch_tensorrt::dump_build_info"],[35,2,1,"_CPPv4N14torch_tensorrt14get_build_infoEv","torch_tensorrt::get_build_info"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability15kDLA_STANDALONEE","torch_tensorrt::kDLA_STANDALONE"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability7kSAFETYE","torch_tensorrt::kSAFETY"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability9kSTANDARDE","torch_tensorrt::kSTANDARD"],[16,4,1,"_CPPv4N14torch_tensorrt7logging5LevelE","torch_tensorrt::logging::Level"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kDEBUGE","torch_tensorrt::logging::Level::kDEBUG"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kERRORE","torch_tensorrt::logging::Level::kERROR"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kGRAPHE","torch_tensorrt::logging::Level::kGRAPH"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level5kINFOE","torch_tensorrt::logging::Level::kINFO"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level15kINTERNAL_ERRORE","torch_tensorrt::logging::Level::kINTERNAL_ERROR"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level8kWARNINGE","torch_tensorrt::logging::Level::kWARNING"],[24,2,1,"_CPPv4N14torch_tensorrt7logging24get_is_colored_output_onEv","torch_tensorrt::logging::get_is_colored_output_on"],[22,2,1,"_CPPv4N14torch_tensorrt7logging18get_logging_prefixEv","torch_tensorrt::logging::get_logging_prefix"],[23,2,1,"_CPPv4N14torch_tensorrt7logging24get_reportable_log_levelEv","torch_tensorrt::logging::get_reportable_log_level"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kDEBUGE","torch_tensorrt::logging::kDEBUG"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kERRORE","torch_tensorrt::logging::kERROR"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kGRAPHE","torch_tensorrt::logging::kGRAPH"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level5kINFOE","torch_tensorrt::logging::kINFO"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level15kINTERNAL_ERRORE","torch_tensorrt::logging::kINTERNAL_ERROR"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level8kWARNINGE","torch_tensorrt::logging::kWARNING"],[26,2,1,"_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE","torch_tensorrt::logging::log"],[26,3,1,"_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE","torch_tensorrt::logging::log::lvl"],[26,3,1,"_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE","torch_tensorrt::logging::log::msg"],[27,2,1,"_CPPv4N14torch_tensorrt7logging24set_is_colored_output_onEb","torch_tensorrt::logging::set_is_colored_output_on"],[27,3,1,"_CPPv4N14torch_tensorrt7logging24set_is_colored_output_onEb","torch_tensorrt::logging::set_is_colored_output_on::colored_output_on"],[28,2,1,"_CPPv4N14torch_tensorrt7logging18set_logging_prefixENSt6stringE","torch_tensorrt::logging::set_logging_prefix"],[28,3,1,"_CPPv4N14torch_tensorrt7logging18set_logging_prefixENSt6stringE","torch_tensorrt::logging::set_logging_prefix::prefix"],[25,2,1,"_CPPv4N14torch_tensorrt7logging24set_reportable_log_levelE5Level","torch_tensorrt::logging::set_reportable_log_level"],[25,3,1,"_CPPv4N14torch_tensorrt7logging24set_reportable_log_levelE5Level","torch_tensorrt::logging::set_reportable_log_level::lvl"],[3,1,1,"_CPPv4I0EN14torch_tensorrt3ptq19Int8CacheCalibratorE","torch_tensorrt::ptq::Int8CacheCalibrator"],[3,7,1,"_CPPv4I0EN14torch_tensorrt3ptq19Int8CacheCalibratorE","torch_tensorrt::ptq::Int8CacheCalibrator::Algorithm"],[3,2,1,"_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator19Int8CacheCalibratorERKNSt6stringE","torch_tensorrt::ptq::Int8CacheCalibrator::Int8CacheCalibrator"],[3,3,1,"_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator19Int8CacheCalibratorERKNSt6stringE","torch_tensorrt::ptq::Int8CacheCalibrator::Int8CacheCalibrator::cache_file_path"],[3,2,1,"_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibratorcvPN8nvinfer115IInt8CalibratorEEv","torch_tensorrt::ptq::Int8CacheCalibrator::operator nvinfer1::IInt8Calibrator*"],[4,1,1,"_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE","torch_tensorrt::ptq::Int8Calibrator"],[4,7,1,"_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE","torch_tensorrt::ptq::Int8Calibrator::Algorithm"],[4,7,1,"_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE","torch_tensorrt::ptq::Int8Calibrator::DataLoaderUniquePtr"],[4,2,1,"_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb","torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator"],[4,3,1,"_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb","torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::cache_file_path"],[4,3,1,"_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb","torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::dataloader"],[4,3,1,"_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb","torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::use_cache"],[4,2,1,"_CPPv4N14torch_tensorrt3ptq14Int8CalibratorcvPN8nvinfer115IInt8CalibratorEEv","torch_tensorrt::ptq::Int8Calibrator::operator nvinfer1::IInt8Calibrator*"],[29,2,1,"_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE","torch_tensorrt::ptq::make_int8_cache_calibrator"],[29,7,1,"_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE","torch_tensorrt::ptq::make_int8_cache_calibrator::Algorithm"],[29,3,1,"_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE","torch_tensorrt::ptq::make_int8_cache_calibrator::cache_file_path"],[30,2,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator"],[30,7,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::Algorithm"],[30,7,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::DataLoader"],[30,3,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::cache_file_path"],[30,3,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::dataloader"],[30,3,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::use_cache"],[36,2,1,"_CPPv4N14torch_tensorrt10set_deviceEKi","torch_tensorrt::set_device"],[36,3,1,"_CPPv4N14torch_tensorrt10set_deviceEKi","torch_tensorrt::set_device::gpu_id"],[49,1,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpecE","torch_tensorrt::torchscript::CompileSpec"],[49,2,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecEN5torch3jit6IValueE","torch_tensorrt::torchscript::CompileSpec::CompileSpec"],[49,2,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorI5InputEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec"],[49,2,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorIN3c108ArrayRefI7int64_tEEEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec"],[49,2,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorINSt6vectorI7int64_tEEEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec"],[49,3,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorIN3c108ArrayRefI7int64_tEEEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec::fixed_sizes"],[49,3,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorINSt6vectorI7int64_tEEEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec::fixed_sizes"],[49,3,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecEN5torch3jit6IValueE","torch_tensorrt::torchscript::CompileSpec::CompileSpec::input_signature"],[49,3,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorI5InputEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec::inputs"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec19allow_shape_tensorsE","torch_tensorrt::torchscript::CompileSpec::allow_shape_tensors"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec10capabilityE","torch_tensorrt::torchscript::CompileSpec::capability"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec5debugE","torch_tensorrt::torchscript::CompileSpec::debug"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec6deviceE","torch_tensorrt::torchscript::CompileSpec::device"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec12disable_tf32E","torch_tensorrt::torchscript::CompileSpec::disable_tf32"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec20dla_global_dram_sizeE","torch_tensorrt::torchscript::CompileSpec::dla_global_dram_size"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec19dla_local_dram_sizeE","torch_tensorrt::torchscript::CompileSpec::dla_local_dram_size"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec13dla_sram_sizeE","torch_tensorrt::torchscript::CompileSpec::dla_sram_size"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec18enabled_precisionsE","torch_tensorrt::torchscript::CompileSpec::enabled_precisions"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec12graph_inputsE","torch_tensorrt::torchscript::CompileSpec::graph_inputs"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec14min_block_sizeE","torch_tensorrt::torchscript::CompileSpec::min_block_size"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec20num_avg_timing_itersE","torch_tensorrt::torchscript::CompileSpec::num_avg_timing_iters"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec14ptq_calibratorE","torch_tensorrt::torchscript::CompileSpec::ptq_calibrator"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec5refitE","torch_tensorrt::torchscript::CompileSpec::refit"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec24require_full_compilationE","torch_tensorrt::torchscript::CompileSpec::require_full_compilation"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec14sparse_weightsE","torch_tensorrt::torchscript::CompileSpec::sparse_weights"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec22torch_executed_modulesE","torch_tensorrt::torchscript::CompileSpec::torch_executed_modules"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec18torch_executed_opsE","torch_tensorrt::torchscript::CompileSpec::torch_executed_ops"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec24truncate_long_and_doubleE","torch_tensorrt::torchscript::CompileSpec::truncate_long_and_double"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec14workspace_sizeE","torch_tensorrt::torchscript::CompileSpec::workspace_size"],[31,2,1,"_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE","torch_tensorrt::torchscript::check_method_operator_support"],[31,3,1,"_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE","torch_tensorrt::torchscript::check_method_operator_support::method_name"],[31,3,1,"_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE","torch_tensorrt::torchscript::check_method_operator_support::module"],[32,2,1,"_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec","torch_tensorrt::torchscript::compile"],[32,3,1,"_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec","torch_tensorrt::torchscript::compile::info"],[32,3,1,"_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec","torch_tensorrt::torchscript::compile::module"],[34,2,1,"_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec","torch_tensorrt::torchscript::convert_method_to_trt_engine"],[34,3,1,"_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec","torch_tensorrt::torchscript::convert_method_to_trt_engine::info"],[34,3,1,"_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec","torch_tensorrt::torchscript::convert_method_to_trt_engine::method_name"],[34,3,1,"_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec","torch_tensorrt::torchscript::convert_method_to_trt_engine::module"],[33,2,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module"],[33,3,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module::device"],[33,3,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module::engine"],[33,3,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module::input_binding_names"],[33,3,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module::output_binding_names"],[70,8,0,"-","torch_tensorrt"]],"torch_tensorrt.Device":[[70,10,1,"","__init__"],[70,11,1,"","allow_gpu_fallback"],[70,11,1,"","device_type"],[70,11,1,"","dla_core"],[70,11,1,"","gpu_id"]],"torch_tensorrt.Input":[[70,10,1,"","__init__"],[70,11,1,"","dtype"],[70,10,1,"","example_tensor"],[70,11,1,"","format"],[70,10,1,"","from_tensor"],[70,10,1,"","from_tensors"],[70,11,1,"","shape"],[70,11,1,"","shape_mode"]],"torch_tensorrt.fx":[[67,9,1,"","InputTensorSpec"],[67,9,1,"","TRTInterpreter"],[67,9,1,"","TRTInterpreterResult"],[67,9,1,"","TRTModule"],[67,12,1,"","compile"]],"torch_tensorrt.logging":[[68,9,1,"","Level"],[68,9,1,"","debug"],[68,9,1,"","errors"],[68,12,1,"","get_is_colored_output_on"],[68,12,1,"","get_logging_prefix"],[68,12,1,"","get_reportable_log_level"],[68,9,1,"","graphs"],[68,9,1,"","info"],[68,9,1,"","internal_errors"],[68,12,1,"","log"],[68,12,1,"","set_is_colored_output_on"],[68,12,1,"","set_logging_prefix"],[68,12,1,"","set_reportable_log_level"],[68,9,1,"","warnings"]],"torch_tensorrt.logging.Level":[[68,11,1,"","Debug"],[68,11,1,"","Error"],[68,11,1,"","Graph"],[68,11,1,"","Info"],[68,11,1,"","InternalError"],[68,11,1,"","Warning"]],"torch_tensorrt.ptq":[[69,9,1,"id1","CacheCalibrator"],[69,9,1,"id2","CalibrationAlgo"],[69,9,1,"id0","DataLoaderCalibrator"],[69,12,1,"","get_batch"],[69,12,1,"","get_batch_size"],[69,12,1,"","get_cache_mode_batch"],[69,12,1,"","read_calibration_cache"],[69,12,1,"","write_calibration_cache"]],"torch_tensorrt.ptq.CacheCalibrator":[[69,10,1,"","__init__"]],"torch_tensorrt.ptq.CalibrationAlgo":[[69,11,1,"","ENTROPY_CALIBRATION"],[69,11,1,"","ENTROPY_CALIBRATION_2"],[69,11,1,"","LEGACY_CALIBRATION"],[69,11,1,"","MINMAX_CALIBRATION"]],"torch_tensorrt.ptq.DataLoaderCalibrator":[[69,10,1,"","__init__"]],"torch_tensorrt.ts":[[71,12,1,"","TensorRTCompileSpec"],[71,12,1,"","check_method_op_support"],[71,12,1,"","compile"],[71,12,1,"","convert_method_to_trt_engine"],[71,12,1,"","embed_engine_in_new_module"]],torch_tensorrt:[[70,9,1,"","Device"],[70,9,1,"","DeviceType"],[70,9,1,"","EngineCapability"],[70,9,1,"","Input"],[70,9,1,"","TensorFormat"],[70,12,1,"","compile"],[70,12,1,"","convert_method_to_trt_engine"],[70,9,1,"","dtype"],[70,12,1,"","dump_build_info"],[67,8,0,"-","fx"],[70,12,1,"","get_build_info"],[68,8,0,"-","logging"],[69,8,0,"-","ptq"],[70,12,1,"","set_device"],[71,8,0,"-","ts"]]},objnames:{"0":["c","macro","C macro"],"1":["cpp","class","C++ class"],"10":["py","method","Python method"],"11":["py","attribute","Python attribute"],"12":["py","function","Python function"],"2":["cpp","function","C++ function"],"3":["cpp","functionParam","C++ function parameter"],"4":["cpp","enum","C++ enum"],"5":["cpp","enumerator","C++ enumerator"],"6":["cpp","member","C++ member"],"7":["cpp","templateParam","C++ template parameter"],"8":["py","module","Python module"],"9":["py","class","Python class"]},objtypes:{"0":"c:macro","1":"cpp:class","10":"py:method","11":"py:attribute","12":"py:function","2":"cpp:function","3":"cpp:functionParam","4":"cpp:enum","5":"cpp:enumerator","6":"cpp:member","7":"cpp:templateParam","8":"py:module","9":"py:class"},terms:{"0":[33,43,44,45,49,52,55,58,60,61,63,64,66,67,68,69,70,71,72,74,75,81,82,83,84,85,87,89,90,92,93],"000":[82,83,84],"0000":76,"01":[61,66,76],"0208":61,"03":76,"0358":61,"0383":61,"04":[61,87],"0435":61,"0464":61,"0530":61,"0678":61,"0805":61,"0818":61,"0932":61,"0x7f6412476370":71,"1":[3,4,33,44,45,48,49,52,54,55,57,60,61,62,63,64,66,67,68,69,70,71,72,73,75,76,79,83,84,86,88,89,90,92,93],"10":[49,61,64,67,71,79,86,87,88,90],"100":[67,89],"1000":87,"1012":54,"1013":54,"1024":[52,70,71,86],"1045":61,"1048576":[45,49,71],"1056":61,"1063":61,"1073741824":[45,49,71],"109":61,"11":[54,61,63,64,75,79,87],"119":88,"12":[54,55,61,75,79,87,88],"120":[61,88],"123":76,"129":88,"13":[75,79],"136":87,"137":88,"138":88,"14":[79,84,87],"1409":90,"15":[75,79],"1502":61,"1549":61,"1556":90,"16":[61,62,70,79,88],"1691":61,"17":79,"18":[61,79],"19":[76,79],"1994":90,"1b":63,"1d":54,"1e":52,"1fec519":75,"2":[33,43,55,60,61,64,66,68,69,70,71,73,75,76,79,81,82,84,85,88,89,90],"20":[55,79,83,84],"2009":90,"2010":90,"2012":76,"2014":90,"2015":63,"2017":63,"2019":63,"2020":[61,65],"2022":63,"2023":90,"2048":[67,89],"2052":[82,83,84],"22":87,"224":[55,67,70,71,83,86,87],"225":[67,87],"229":87,"23":[49,54,71,76],"234375":87,"24":54,"244":[70,71],"248":54,"249":54,"25":[61,67,89],"256":87,"258":75,"27":[55,61],"28":61,"2802":61,"2822":75,"287":75,"29":61,"2c3":76,"3":[45,49,52,54,55,57,61,64,66,68,69,70,71,75,76,79,83,86,88,89,90,92,93],"30":[83,84],"300":[52,92],"31":61,"32":[52,61,62,70,88,90,93],"320":90,"32bit":52,"33":61,"33554432":[67,89],"346":61,"35":61,"36":61,"3677":54,"37":61,"38":88,"39":88,"3d":89,"4":[55,57,61,64,66,68,73,75,76,79,82,84,89],"406":87,"429688":87,"4465":90,"456":87,"468750":87,"4822":90,"485":87,"4914":90,"5":[52,55,57,58,61,63,64,68,75,76,79,82,87,88,89],"50":86,"512":[52,70,71,86],"523438":87,"53":76,"536870912":[45,49,71],"539":61,"56":61,"576":61,"6":[54,55,57,61,64,66,70,79,88],"622":54,"64":[62,70,89],"64bit":52,"664062":87,"7":[55,57,58,61,63,79,82,83,84],"72048":64,"7302":76,"8":[3,52,54,61,63,64,70,75,76,79,83,87],"8000":87,"8001":87,"8002":87,"84":[61,88],"9":[61,79,87],"90":87,"92":87,"9223372036854775807":66,"96":63,"abstract":[57,60,76],"boolean":[70,89],"break":[75,89],"byte":[69,70,71,86],"case":[0,1,2,46,49,53,55,57,60,64,70,89,90,91],"catch":[54,61],"char":[3,4,44,52,61],"class":[29,30,44,45,46,51,57,60,61,62,68,71,75,76,82,86,88,89,90],"const":[0,1,2,3,4,29,30,31,32,33,34,36,44,45,46,54,60,61,66,90],"default":[0,1,2,3,4,16,29,30,33,43,45,46,48,49,52,55,61,62,64,67,70,71,73,74,75,89,90,92],"do":[53,54,55,60,61,62,74,76,88,89,90,93],"enum":[0,1,2,42,45,46,68,71,90],"export":64,"final":[53,55,56,58,64,82,83,84,86],"float":[49,52,61,62,66,70,82,84,88,90,92],"function":[0,1,2,3,4,46,48,49,54,55,57,60,61,64,82,83,84,86,87,88,89,90,92,93],"import":[52,54,55,61,62,64,73,75,87,88,89,91,92],"int":[0,3,4,36,44,45,49,52,55,61,66,67,69,70,71,73],"long":[49,52,53,70,75,76],"new":[0,1,2,3,4,32,33,46,48,49,55,57,58,60,61,68,71,75,81,83,84,85,87,89],"public":[0,1,2,3,4,44,45,46,47,48,49,76,90],"return":[0,1,2,3,4,23,24,29,30,31,32,33,34,35,42,43,44,45,46,54,55,56,57,58,60,61,62,67,68,70,71,82,87,88,89,90],"short":[54,75,76],"static":[48,49,53,60,61,70,71,73],"super":[44,82,88],"throw":[52,54,61],"true":[0,1,2,4,46,49,54,55,60,61,66,67,70,71,73,76,82,83,84,87,89,90,92,93],"try":[58,61,75,76,92],"var":66,"void":[3,4,25,26,27,28,36,37,42,44,45],"while":[55,64,86,87,90],A:[4,29,30,32,33,47,48,54,55,60,64,67,70,71,76,87,89,90],And:61,As:[55,61,89],At:74,But:[61,75],By:[29,30,51,55,73,88],For:[53,55,61,64,67,73,75,76,82,86,87,88,89,90,91,92],If:[27,33,53,54,55,61,62,64,67,68,70,73,75,82,87,89,90,91,93],In:[0,1,2,46,53,55,56,57,58,60,62,64,65,75,76,78,81,85,86,87,89,90,91],Is:[24,70],It:[52,54,55,56,58,60,64,73,75,86,89],Its:[60,75],Not:3,On:55,One:[61,75,76,86,89],Or:75,THE:75,TO:61,That:75,Thats:61,The:[1,46,48,49,52,53,54,55,56,57,58,60,62,64,68,70,71,73,76,85,86,87,88,89,90,92],Then:[55,64,90,92],There:[4,53,58,60,63,64,76,86,87,88,89,90,91],These:[53,55,57,73,75,87,90],To:[1,46,52,55,61,62,64,73,87,88,92],Will:31,With:[61,73,75,87,90],_:[69,75,89],___torch_mangle_10:88,___torch_mangle_4847:57,___torch_mangle_5:88,___torch_mangle_9:88,__and__:66,__attribute__:43,__derive_index:66,__getitem__:66,__gnuc__:43,__init__:[69,70,75,82,88],__is__:66,__isnot__:66,__main__:[82,83,84],__name__:[82,83,84],__not__:66,__or__:66,__range_length:66,__round_to_zero_floordiv:66,__torch__:[57,61,88],__torch___pytorch_detection_ssd_src_model_ssd300_trt_engin:57,__torch___torchvision_models_resnet____torch_mangle_4847_resnet_trt_engin:57,__visibility__:43,__xor__:66,_all_:54,_c:[70,71,92],_convolut:[61,66],_devic:71,_dynamo:[82,83,84],_enum:70,_input:[70,71],_jit_to_backend:92,_jit_to_tensorrt:71,_rendered_examples_jupyt:85,_rendered_examples_python:85,_script:[70,71],_set:82,_shapemod:70,_theme:80,_validate_not_a_forked_repo:87,a1b:76,aarch64:58,ab:66,abi:91,abl:[53,54,60,65,89,90,92],about:[52,53,57,60,61,64,70,73,87],abov:[25,55,61,64,68,74,75,83,84,89],absolut:52,ac:78,acc_mod:89,acc_norm:89,acc_op:89,acc_op_convert:89,acc_ops_sigmoid:89,acc_trac:89,acceler:[67,81,85,93],accept:[48,52,57,60,61,62,70,82],access:[54,60,61,65,73,89,92],accord:[60,71],accordingli:[73,89],account:87,accumsan:78,accumul:[49,71],accuraci:[86,90],achiev:[55,86],aco:66,acosh:66,acoust:86,acquir:61,across:[49,52,54,55,71,73],acthardtanh:60,action:[75,89],activ:[61,71,75,86,89,90,93],activationtyp:[60,89],actual:[54,57,60,61,68,88,89],ad:[25,52,53,55,89],adaptive_avg_pool1d:66,adaptive_avg_pool2d:66,adaptive_avg_pool3d:66,adaptive_max_pool1d:66,adaptive_max_pool2d:66,adaptive_max_pool3d:66,add:[26,53,54,55,60,61,62,63,64,66,68,73,75,80],add_:[54,61,66],add_activ:89,addactiv:60,addit:[54,61,63,70,86,89],addlay:61,address:76,addshuffl:61,adipisc:[76,78],adjac:[55,75],adjust:75,adopt:86,advanc:[76,81,85,90],advis:75,aenean:78,afford:89,aforement:87,after:[52,53,54,55,61,62,63,65,82,83,84,87,88,89,91],again:[44,57,60,75],against:[52,61],agx:45,ahead:61,aim:54,algo_typ:[69,90],algorithm:[3,4,29,30,44,69,89,90],algorithm_selector:89,alias:43,align:75,align_corn:66,aliquam:78,aliquet:[76,78],all:[16,42,43,44,45,49,52,54,55,57,61,62,63,64,68,70,75,76,85,86,87,88,89,90,91],alloc:60,allow:[48,49,52,53,54,55,63,70,71,73,83,84,89],allow_gpu_fallback:[45,46,70,71,90,92,93],allow_shape_tensor:[45,49,71],allow_tf32:66,almost:61,alpha:[66,76,89],alreadi:[52,53,54,61,90],also:[29,53,60,61,62,63,64,65,73,75,76,85,86,90],altern:[48,55,62,86],although:75,altogeth:[55,73],alwai:[3,4,27,52,75],amet:[76,78],an:[2,3,4,48,49,52,53,54,55,56,57,58,60,61,62,63,64,65,67,69,70,71,73,75,76,82,86,87,88,89,90,91],analogu:60,analysi:55,analyt:73,analytics_id:73,ancient:75,ani:[48,52,53,60,61,62,64,66,69,70,71,73,75,89,90],ann:75,annot:[60,61],anonym:75,anoth:[62,75,76,88],ant:78,anyon:76,anyth:[75,76,91],aot:[61,65],api:[55,58,60,61,62,70,71,74,81,82,85,86,87,89,90,91,92],appear:[55,75],append:66,appli:90,applic:[1,29,46,52,54,58,61,62,91,92,93],approach:55,appropri:63,apr:61,ar:[42,46,49,52,53,54,55,57,58,60,61,63,64,65,70,71,73,75,76,77,86,87,88,89,90,91,92],arab:76,arang:66,architectur:[64,65,86],archiv:64,arcu:[76,78],area:77,aren:61,arg:[53,61,69,70,79,86,89],arg_replacement_tupl:89,argc:61,argmax:66,argmin:66,argument:[48,52,54,57,60,61,62,70,71,75,76,89],argv:61,arithmet:55,around:[54,57,60,75,78,88],arrai:[3,4,33,53,71],arrayref:[45,48,49],artifact:63,arxiv:90,as_numpi:87,asin:66,asinh:66,aspect:52,assembl:[53,61],assign:[3,4,74],associ:[53,60,61],associatevalueandivalu:60,associatevalueandtensor:[60,61],assum:[33,92],atan:66,atanh:66,aten:[49,54,55,59,60,61,66,71,82],atol:52,attrdict:87,attribut:[54,55,57,61,75,89],auctor:78,audio:86,augu:78,author:76,auto:[44,55,60,61,75,76,90,93],autodoc:[75,76],automat:[61,75],avail:[52,60,64,73,89,93],averag:[49,52,71],avg:52,avg_pool1d:66,avg_pool2d:66,avg_pool3d:66,awai:75,awaken:75,axi:66,b0:86,b:[63,64,66,76,87],b_hh:66,b_ih:66,back:[54,55,57,58,61,70,75,88],back_insert:44,backend:[71,74,81,82,84,85,92],backend_kwarg:82,background:[75,88],backlink:75,backward:89,bar:[73,75],base:[37,50,57,62,64,67,68,69,70,75,84,86,88,90],bash:64,basi:75,basic:[52,55,76,85,87,89],batch:[3,4,44,67,83,84,87,89,90,93],batch_norm:[60,66],batch_siz:[44,90],batched_data_:44,batchnorm:54,batchtyp:44,bathroom:75,bazel:[58,64],bazel_vers:64,bazelbuild:64,bazelisk:64,bazelvers:64,bdist_wheel:64,beat:76,becaus:[60,61,64,67,88,89],becom:60,bee:75,been:[53,60,61,76],befor:[49,54,55,58,60,61,64,65,71,87,89],beforehand:61,begin:[44,64,75,82,89],beginn:88,begun:75,behav:77,behavior:[49,55,70,71,89],behind:75,being:[61,89],belong:75,below:[33,55,60,61,62,64,75,87,89],benchmark:66,benefit:[60,61],bert:84,bertmodel:84,besid:75,best:[64,75,89],beta:[66,71,89],better:[86,88],between:[54,55,60,64,75,76,90],bia:[54,61,66],bibendum:78,bibliograph:76,bigger:75,bin:64,binari:[44,90],binary_data:87,bind:[3,4,33,44,71,75],bird:87,bit:[49,60,61,70,71,89],bitbucket:73,bitbucket_url:73,bitwise_not:66,blandit:78,blank:75,blob:[59,73,90],block0:54,block1:54,block:[52,53,54,55,79],blue:75,bmm:66,bodi:[75,76],bold:75,bool:[0,1,2,3,4,24,27,30,31,42,44,45,46,49,54,60,61,66,67,68,70,71,73,90],border:75,both:[55,64,67,73,75,88,90],bottom:73,bound:70,boundari:[55,68,69],box:75,bracket:75,branch:64,bread:75,brief:55,briefli:88,brontosaurus:75,browser:75,bsd:[42,43,44,45],buffer:[3,4,89],bug:64,bui:76,build:[29,30,35,49,52,53,56,58,60,61,70,74,79,83,84,89,90],build_fil:64,build_model:89,buildcommandarg:63,builddirectori:63,builder:89,builderconfig:45,buildroot:63,built:[33,52,57,58,64,71],builtin:89,button:[73,75],bytearrai:[71,89],c10:[0,1,45,46,48,49,61,90],c:[42,43,44,45,52,58,62,63,66,67,76,87,91,93],c_api:59,c_str:[60,61],cach:[3,4,29,30,44,52,61,67,69,89,90],cache_:44,cache_fil:[44,69,90],cache_file_path:[3,4,29,30,44],cache_file_path_:44,cache_size_:44,cachecalibr:[69,90],cackl:76,calcul:[48,53,55,61],calibr:[3,4,29,30,44,49,52,61,69,71,90],calibration_cache_fil:[29,30,90],calibration_dataload:[30,90],calibration_dataset:90,calibrationalgo:[69,90],call:[29,30,32,49,54,57,60,61,67,71,75,82,83,84,86,88,89,92],call_funct:89,callabl:70,callmethod:88,can:[0,1,4,29,30,34,46,47,48,49,52,53,54,55,56,57,58,60,61,62,64,70,71,73,75,81,82,83,84,85,86,87,88,89,90,91,92],canada:76,cannot:[48,54,55,64,70,71,74,88,89],canon:73,canonical_url:73,capabl:[17,45,49,52,57,70,71,92],capit:75,caption:[75,78],cast:[3,4,54],cat:[55,64,66],caught:54,caus:[60,64,73,82,83,84],cd:[64,87],cdll:61,ceil:66,ceil_mod:66,cell:76,centercrop:87,cerr:61,certain:[64,82,89],cfg:55,chain:60,challeng:87,chanc:60,chang:[29,54,55,58,63,71,73,87,89,90],changelog:79,channel:[2,70,74],channel_last:[70,71,86],channels_last:70,charact:75,check:[0,1,31,46,52,54,60,61,64,71,87,89,91],check_method_op_support:71,check_method_operator_support:[41,45,50],checkmethodoperatorsupport:61,child:76,children:89,choic:[64,69],choos:[88,89],cifar10:90,cifar:90,clamp:66,clamp_max:66,clamp_min:66,class_count:87,classif:[61,86,88],classifi:[76,86],classification_index:87,classmethod:70,clean:[55,63,75,82,83,84],cleanli:55,clear:44,cli:[52,62],click:63,clickabl:75,clone:[63,66],close:61,closer:54,closet:75,cmake:63,cmake_build_typ:63,cmake_cuda_flag:63,cmake_cxx_flag:63,cmake_module_path:63,cmakecommandarg:63,cmakeset:63,cnn:86,co:[66,76,86],code:[55,58,61,65,74,76,82,83,84,85,88,89,90],coeffici:86,collapse_navig:73,collat:76,collect:[55,61,62,71],colon:75,color:[24,27,68,75],colored_output_on:[27,42,68],column:76,com:[59,61,64,82,83,84,87,90,91],combin:[55,89],come:[64,74,87,89],command:[52,61,64,75,76,87,88],comment:[64,75],commodo:78,common:[53,54,67,75,89],common_subexpression_elimin:54,commonli:76,commun:[49,52,61,63,71],compar:[62,89],comparis:[0,2],comparison:[1,46],compat:[0,1,46,54,57,63,64,71,89],compil:[31,34,41,45,49,50,52,54,55,57,60,62,67,68,70,71,73,87,88,89,90,91,92,93],compilation_kwarg:84,compilationset:82,compile_engine_and_inf:[82,83,84],compile_spec:[90,93],compilegraph:[61,90],compilesepc:33,compilespec:[3,4,21,32,34,41,45,50,55,61,71,90,93],compilespecstruct:50,complet:[55,61,88],complex:[47,49,62,64,88],complianc:52,compliat:90,complic:64,compon:[56,58,88,91],compos:[87,88,89,90],composit:61,compound:86,comput:[49,75,86,89,90],conceiv:75,concept:85,concorr:87,condimentum:78,condit:[55,75],conf:[73,80],confidence_scor:87,config:[64,67,87,89],configur:[32,34,48,61,64,65,70,71,79,87,90],configurationtyp:63,configureset:63,congu:78,connect:[54,71,75,87,93],consectetur:[76,78],consecut:55,consid:[55,61,71],consider:87,consist:[54,75,89],consol:52,consolid:[55,88],constant:[53,54,55,61],constant_pad_nd:66,constexpr:[0,1,2,45,46],construct:[0,1,2,3,4,46,48,49,53,54,56,58,60,61,69,70,75,76,89,90],constructor:[0,2,46,48,49,57,88],consult:74,consum:[4,53,88],contact:76,contain:[30,31,52,53,54,55,60,61,64,67,70,75,76,87,88,89,90,91],content:[79,87,90],context:[53,56,57,58,68],contextnet:86,contigu:[2,48,49,52,70,71],continu:[75,89,91],contributor:61,control:[88,89],conv1:[61,88],conv2:[61,88],conv2d:88,conv:[49,52,61],conval:78,convect:48,conveni:[84,86,90],convent:33,converison:89,convers:[54,55,57,61,70,71,89],conversionctx:[60,61],convert:[3,4,31,32,34,52,54,55,56,58,62,65,70,71,83,84,86,91,92],convert_method_to_trt_engin:[41,45,50,70,71,92],convertgraphtotrtengin:61,convien:49,convienc:[3,4,49],convolut:[71,90,93],convtert:89,coordin:58,copi:[44,60,66,69,76,87,89],copy_:66,copyright:[42,43,44,45,61,76],core:[45,52,54,55,58,61,70,93],core_id:70,corpor:[42,43,44,45],corpu:86,correct:[57,64,73],correctli:64,correctness_atol:67,correctness_rtol:67,correspond:[60,64,89],cosh:66,could:[55,83,84,89],count_include_pad:66,coupl:[53,58,89,91],cout:61,cover:[85,86],cp:64,cpp:[14,15,42,43,44,45,51,54,58,61,64,90],cpp_frontend:90,cppdirectori:50,cppdoc:61,cpu:67,cra:78,creat:[29,30,33,52,53,55,57,60,61,65,71,75,87,89],credit:61,criteria:[55,56,58],cross:75,cs:90,csrc:[54,59],cstddef:90,ctestcommandarg:63,ctrl:63,ctx:[60,61],ctype:61,cuda113:64,cuda:[49,57,61,62,63,64,67,70,87,89,90,92],cuda_graph_batch_s:[67,89],cuda_runtim:[21,45],cudafloattyp:61,cudasetdevic:36,cudnn:63,cudnn_en:66,cudnn_root_dir:63,cumsum:66,curabitur:78,curl:[64,75],current:[23,55,57,60,63,64,67,71,73,89],cursu:78,custom:[52,64,81,85,89],custom_class:[21,45],custom_mapp:89,customclasshold:[45,48],cut:75,cxx11:91,d:[52,75,76,93],d_silence_experimental_filesystem_deprecation_warn:63,dapibu:78,data:[0,2,3,4,29,30,44,46,48,49,52,53,55,56,58,60,62,66,67,69,70,71,75,79,86,89,90],data_dir:90,data_item_1:74,data_typ:87,dataclass:[82,89],dataflow:[60,61],dataload:[4,29,30,44,49,69,90],dataloader_:44,dataloadercalibr:[69,90],dataloaderopt:90,dataloaderuniqueptr:[4,44],dataset:[29,69,86,90],datatyp:[1,21,38,45,46,48,49,50,62,70,71,87],datatypeclass:50,date:76,david:76,dbg:64,dcmake_build_typ:64,dcmake_module_path:64,dead_code_elimin:54,deal:60,debug:[16,27,45,49,52,60,63,68,71,82,83,84,92],debugg:[52,71],decid:70,declar:64,deconvolut:93,decor:89,dedic:[54,76],deep:[60,65,73,90,93],deeplearn:[59,89],def:[62,75,82,87,88,89],defin:[0,1,2,3,4,33,43,46,47,48,49,51,52,61,62,70,73,82,84,86,88,89,90],definit:[51,60,75],deiti:75,delet:[0,1,2,45,46,54],delimit:54,demo:[75,90],demonstr:[75,76,77,86,87,90],demostr:86,denot:75,dep:[63,64],depend:[29,35,53,58,61,62,63,87,89,91],depickl:57,deploi:[56,58,61,65,87,90],deploy:[52,61,62,86,87,90,91,93],deprec:[66,89],depth:[73,86],descclassnam:75,descnam:75,describ:[49,55,60,81,85,87,88,92],descript:[55,76],deseri:[61,70,71],design:[86,89,93],desir:[76,90],desktop:63,destini:76,destroi:[60,76],destructor:60,detail:[61,87,88,89,91],detect:[48,57],determin:[54,89],determinist:66,dev0:75,develop:[61,63,64,65,75,76,89],deviat:52,devic:[21,33,36,38,45,49,50,52,57,62,66,67,69,70,71,86,90,92,93],device_typ:[45,46,70,90,92,93],deviceclass:50,devicetyp:[21,38,45,46,50,70,71,90,92,93],devicetypestruct:50,diam:78,dict:[70,71],dictionari:[70,71,82,92],dictum:78,dictumst:78,did:75,didn:75,differ:[29,54,55,58,64,65,73,88,89],dignissim:78,dilat:66,dim0:66,dim1:66,dim:[66,67,87,89],dim_int:66,dim_intlist:66,dimens:[48,54,67,86,89],direct:[79,91],directli:[60,63,64,65,69,81,82,85,90],directori:[18,19,20,21,42,43,44,45,50,63,64,90],disabl:[52,68,73,74],disable_memory_format_check:70,disable_tf32:[45,49,71,90],disclos:64,disconnect:75,discret:75,discuss:87,displai:[52,68,73],display_github:73,display_gitlab:73,display_vers:73,dist:64,distdir:64,distribut:[61,70,90,91],div:[55,66],div_:66,div_lgamma:55,divisor_overrid:66,django:74,dl:75,dl_open:91,dla:[1,45,46,49,52,65,70,71],dla_cor:[45,46,52,70,90,92,93],dla_global_dram_s:[45,49,52,71],dla_local_dram_s:[45,49,52,71],dla_sram_s:[45,49,52,71],dla_standalon:52,dlacor:52,dll:52,do_not_merg:55,doc:[58,59,64,73,74,75,80],docker:87,docsrc:58,docstr:[62,75,76],document:[42,43,44,45,50,58,61,73,75,76,80,87,88,90,91,92],docutil:[75,76],doe:[43,44,54,55,60,75,83,84,89,90],doesn:[61,64,75,88],dolor:[76,78],domain:[48,70,76,90],don:[60,73,75,76,87,89,90],done:[53,55,58,87],donec:[76,78],dont:42,dothismethod:75,dotpai:74,dotpayprovid:74,doubl:[45,48,49,52,71,75],down:[64,73,89],download:[64,79,82,83,84,85,87,90],downstream:86,doxygen_should_skip_thi:[44,45],dpython:[70,71],dram:52,dream:76,driver:64,drop:[64,73],dt:75,dtensorrt_root:64,dtorch_dir:64,dtyep:67,dtype:[45,48,49,52,62,66,67,70,71,84,86,89],dual:75,due:[3,4,64,74,75],dui:[76,78],dump:[37,52,64],dump_build_info:[38,45,50,70],durat:75,dure:[49,52,55,60,69,86,90,91],dynam:[48,49,67,70,71,89],dynamic_batch:[67,89],dynamo:[82,83,84],e:[29,30,52,54,60,61,63,64,67,70,88,89,90],each:[3,4,49,53,54,55,57,60,61,64,67,73,75,89],ear:75,earli:89,eas:43,easi:[52,53,54,61,90],easier:[56,58,60,61,89,90],easiest:64,easili:[3,4],echo:75,edg:75,edit:[63,73],edu:90,effect:[54,61,73,86,89,90],effici:60,efficientnet:86,efficitur:78,eg:87,egesta:78,eget:78,either:[47,48,52,60,61,62,64,70,71,73,75,88],el:66,eleifend:76,element:[57,75,76,79,89],element_typ:44,elementum:78,elit:[76,78],elk:75,els:[43,44,48,71,75,76],elu:66,emb:[33,52,71,76],embed:[52,57,66,71,75,93],embed_engine_in_new_modul:[41,45,50,71],emit:53,emphasi:75,empti:[49,67,71,76,88],emum:[16,17],en:73,enabl:[3,4,24,49,52,55,56,58,67,68,69,71,73,83,84,89],enable_precis:61,enabled_precis:[45,49,61,62,70,71,82,83,84,87,90,92,93],enalbed_precis:93,encod:[57,86],encompass:71,encount:[55,64,82,83,84],encourag:87,end:[44,52,60,61,66,71,75,82,83,84,90],end_dim:[61,66],endif:[43,44,45],energi:75,enforc:61,engin:[0,1,17,32,33,34,45,46,48,49,52,53,55,56,58,61,62,65,67,70,71,73,83,84,90,91,92,93],engine_converted_from_jit:61,enginecap:[38,45,49,50,70,71,92],english:86,enhanc:75,enim:78,ensur:[29,54,55,63],enter:[53,70],entir:75,entiti:75,entri:[49,60],entropi:[29,30,90],entropy_calibr:69,entropy_calibration_2:[69,90],enumer:[0,1,2,16,17,46],environ:[87,89],ep:66,eq:[66,75],equat:75,equival:[32,56,58,60,61,71,83,84,88,90],equivil:34,erat:78,erf:66,eric:75,ero:78,error:[16,49,52,53,54,58,61,64,68,71,75,89],essenc:75,essenti:89,est:78,et:78,etc:[73,75,89,93],etiam:78,eu:78,euismod:78,eval:[61,62,82,83,84,87],evalu:[56,57,58],evaluated_value_map:[53,60],even:61,event:48,everi:[55,61,67],everyth:16,ex:[0,1,2,33,46,71,76,78],exact:87,examin:89,exampl:[48,55,57,58,60,61,62,63,65,68,70,71,73,74,76,79,81,82,83,84,85,87,88,89,90,91],example_tensor:70,exceedingli:75,except:89,exception_elimin:54,excerpt:76,excit:86,execpt:54,execut:[33,49,52,54,56,57,58,61,64,67,70,71,87,88,89,90],execute_engin:[57,61],exert:75,exeuct:57,exhaust:61,exist:[4,31,32,34,64,69,70,71,86,89,90],exit:[82,83,84,87],exp:66,expand:[54,66],expand_a:66,expanded_asset:64,expect:[48,54,60,61,62,70,86],experiment:[71,89],explain:89,explan:89,explic:44,explicit:[0,1,2,3,4,45,46,54,65,67,75,89,90],explicit_batch_dimens:[67,89],explicit_precis:67,explicitli:[55,56,58,62,71,90,92],explict:44,explictli:0,explor:85,expon:66,expos:90,express:75,ext:[75,76],extend:[56,58,60,61,63,66,86],extens:63,extent:[61,65],extern:[73,75],extra:61,extract:[61,86],extrem:75,ey:75,f16:[52,61,93],f32:52,f:[64,75,88,89],facilisi:78,fact:64,facto:75,factori:[4,29,30,90],fail:[61,93],fake_quantize_per_channel_affin:66,fake_quantize_per_tensor_affin:66,fall:70,fallback:[52,56,58,60,93],fals:[0,1,2,3,4,44,45,46,49,61,66,67,70,71,73,74,75,76,82,89,90,92],fame:78,familiar:87,far:[75,89],fashion:[61,86],fast:[49,52,71],faster:86,faucibu:78,fc1:[61,88],fc2:[61,88],fc3:[61,88],fc:[49,52,54],feat:[61,88],featur:[52,55,61,86,89,90,92],fed:[3,4,48],feed:[29,30,61],feedforward:86,feel:65,feli:78,feugiat:[76,78],few:[64,70,89],field:[3,4,67,70,90],fifth:76,figur:[55,76,78],file:[0,1,2,3,4,5,6,7,8,9,10,11,12,46,47,48,49,52,55,57,58,61,63,64,67,69,70,71,73,74,76,80,87,89,90],file_path:52,filepath:63,find:[4,61,64,89],finder:64,finibu:78,finish:89,first:[48,53,54,61,62,75,76,82,87,89,90],firstli:87,fit:75,fix:[49,75,89,93],fixed_s:[45,49],flag:[52,55,56,58,62,64,69,91],flaten:49,flatten:[45,47,61,66,88],flatten_convert:61,flesh:87,float16:[52,70],float32:[48,49,52,70,71,89],float64:71,float_int:66,floor:66,floor_divid:66,floordiv:66,flow:[60,75,86,88,89],flox:75,flush:75,fly:88,fold:76,folder:[63,89],follow:[33,52,55,57,61,63,64,71,73,75,76,80,81,85,86,87,88,89,90,91],foo:[75,76,89],foo_kwarg:89,foo_nod:89,forc:[52,71,73,89],force_fp32_output:89,forced_fallback_op:55,form:[53,62,70,75,87],format:[33,45,48,49,52,62,66,70,71,75,76,86,87],forth:76,forum:64,forward:[29,30,32,33,55,57,60,61,62,70,71,82,88,90,92],found:[42,43,44,45,61,64,75,90,91],four:[75,76],fp16:[0,48,49,52,61,62,65,67,89,93],fp32:[0,48,49,52,65,71,86,87,89,90],frac:75,freed:60,freeze_modul:54,friend:45,fringilla:78,from:[0,1,2,3,4,29,30,44,46,48,49,52,53,54,55,56,57,58,60,61,63,65,67,71,73,74,75,76,84,86,87,88,89,90],from_pretrain:84,from_tensor:[70,89],frontend:[62,65,81,83,84,85],fssl:64,fstream:[20,44],full:[45,49,52,60,61,68,82,83,84,87,90,91,93],fulli:[31,52,54,61,71,90,93],further:[55,89],fusc:78,fuse_addmm_branch:54,fuse_flatten_linear:54,fuse_linear:54,fusion:[60,89],futur:[71,89],fx2trt:67,fx2trt_exampl:89,fx:[62,65,70],g:[29,30,52,54,63,64,67,70,75,89,90],g_:75,galleri:[82,83,84,85],gamma:66,gatewai:74,gather:55,gaurd:43,gcc:[58,61],ge:66,gear:90,gener:[3,4,29,52,54,57,58,60,61,63,64,67,73,75,76,79,82,83,84,85,88,89,90],get:[0,1,2,3,4,23,35,44,46,54,55,60,61,64,68,70,86,87,89,90],get_batch:69,get_batch_impl:44,get_batch_s:69,get_build_info:[38,45,50,70],get_cache_mode_batch:69,get_is_colored_output_on:[39,42,50,68],get_logging_prefix:[39,42,50,68],get_output:89,get_reportable_log_level:[39,42,50,68],getattr:[54,57,61,88],getbatch:[3,4,44],getbatchs:[3,4,44],getdimens:[60,61],getoutput:[60,61],git:79,github:[59,61,64,73,82,83,84,87,90,91],github_url:73,gitlab:73,gitlab_url:73,give:[73,75,89],given:[48,49,52,54,61,62,67,69,70,71,88,89,92],global:[26,52,61],gnu:64,go:[44,54,55,61,65,82,83,84,86,87,88,89],goal:60,goe:[75,89],good:[44,60,75,89],goodger:76,googl:73,got:[61,75],gpu:[1,32,34,36,45,46,52,61,70,71,87,89,90,92,93],gpu_id:[36,45,46,52,70,71,90,92,93],graph:[16,31,32,34,45,49,52,53,55,56,58,60,61,65,67,68,71,83,84,86,88,89],graph_input:[45,49],graph_modul:[67,70],graphinput:[21,38,45,49,50],graphinputsstruct:50,graphmodul:[62,67,70],gravida:78,great:[61,75],greater:68,greedi:55,group:[66,75,76],grpc:87,gru_cel:66,gt:66,gtc:65,guangzhou:76,guarante:55,guard:54,guard_elimin:54,gui:75,guid:[74,85],gulf:87,gz:[75,76,90],h:[0,1,2,3,4,5,6,7,8,9,10,11,12,15,46,47,48,49,50,51,52,54,61,90],ha:[49,53,54,55,56,58,60,61,63,67,75,76,86,88,89,90],habit:78,habitass:78,hac:78,hack:44,hakaimagazin:87,half:[52,61,62,70,75,82,83,87,90,92,93],hand:87,handl:[54,55,57,89],happen:[88,89],hardtanh:[60,66],hardtanh_:66,hardwar:93,has_batch_dim:67,hash:64,have:[29,33,44,52,53,54,55,60,61,62,64,65,67,70,71,75,83,84,86,87,88,89,90],haven:61,header:[61,73,75,76,87],heart:76,heaven:75,heck:75,heh:76,hehe:76,height:75,help:[27,52,53,60,61,86,89,91],helper:60,hendrerit:78,here:[44,53,55,57,61,64,73,75,76,87,88,89,90,91],hermet:64,hexagram:75,hfile:50,hi:[66,75,76],hidden:[43,73],high:[48,54,55,73],higher:[54,73,75,88],highli:[86,87],highlight:75,hinton:90,hit:55,hold:[46,47,48,53,60,90],holder:[57,77],holi:75,home:64,hood:58,hope:76,host:[49,52,64,71,87],how:[3,4,64,75,77,79,82,86,87,88,91,92],howev:[29,64,73,74,87],html:[59,64,75,88,90],html_theme:80,html_theme_opt:73,html_theme_path:80,http:[59,61,64,73,75,82,83,84,86,87,88,90,91],http_archiv:64,httpclient:87,hub:87,huggingfac:86,human:75,humankind:76,hx:66,hybrid:71,hyperlink:75,hyphen:75,i8:52,i:[52,54,60,61,66,75,76,88,90],iaculi:78,icon:[73,75],id:[36,45,52,70,73,74,78,93],idea:[54,75],ident:52,identifi:55,idx:66,ifndef:[44,45],ifstream:44,ignor:70,iii:76,iint8calibr:[3,4,29,30,44,45,49,71,90],iint8entropycalibrator2:[3,4,29,30,44,90],iint8minmaxcalibr:[29,30,90],ilay:60,illustr:[86,89],imag:[87,90],imagenet:86,imagenett:86,images_:90,img1:87,img:87,img_path:87,imperdiet:78,implement:[3,4,54,55,57,61,74,89,90,91],implic:54,implicit:[66,67,75,89],implicitli:70,implictli:70,improv:76,in_shap:61,in_tensor:88,incas:44,includ:[13,15,16,35,37,42,43,44,45,51,52,55,56,57,58,61,64,67,73,75,81,85,88,89,90],includedirectori:50,includehidden:73,incompat:64,incorpor:76,indent:75,index:[33,59,65,66,71,73,79,90],indic:[66,73,75],indirect:75,individu:62,inetworkdefinit:53,infer:[54,61,70,71,81,82,85,86,89,90],inference_output:87,inferenceservercli:87,inferinput:87,inferrequestedoutput:87,info:[16,32,34,45,52,60,61,68,70],inform:[25,33,35,37,48,52,53,55,57,61,64,65,67,68,70,75,88,89,90,92],infrastructur:[87,90],ingest:58,inherit:[50,89,90],inheritenviron:63,initi:[75,82,83,84],injuri:75,inlin:[0,1,2,3,4,29,30,44,46,48,54,61,76,79],inner:[49,76,86],input0:[61,62],input1:[61,62],input2:61,input:[3,4,21,29,33,38,44,45,47,49,50,52,53,54,55,57,60,61,62,66,67,68,70,71,76,82,86,87,88,89,90,92,93],input_0:[57,61],input__0:87,input_binding_nam:[33,45,71],input_data:[62,88],input_file_path:[52,93],input_is_dynam:45,input_nam:[67,89],input_s:[55,61],input_scal:66,input_shap:[90,93],input_signatur:[45,47,49,62,71],input_spec:[52,67,89],input_tensor_spec:[67,70,89],input_v:89,inputclass:50,inputrang:[55,61],inputtensorspec:[67,70,89],insert:[61,90],inserting_befor:89,insid:[75,87],inspect:[60,61,88],instal:[61,65,79,87,91],installroot:63,instanc:[54,61,69,86,88],instance_norm:66,instanti:[56,57,58,60,61],instatin:[0,1,2,46],instead:[49,52,53,54,61,64,91],instnanti:57,instruct:[55,56,58,61,64,87,89],insur:64,int32:[70,71,84,86],int64:[0,70,71],int64_t:[45,46,48,49,90,93],int8:[0,44,48,49,52,65,70,71,90,93],int8_t:45,int8cachecalibr:[20,29,40,44,50],int8cachecalibratortempl:50,int8calibr:[3,20,30,40,44,50],int8calibratornamespac:50,int_float:66,integ:[70,78],integr:[65,82],intend:[64,82,83,84],intent:[54,75],interact:[75,82,83,84],interdum:78,interest:[54,75],interfac:[0,1,2,46,57,58,60,90],interfer:75,intermedi:[16,49,52,68,71,88],intern:[1,16,46,60,61,68,75],internal_error:68,internalerror:68,interpol:75,interpret:[57,75,89],interv:70,intro_to_torchscript_tutori:88,introduc:[86,89],invoc:82,invok:[61,88,89],io:[44,87],iostream:[20,21,44,45,61],ipso:75,ipsum:[76,78],ipynb:[82,83,84],ir:[56,58,60,62,70,82,83,84,88],is_aten:67,is_floating_point:66,is_train:90,iscustomclass:60,ishap:49,ishapelay:71,isinst:89,isn:[73,75],issu:[3,4,61,64,82,83,84],istensor:60,istream_iter:44,it_:44,ital:75,item:[74,76],itensor:[53,60,61,89],iter:[20,44,49,52,53,69,70,71],its:[29,53,55,57,60,64,75],itself:[0,1,2,46,52,54,64,87,92],iv:76,ivalu:[45,47,49,53,57,60,61],jan:76,jetpack:64,jetpack_5:64,jetpack_x:64,jetson:86,jit:[31,32,33,34,45,47,49,52,53,54,55,56,57,58,59,60,61,62,70,71,87,88,92],jp_workspac:64,jpg:87,json:63,jump:87,jupyt:[82,83,84,85],just:[44,45,54,55,61,62,65,68,75,77,86,88,89,91,92],justo:[76,78],k:[66,90],kbool:[0,45],kchannelslast:[2,45],kchar:[0,45],kclip:60,kcontigu:[2,45,48],kcpu:[1,46],kcuda:[1,46,55,61],kdebug:[16,42,44],kdla:[1,45,46,93],kdla_standalon:[17,45],keepdim:66,kei:[75,87,88],kept:76,kernel:[48,49,52,60,70,71,89],kernel_s:66,kerror:[16,42],keyboard:75,keyword:[70,71,82,84],kf16:[90,93],kfloat:[0,45,49],kgpu:[1,45,46],kgraph:[16,42,54],khalf:[0,45,61],ki8:90,kind:[53,89],kinfo:[16,42,44],kint:[0,45],kinternal_error:[16,42],klong:[0,45],know:[42,60,73,75],knowledg:75,kriz:90,krizhevski:90,ksafeti:[17,45],kstandard:[17,45,49],ktest:90,ktrain:90,kunknown:[0,2,45],kwarg:[69,70,86,89],kwarn:[16,42],l:66,label:[75,86,87,90],lacinia:78,lack:[55,56,58,89],lacu:78,laid:61,lambda:[60,61,75,87],lang:74,languag:[74,75,76,87,88],laoreet:78,larg:[56,58,61,73,75,86,90],larger:[55,73,86],largest:66,last:[2,54,70,89],lastli:87,later:[29,61],latest:[63,64,73],launch:87,layer:[46,49,52,53,54,60,61,71,86,87,89,90,93],layer_norm:66,layout:[2,48,66,70,71],ld_library_path:64,ld_preload:91,ldd:64,le:66,lead:75,leader:75,leaky_relu:66,leaky_relu_:66,learn:[61,65,87,90,93],leas:76,least:[75,76],leav:54,lectu:[76,78],left:[73,75],legacy_calibr:69,legend:75,len:66,lenet:[61,88],lenet_script:[61,88],lenetclassifi:88,lenetfeatextractor:88,length:[3,4,44,66,76,89],leo:78,let:[46,52,54,60,70,71,73,75,86,87,89],letter:[76,86],level:[23,25,26,39,42,44,50,54,55,58,68,71,79,87,88,89],levelnamespac:50,leverag:[81,85,89,90],lgamma:55,lib:[54,61,63,64],libero:[76,78],librari:[35,42,43,44,45,52,56,57,58,60,61],libtorch:[4,37,60,61,63,64,90],libtorch_pre_cxx11_abi:64,libtorchtrt:[52,61,64],libtorchtrt_plugin:91,libtorchtrt_runtim:91,licens:[42,43,44,45,61,63],light:75,ligula:78,like:[52,53,54,57,60,61,62,64,74,75,87,88,89,90,91],limit:[54,68,74,90],line:[52,61,76],linear:[2,55,66,70,88],link:[52,53,61,65,73,74,79,91],linux:[58,61,64],list:[18,19,20,21,31,49,51,53,55,57,60,61,62,64,66,67,69,70,71,79,87,89],listconstruct:[53,55,57,61],listunpack:[57,61],liter:76,literal:76,literal_block:75,live:[60,75],ll:89,lo:66,load:[52,55,57,61,62,69,71,86,87,89,90,91,92],load_librari:91,loading_data_recip:90,loborti:[76,78],local:[52,54,61,73],localhost:87,locat:[64,90],lock:74,log:[15,16,19,20,38,44,50,51,54,60,65,66,67,70,83,84,89],log_debug:60,logger:68,logger_level:67,loggingenum:50,logic:89,login:87,logist:89,loglevel:68,logo_onli:73,lone:76,longer:[73,91],look:[53,54,87,88,90,92],loop:[55,89],loop_unrol:54,lorem:[76,78],lose:73,loss:[86,90],lot:60,low:[48,89],lower:[16,67,68,70,76,83,84,86,89],lower_exampl:89,lower_graph:54,lower_precis:[67,89],lower_tupl:54,loweralltupl:54,lowerprecis:[67,89],lowersimpletupl:54,lstm_cell:66,lt:66,ltorchtrt:91,luctu:78,lvl:[25,26,42],m:76,machin:[57,87,90],macro:[5,6,7,8,9,10,11,12,15,18,20,21,42,44,45,50,51],mad:75,made:[54,56,58,75],maecena:78,magna:78,mai:[53,55,57,58,61,62,71,75,76,82,83,84,87,88,89,90],main:[54,55,56,57,58,60,61,73,75,77,89],mainli:89,maintain:[55,57,60],major:[58,89],make:[53,61,62,64,75,77,81,85,86,87,89,90,93],make_data_load:[4,90],make_int8_cache_calibr:[40,44,50,90],make_int8_calibr:[29,40,44,50,90],malesuada:78,man:[75,76],manag:[49,52,53,56,58,60,61,63,68,70,71],mangag:54,mani:[55,73,75,76,89],mantissa:[49,71],manual:[74,75,89],map:[1,46,53,54,56,58,60,61,82,86,87,89,90,92],mapper:89,mark:[54,55,73],marknodesforfallback:54,markup:[76,79],markup_process:75,mask:66,masked_fil:66,massa:78,master:[59,64,90,91],mat2:66,match:[54,64],math:79,matmul:[54,61,66],matrix:59,matter:89,matti:76,matur:58,mauri:[76,78],max:[48,52,60,66,70,73],max_batch_s:[67,87,89],max_c:52,max_h:52,max_input_shap:67,max_n:52,max_pool1d:66,max_pool2d:[61,66,88],max_pool3d:66,max_shap:[45,48,62,70,71,86,89],max_val:[60,66],max_w:52,max_workspace_s:[67,89],maximu:78,maximum:[48,49,52,67,71,83,84,87,89],mayb:75,mb:52,md:59,me:[75,76],mean:[55,60,65,66,67,82,87,89],mechan:[60,86,89],medium:75,meet:70,member:[46,47,48,49,70],memeori:2,memori:[20,21,44,45,54,60,61,62,70,71],memory_format:[66,70],memoryformat:[2,45],men:75,mental:75,menu:[52,73,75],menuselect:75,merg:55,messag:[16,25,26,52,68],meta:[79,89],metadata:[49,52,57,60,71,73],meth:75,method:[31,32,33,34,48,52,54,60,61,64,70,71,75,86,88,92],method_nam:[31,34,45,52,61,70,71],metu:78,mi:78,microsoft:63,middl:75,might:[54,64,73],min:[48,52,60,66,70],min_acc_module_s:67,min_block_s:[45,49,55,71,82,83,84],min_c:52,min_h:52,min_input_shap:67,min_n:52,min_shap:[45,48,62,70,71,86,89],min_val:[60,66],min_w:52,mind:75,mine:75,minim:[67,90],minimum:[48,49,52,55,68,71],minmax:[29,30,90],minmax_calibr:69,minor:63,minut:[82,83,84],misbuild:73,miss:[61,75],mkdir:64,mm:87,mmb:75,mobilenet_v2:92,mobilenetv2:86,mod:[52,55,61,79,89,90],mode:[62,89,90],mode_:90,model:[52,55,57,61,62,65,67,68,81,85,88,90,92],model_half:82,model_nam:87,model_repositori:87,model_torchtrt:68,model_trt:68,modifi:[55,76,89],modul:[31,32,33,34,45,49,52,55,56,57,58,60,62,63,64,65,67,68,69,70,71,74,75,76,82,86,89,90,92,93],modular:61,module_fallback:54,module_nam:52,molesti:78,momentum:66,morbi:78,more:[53,61,62,64,65,70,73,76,83,84,87,88,89,90,91,92],most:[58,64,67,87,89,91],mother:75,motion:75,mous:75,move:[30,44,54,57,61,71,90],ms:63,msg:[26,42,68],msvc:63,msvc_x64_x64:63,mu:75,much:[60,73,75,90],mul:[55,66],mul_:66,multi:52,multipl:[57,75,76,87,90],multipli:[49,71],must:[33,48,49,52,54,55,60,61,64,67,70,71,75,76,89,91],mutil:76,my:75,my_pytorch_model:89,myclass:75,mymodel:[55,62],myself:76,n:[52,60,61,90],nabla:75,nam:[76,78],name:[3,4,31,33,34,44,55,57,60,61,63,64,67,68,69,70,71,75,76,87,88,89,92],namedtupl:89,namespac:[42,43,44,45,51,54,65,90],narrow:66,nativ:[58,59,61],native_funct:59,natur:75,nav:[73,79],navig:73,navigation_depth:73,nbbind:[3,4,44],nchw:[2,70,71],ne:[54,66],nec:78,necessari:[42,91],need:[0,1,2,25,29,43,46,53,54,60,61,62,64,67,75,86,87,89,90,91],neg:66,negative_slop:66,nequ:[76,78],nest:[45,49,50,75,76],net:[60,61,75,76],netu:78,network:[29,30,60,61,86,87,89,90,93],neural:93,new_batch_size_input:83,new_batch_size_output:83,new_input:[83,84],new_lay:60,new_local_repositori:64,new_output:[83,84],new_siz:90,newer:64,next:[3,4,53,57,67,73,75,76,82,87,90],ngc:[64,87],nhwc:[2,52,70],nibh:[76,78],nice:64,nickel:75,night:76,nightli:89,ninja:[63,64],nisi:78,nisl:78,nlp:[29,30,90],nn:[54,59,61,62,67,70,71,82,88,89],node:[54,55,56,58,60,61,67,86,89],node_info:[60,61],noexcept:[44,90],noexceptoverrid:[3,4],non:[76,78],non_block:66,none:[60,66,67,68,69,70,71,73,75,82,89],nonetheless:75,nonexist:75,norm:66,normal:[0,1,2,46,61,75,87,88,89,90,93],normalized_shap:66,noskipw:44,notat:70,notatemoduleforfallback:54,note:[1,46,48,60,61,63,64,70,73,75,89,93],notebook:[58,65,82,83,84,85],now:[54,55,58,60,61,64,75,89,92],np:87,nu:75,nulla:78,nullptr:[44,45,49],num:52,num_avg_timing_it:[45,49,71,92],num_it:52,num_op:52,num_work:90,number:[3,4,49,52,54,55,60,61,62,67,70,71,73,81,83,84,85,86,89],numel:66,numer:[52,76,89],numpi:87,nunc:78,nvcr:87,nvidia:[32,34,42,43,44,45,52,59,61,64,70,71,82,83,84,87,89,93],nvinfer1:[3,4,29,30,44,45,49,60,90],nvinfer:[20,44],o:[64,75,87],obj:66,object:[0,1,2,3,4,46,48,49,52,57,60,68,69,70,71,90,92],obtain:86,obvious:88,occasion:[82,83,84],odio:[76,78],off:[55,57],offici:64,ofstream:[44,61],often:75,oh:76,ok:[61,75],okai:49,older:58,onc:[42,43,44,45,53,54,55,57,87,89,90,91],one:[47,54,60,61,62,68,70,75,82,83,84,87,88,89],ones:[42,55,56,58,61,64,75],onli:[1,3,4,16,29,44,46,48,52,54,55,58,60,64,67,68,70,75,89,90,91,93],onnx:54,onto:[52,57],op:[52,53,54,55,56,58,60,61,70,82,91],op_and_target:89,op_nam:52,open:[63,86,87],oper:[0,1,2,3,4,31,44,45,46,49,52,53,54,55,56,57,58,60,62,65,70,71,83,84,89,90,93],oppos:71,opset:[56,58],opt:[48,64,70],opt_c:52,opt_h:52,opt_n:52,opt_shap:[45,48,62,70,71,86],opt_w:52,optim:[48,52,54,61,62,65,67,83,84,86,88,89],optimin:48,optimiz:88,optimization_level:82,optimization_profile_field:70,optimize_target_shap:89,optimized_input_shap:67,optimized_model:[82,83,84],optimized_model_custom:82,optimz:87,option:[44,48,52,55,56,58,63,64,69,70,71,75,79,82,89,90,91,93],orchestra:75,orci:78,order:[33,49,55,60,61,62,64,67,71,89],org:[59,61,64,73,75,88,90],organ:76,origin:[33,67,89],ornar:[76,78],os:45,ostream:45,other:[0,1,2,45,46,52,53,54,57,61,62,64,65,66,74,75,89,91],otherwis:[64,67,89,91],our:[55,58,61,87,88],out:[31,44,53,54,55,56,58,60,61,63,64,68,71,75,87],out_shap:61,out_tensor:[60,61],output0:54,output:[24,27,33,49,52,53,54,55,57,60,61,64,68,71,73,75,76,86,87,89],output__0:87,output_binding_nam:[33,45,71],output_file_path:[52,93],output_nam:[67,89],output_pad:66,output_s:66,outself:61,outsid:75,over:[56,58,75,87,89],overal:86,overrid:[29,30,44,70,89,90],overview:[59,65,82],own:[55,60,61,64,75,87],p:[52,61,66,87,93],packag:[52,54,61],pad:66,padding_idx:66,page:[65,77,79,87],pair:[54,60,64,75,86,90],pane:75,paragraph:[76,79],param:[69,74],paramet:[0,1,2,3,4,25,26,27,29,30,31,32,33,34,36,46,48,49,53,54,60,61,67,68,70,71,79,88,89],parent:[14,15,18,19,20,21],pars:[61,75],parser:75,part:[52,55,58,73,74,75,89],partial:[52,75],partit:54,partitioninfo:55,pass:[33,53,55,56,57,58,60,61,64,68,69,71,88,89,90],past:75,path:[4,13,14,15,29,30,52,61,63,64,69,70,87,88,89,90],path_to_torchtrt_root:64,pathwai:88,pattern:[60,61,70],payment:74,pbtxt:87,peephole_optimz:54,pellentesqu:78,peopl:75,pep:75,perforamnc:89,perform:[29,30,86,87,90],permit:75,permut:[66,89],persist:75,pharetra:78,phase:[16,60,61],phasellu:78,phi:75,philosoph:75,phrase:75,pi:75,pick:88,pickler:57,piec:86,pil:87,pin:74,pin_memori:66,pip3:64,pip:[64,87],pipelin:[52,93],piplein:61,pixel_shuffl:66,pl:74,place:[48,54,64,75,76,77,89,90],placerat:78,plan:[52,58],platea:78,platform:[45,52,58,63,64,87,93],pleas:[61,64,75,87,89],plugin:89,point:[61,70,73,74,75,87],pointer:[3,4,90],polish:74,pool:93,pop:57,popul:67,popular:[64,74,86],portabl:[57,71],portion:[55,75],porttitor:[76,78],posit:[52,70,73,89],possibl:[64,75,86,87],post:[29,30,49,52,61,65],posuer:[76,78],potenti:[49,78],pow:66,power:[61,75,86,89],pr:61,praesent:78,pragma:[42,43,44,45,90],pre:[33,54,69,71,90,91],pre_cxx11_abi:64,preced:75,precis:[49,52,61,62,65,70,83,84,89,90,93],prefer:61,prefix:[27,28,42,68,75],preinstal:64,prelu:66,prepar:[87,89],preprint:90,preproc:69,preprocess:[87,90],prerequisit:63,present:63,preserv:[75,88,90],prespect:88,press:[63,75],pretium:78,pretrain:[83,86,87,92],pretti:61,prev_next_buttons_loc:73,prevent:[49,52,55],previou:[63,73,82],previous:[29,33,61],prim:[53,54,55,57,61,66,88],prim_devic:66,primal:75,primarili:[58,61],print:[16,31,44,61,68,70,71,75,83,84,87,92],priorit:64,privat:[3,4,44,45,90],problem:75,problemat:75,proce:87,proceed:87,process:[52,55,74,75,82,86,87,88,90,92],prod:66,produc:[48,53,57,60,61,70,75,86],product:49,profil:[48,67],profiling_verbos:89,program:[18,19,20,21,29,51,52,56,57,58,65,88],programm:75,progress:76,proin:78,project:[64,74,79],projectdir:63,promis:89,prop:67,properli:64,properti:73,propog:54,prose:75,provid:[3,4,49,52,55,57,60,61,62,64,67,70,71,75,81,82,85,87,89,90,91,92],providi:[56,58],provok:75,pt:[52,61,87,89],ptq:[3,4,15,18,19,38,50,51,52,65,70,71],ptq_calibr:[3,4,45,49,90],ptqtemplat:50,publish:87,pull:[64,87],purchas:74,pure:31,purpos:[64,86,87,89],puru:78,push:57,push_back:[44,55],put:[75,86],put_binding_nam:71,pwd:[63,87],py3:87,py:[54,58,61,64,73,75,80,82,83,84,88,89,90],pyindex:[64,87],pypi:64,python3:[54,61,64],python:[52,55,58,61,67,70,71,75,76,82,83,84,85,86,87,89,91,92,93],python_api:59,pytorch:[33,48,49,52,54,55,56,57,58,60,61,62,64,69,70,71,81,85,87,88,90,91],pytorch_libtorch:87,pytorch_sphinx_them:[73,80],qat:86,qualnam:[68,69],quant_max:66,quant_min:66,quantiz:[29,30,52,61,65],quantizatiom:49,quartznet:86,question:61,qui:[76,78],quickli:[52,61,90],quisqu:78,quit:[60,61,86],quot:76,r:75,rais:[54,89],raiseexcept:54,ram:[49,52,71],rand:[61,82,89],randint:84,randn:[55,61,70,71,83,92],rang:[48,49,52,70,86,89],rank:73,rather:54,raw:73,re:[75,89],read:[3,4,29,30,44,73,75,90],read_calibration_cach:69,readcalibrationcach:[3,4,44],reader:75,realiz:57,realli:60,reason:[0,88,89],reattribut:76,recalibr:29,receiv:89,recip:90,reciproc:66,recognit:[86,90],recomend:[29,30],recommend:[29,30,61,64,75,87,89],recompil:[83,84],record:[53,88],recurs:53,redistribut:76,reduc:[54,55,56,58,86,89,90],redund:89,ref:75,refer:[48,56,58,61,74,79,87,89,90],referenc:64,refit:[45,49,71,92],reflect:45,reflection_pad1d:66,reflection_pad2d:66,regard:[64,75],regardless:[76,83,84],region:89,regist:[33,57,60,71,89],register_acc_op:89,register_acc_op_map:89,register_custom_acc_mapper_fn:89,registernodeconversionpattern:[60,61],registr:89,registri:[53,61],reinterpret_cast:44,rel:[52,55],relat:[46,75,82,83,84],relationship:50,releas:[63,75,81,85],reload_model_output:89,reload_trt_mod:89,relu:[55,61,66,82,88],relu_:66,relwithdebinfo:63,remain:[54,90],rememb:89,remov:73,remove_contigu:54,remove_dropout:54,remove_to:54,render:73,rent:76,reorder:55,repack:57,repeat:[52,66],repeat_interleav:66,replac:[55,64],replication_pad1d:66,replication_pad2d:66,replication_pad3d:66,repo:63,report:[23,44],reportable_log_level:68,repositori:[58,73,80,87],repres:[48,49,60,68,75,89],represent:[54,60,86,88,89],request:[61,70,87],requir:[29,49,52,53,54,61,68,70,71,73,87,89,90,91],require_full_compil:[45,49,71],requires_grad:66,research:89,reserv:[42,43,44,45],reset:[44,82,83,84],reshap:[66,87],resiz:87,resnet18:83,resnet50:87,resnet:[57,81,85,86,87],resnet_trt:57,resolut:86,resolv:[53,54,56,58,82,83,84],resourc:[53,90],respons:[29,57,75],rest:[75,76,89],restrict:[49,71],restructuredtext:[75,76],result:[53,54,55,62,68,71,73,87,88],ret:54,reus:[54,89,90],revert:73,revis:[75,76],revisit:75,rewrit:55,rfc:75,rho_:75,rhoncu:78,right:[42,43,44,45,54,58,60,63,75],risu:78,rm:87,rn50_preprocess:87,role:75,roll:66,roman:76,room:75,root:[42,43,44,45,64,73,90],roughli:55,round:[49,71],rounding_mod:66,row:76,rst:[73,75],rsub:66,rtol:52,rule:[64,71,89],ruler:75,run:[1,34,46,49,52,53,54,55,56,57,58,60,61,62,64,65,67,70,71,75,82,83,84,86,87,88,89,90,91,92,93],running_mean:66,running_var:66,runtim:[61,65,82,83,84],runtimeerror:89,rutrum:[76,78],s:[48,49,55,57,60,61,63,64,65,67,70,73,75,76,86,87,88,89,90],safe:[60,71],safe_dla:70,safe_gpu:70,safeti:[49,52,70],sage:75,sagitti:[76,78],sai:[76,86],said:75,same:[55,57,61,64,73,75,83,84,87,88,89,92],sampl:[62,75,82,83,84,87,89,90],sample_input:[82,89],sample_inputs_half:82,sapien:78,satisfi:[55,89],save:[29,44,52,57,61,62,70,71,86,87,89,91],save_timing_cach:[67,89],saw:61,scalar:[60,66],scalaropt_dim:66,scalartyp:[0,45,66],scale:[66,86,90],scale_factor:66,scale_grad_by_freq:66,scales_d:66,scales_h:66,scales_w:66,scatter:66,scelerisqu:78,schedul:[70,87],schema:[60,61],scheme:89,scientist:75,scope:[54,82,83,84],scratch:29,scratch_spac:87,screen:73,script:[31,54,55,61,62,70,71,82,83,84,88,92],script_model:[88,92],scriptclass:71,scripted_model:93,scriptmodul:[61,62,70,71],scroll:[73,77],sdk:59,se:86,seamlessli:65,search:[65,73],second:[54,62,75,82,83,84,89],secondli:87,section:[61,73,75,76,77,79,87,89,90],sed:[76,78],see:[31,54,55,57,61,62,64,70,71,75,82,88,89],seen:[75,76],segment:[55,83,84,86],segmentmodelwithdependencyawar:55,select:[17,29,30,34,49,52,57,62,63,64,66,70,71,74,77,89,90],self:[54,57,60,61,62,66,69,82,86,88,93],self_1:[57,61],self_int:66,sell:76,seller:74,seller_id:74,sem:78,semant:75,semper:78,send:87,senectu:78,sens:[61,75],sentenc:[75,86],sentinel:[0,2],separ:[55,56,58],sequenc:[67,70,71,75,86,89],seri:55,serial:[33,34,52,56,58,61,70,71],seriali:71,serializ:[57,88],serialized_cach:[67,89],serialized_engin:71,seril:57,serv:[52,57,65,89],servic:75,session:75,session_nam:75,set:[3,4,16,21,25,27,29,32,34,36,45,46,48,49,52,53,54,55,56,57,58,61,62,63,64,65,67,68,70,71,73,77,80,86,88,89,90,93],set_data_from_numpi:87,set_devic:[38,45,50,70],set_is_colored_output_on:[39,42,50,68],set_logging_prefix:[39,42,50,68],set_reportable_log_level:[39,42,50,68],setalpha:60,setbeta:60,setnam:[60,61],setreshapedimens:61,setup:[43,87,90],sever:[16,26,68],sh:64,sha256:64,shape:[45,47,48,49,52,55,60,62,66,67,70,71,87,89,93],shape_mod:70,shape_rang:[67,89],share:[49,52,63,64,71],shell_command:75,shift:[63,64,66,75],ship:[61,91],shorthand:75,should:[0,3,4,29,45,49,52,53,54,55,56,58,60,65,68,70,71,73,75,78,87,89,90],show:[73,75,86],shown:[61,73,75],shuffl:[61,90],side:[54,61,73],sidebar:[73,79],sigmoid:[66,89],sigmoid_:66,sign:87,signatur:71,signifi:[48,54],signific:75,significantli:[54,73],similar:[60,61,89,91,92],simonyan:90,simpil:90,simpl:[75,76,86,87,88,89],simplest:87,simpli:[54,82,86],simplifi:53,simul:86,sin:[66,75],sinc:[54,61,75,88,89,90],sing:75,singl:[48,52,54,55,61,70,75,88,89,90],singular:60,sinh:66,sink:75,sit:[76,78],site:[54,61,64,75],six:75,sixth:76,size:[3,4,44,48,49,52,54,55,61,66,67,70,71,73,83,84,86,89,90],size_t:[3,4,44,90],skip:52,slash:73,slice:66,slightli:89,sm:57,small:[54,87],smaller:86,so:[0,44,52,53,54,57,58,60,61,64,65,67,74,75,76,82,83,84,89,90],sodal:78,softmax:[54,66,89],softwar:[49,52,71,75],sole:[62,90],sollicitudin:78,solv:87,some:[53,54,55,56,57,58,60,61,74,75,89,90],some_funct:75,someth:[43,54,75,87],someurl:75,sort:[60,66,92],sourc:[42,43,44,45,58,63,67,68,69,70,71,82,83,84,85,89],sourceforg:[75,76],space:[75,76,90],spaces_and_linebreak:75,span:76,spars:[52,66],sparse_weight:[45,49,71,89],sparsiti:[49,52,71,89],spec:[45,48,49,52,68,70,71,92],special:55,specif:[32,49,54,56,58,70,71,75,85,86],specifi:[3,4,33,52,60,62,64,65,68,70,71,73,75,87,89,92],specifii:70,speech:86,speedup:86,sphinx:[73,74,75,76,80,82,83,84,85],sphinx_rtd_them:[75,76],spin:87,spirit:75,split:[55,66,89],split_siz:66,split_with_s:66,sqrt:66,squar:66,squeez:[66,86],sram:52,src:[57,59,66],ss:44,ssd300_trt:57,ssd:57,ssd_trace:52,ssd_trt:52,sstream:[20,44],stabl:[59,71,73],stack:[57,66,90],stage:[53,89],stand:[57,75],standalon:75,standard:[52,57,65,75,86,91,92],stapl:76,start:[53,55,61,64,66,68,69,76,86,89,92],start_dim:[61,66],start_step:66,state:[53,60,61],statement:[54,75],static_cast:44,statu:[44,76],std:[3,4,22,26,28,29,30,31,33,34,35,42,44,45,47,48,49,55,61,87,90,93],stdout:[37,68,70],steamlin:90,step:[55,65,66,86,89,90],stick:73,sticki:[73,79],sticky_navig:[73,77],still:[44,55,82,89,90],stitch:[55,61],stop:61,storag:90,store:[2,4,49,52,53,57,60,61,71,88,89],str:[19,43,44,50,66,68,70,71,89],straight:60,strang:75,strategi:[55,70],street:76,strict:91,strict_type_constraint:89,stride:66,string:[3,4,18,20,21,22,26,28,29,30,31,33,34,35,42,44,45,49,55,57,60,61,63,70,73,90],stringstream:44,strip_prefix:64,strong:75,strongli:75,struct:[1,21,38,41,45,90],structur:[29,46,49,55,58,60,73,75,79,87,88],structuredtext:75,stub:76,stuff:75,style:[42,43,44,45,73,75,76],style_external_link:73,sub:[66,75,82,88],sub_:66,subdirectori:51,subexpress:54,subgraph:[49,52,53,54,60,61],subject:58,submenu:79,submodul:[67,88],suboptim:55,subscript:75,subsect:75,subset:[86,90],substitut:75,subtitl:75,subtre:80,subword:86,success:63,sudo:64,suffic:54,suggest:87,suit:65,suitabl:89,sum:[49,66,71,89],superscript:75,supervis:86,suppli:75,support:[0,1,2,27,31,46,48,49,52,55,59,61,62,63,64,65,67,70,71,73,74,83,84,87,88,89,93],sure:[61,62,64,87,93],suscipit:[76,78],suspendiss:78,symbol:[33,64,71,75,89,91],symlink:80,system:[53,60,64,65,71],t1:66,t2:66,t:[0,1,2,45,46,54,60,61,64,66,70,73,75,76,87,88,89,90],t_:75,tabl:[64,79],tag:[75,87],take:[31,32,33,34,53,56,57,58,60,61,67,70,71,73,75,82,86,89,90,92],taken:75,talk:65,tan:66,tanh:66,tanh_:66,tar:[64,75,90],tarbal:[61,90],target:[1,33,45,46,48,49,52,55,57,58,62,63,65,70,71,89,90,92,93],targets_:90,task:[29,30,86,89,90],techinqu:61,techniqu:90,tell:[54,55,56,57,58,60,75],tellu:78,tem:52,templat:[20,40,44,45,50,61,73],temporari:89,tempu:78,tensor:[2,33,44,45,48,49,52,53,54,55,57,60,61,62,66,67,70,71,82,86,88,89,90],tensor_domain:[45,48,70],tensor_mod:66,tensor_scalar:66,tensor_tensor:66,tensorcontain:60,tensorformat:[21,38,45,48,50,70],tensorformatenum:50,tensorlist:[55,60],tensorrt:[0,1,3,4,29,30,31,32,33,34,37,44,45,46,48,49,52,53,54,55,56,58,60,67,69,70,71,81,82,88,90],tensorrt_bind:70,tensorrt_convert:89,tensorrt_root:63,tensorrtcompilespec:[71,92],tensort:89,teo:52,term:[63,70,75,76,86,90],termin:[27,52,61],test:[52,55,58,63,64,75,76,86,87,89,90],test_acc_trac:89,test_ptq_dataloader_calibr:90,test_ptq_trt_calibr:90,test_py_modul:[75,79],test_segment:55,testing_dataload:90,testing_dataset:90,text:[68,76,78,86],tf32:[49,52],than:[54,65,74,75,86,91],thats:[53,90],the_model_repositori:87,thei:[46,52,53,54,57,60,62,64,70,73,75,89],them:[54,55,57,61,64,73,86,89],theori:[53,75],therebi:[57,86],therefor:[29,57,61,75,86,89],theres:91,therfor:91,theta:75,thi:[0,1,2,29,30,42,43,44,45,46,47,48,49,52,53,54,55,56,57,58,60,61,63,64,67,70,71,73,74,75,77,78,81,82,83,84,85,86,87,88,89,90,91,92],thicker:75,thin:75,thing1:75,thing2:75,thing3:75,thing:[64,75,89],think:[60,75],third:[76,89],third_parti:[58,64],this_arg_is_opt:89,those:[53,75],though:[52,58,60,61,88],thought:75,three:[48,56,58,67,70,75,76,86,87,89],threshold:52,through:[48,53,54,55,57,61,62,65,68,69,75,86,89],throught:89,thrown:[49,71],thu:75,tile_to_repeat:54,time:[49,52,53,54,55,56,57,58,60,61,67,71,73,75,82,83,84,89,90],timing_cach:89,timing_cache_prefix:[67,89],tincidunt:78,tini:90,titles_onli:73,tmp:61,toctre:73,tocustomclass:60,todim:61,todo:[73,89],togeth:[53,60,61],token:86,toler:52,too:[64,73,75,76],tool:[60,61,63,86,89],toolchain:[58,64],top:[58,73,77],topk:66,torch:[0,1,2,4,20,21,29,30,31,32,33,34,37,44,45,46,47,48,49,52,53,54,55,56,57,58,60,64,67,70,71,88,90,93],torch_compil:[82,83,84],torch_compile_advanced_usag:82,torch_compile_resnet_exampl:83,torch_compile_transformers_exampl:84,torch_dir:63,torch_executed_modul:[45,49,55,71],torch_executed_op:[45,49,55,71,82,83,84],torch_scirpt_modul:88,torch_script_modul:61,torch_tensorrt:[0,1,2,3,4,14,16,17,42,43,44,46,47,48,49,50,51,52,55,61,62,65,81,82,85,86,87,89,90,91,92,93],torch_tensorrt_build:63,torch_tensorrt_export:43,torch_tensorrt_major_vers:[19,43,50],torch_tensorrt_minor_vers:[19,43,50],torch_tensorrt_patch_vers:[19,43,50],torch_tensorrt_vers:[19,43,50],torch_tensorrtfil:50,torch_tensorrtnamespac:50,torchbind:57,torchhub:87,torchscript:[19,21,38,43,45,49,50,52,55,56,57,58,62,67,70,71,86,92,93],torchscriptstruct:50,torchtrt:[43,55],torchtrt_api:[0,2,19,22,23,24,25,26,27,28,31,32,33,34,35,36,37,42,43,44,45,48,49,50],torchtrt_check:60,torchtrt_hidden:[19,43,50],torchtrt_runtime_exampl:91,torchtrt_unus:60,torchtrtc:[64,65,93],torchvis:[57,83,87,90,92],toronto:90,tortor:78,total:[82,83,84],totensor:[87,90],tovec:61,toward:90,trace:[55,61,71,88,89],traced_model:88,track:[60,90],tradit:[48,71,90],traget:32,trail:73,train:[29,30,49,52,61,62,65,66],trainabl:54,transcrib:86,transfer:74,transform:[61,81,85,87,90],transformed_img:87,translat:61,transmit:75,transpos:[66,89],trash:75,travers:[55,56,58],treat:52,tree:[42,43,44,45,73,90,91],trigger:[55,61,89],trim:90,tristiqu:78,triton:65,triton_to_np_dtyp:87,tritoncli:87,tritonserv:87,trt:[0,1,3,4,46,48,53,54,57,60,61,66,83,84,89],trt_interpreter_result:89,trt_lenet_script:61,trt_mod:[55,61,90,93],trt_model:[55,87,92],trt_ts_modul:[55,62],trtinterpret:[67,89],trtinterpreterresult:[67,89],trtmodul:[67,89],truncat:[49,52,71],truncate_long_and_doubl:[45,49,71],ts:[43,52,55,61,62,65,70,88,92],ts_model:[55,61],tt:75,tue:76,tup:66,tupl:[57,62,67,70,71,89],tupleconstruct:[54,57],tupleindex:66,tupleunpack:54,turn:[67,89],turpi:78,tutori:[88,90],two:[52,54,60,62,64,75,76,80,87,88,89,90],type:[0,1,2,30,49,50,52,53,55,57,60,61,62,63,67,68,69,70,71,75,86,89,90],type_fp32:87,typenam:[3,4,29,30,44],typic:[53,60,87],ugli:75,ui:74,uint64_t:[45,49],ultric:78,un:90,unabl:[60,61],unbind:66,unbroken:75,uncas:[84,86],uncom:64,under:[42,43,44,45,58,75,89],underli:[0,1,2,46,60],uniformli:86,union:[60,61,70,71],uniqu:[4,62],unique_ptr:[4,30],unit:89,univers:75,unknown:70,unless:89,unlik:[64,65,92],unlimit:73,unpack_addmm:54,unpack_log_softmax:54,unqiue_ptr:4,unreferenc:75,unrestrict:75,unsqueez:66,unstabl:58,unsupport:[31,49,63],unsur:60,untest:58,until:[53,55,58,60,64],unwrap:60,unwraptodoubl:60,unwraptoint:61,unzip:64,up:[53,54,55,56,57,58,75,82,83,84,86,88,89],updat:[63,67,89],upload:87,upon:[73,82,83,84],upper:76,upsample_bilinear2d:66,upsample_linear1d:66,upsample_nearest1d:66,upsample_nearest2d:66,upsample_nearest3d:66,upsample_trilinear3d:66,upscale_factor:66,upstream:61,uri:75,url:[64,73,87],urna:78,us:[0,1,2,3,4,29,30,32,34,36,43,44,45,46,48,49,52,53,55,57,58,60,63,65,67,68,69,70,71,73,74,75,76,81,85,87,88,89,90,91,93],usag:[61,69,75,81,85,89],use_cach:[3,4,30,44,69,90],use_cache_:44,use_cmake_generated_export_head:43,use_experimental_fx_rt:67,use_input_stat:66,use_python_runtim:82,use_subset:90,usecas:[62,64,85],user:[42,48,55,56,57,58,61,62,64,75,76,85,87,90],using_int:[61,66],usr:64,usual:[73,89],ut:78,utf:[75,76],util:[60,61,71,82,83,84,86,87,90],v0:[72,87],v143:63,v2:[29,30,75],v:[52,76,87],valid:[1,46,55,60,70],valu:[0,1,2,16,17,45,46,48,53,55,57,60,61,63,66,68,69,70,73,82,83,84,86],value_tensor_map:[53,60],vanilla:89,vari:67,variabl:[48,63,70,89],variant:91,varient:54,varieti:87,variou:[89,93],variu:78,vcs_pageview_mod:73,vec:66,vector:[20,21,33,44,45,47,48,49,55,57,61,90,93],vehicula:78,vel:78,velit:78,venenati:78,verbios:52,verbos:[52,67,76,83,84,89],verbose_log:[67,89],veri:[76,77,87,89,90,92],verifi:55,verison:64,version:[35,37,58,63,64,73,76,86,87,89],vertic:[73,75],vestibulum:[76,78],vgg16:90,vgg:90,vi:75,via:[62,65,70,71,73,79,82,83,84,86,89,90,91],view:[66,73],virtual:90,vision:[87,89],visitor:73,vita:[76,78],vivamu:78,viverra:78,vm:76,volutpat:78,vs:[0,1,2,46,54,71,92],vscode:63,vulput:78,w:52,w_hh:66,w_ih:66,wa:[54,57,61,75,89],wai:[52,61,64,81,85,86,88,89,90],walkthrough:86,want:[42,55,61,67,82,87,88,89,90,92],warn:[16,44,52,60,68],wash:75,we:[42,44,53,54,55,56,57,58,60,61,67,73,75,81,82,83,84,85,86,87,88,89,90],weak:75,web:75,websit:64,weight:[48,49,52,53,61,66,71,75,86,89],welcom:[61,89],well:[61,64,68,75,90],were:61,wget:87,what:[4,54,61,62,75,88,89],whatev:[57,89],wheel:64,when:[27,44,45,46,52,53,54,55,56,57,58,60,61,64,68,70,71,73,75,77,86,88,89,90],where:[53,54,60,61,71,76,89,90],wherev:89,whether:[4,52,67,70,74,83,84,89,90],which:[1,2,29,32,34,46,49,53,54,55,56,57,58,60,61,62,64,67,69,70,71,73,75,76,82,86,87,88,89,90,91,92],white:75,whitespac:75,whl:64,who:75,whole:89,whose:[54,89],why:75,wide:79,width:[75,86],win:63,window:75,window_nam:75,wish:76,within:[49,52,56,58,71,73,75],without:[55,60,61,73,75,90],wl:91,wooden:75,word:[75,86],work:[44,54,58,60,75,76,82,89,90],worker:90,workflow:[67,83,84,86,89,92],workspac:[49,52,64,67,71,82,83,84,89],workspace_s:[45,49,52,71,83,84],workspacefold:63,world:75,would:[52,60,61,62,64,87,89,91,92],wp:87,wrap:[56,57,58,61,75,78,82,83,84,89,92],wrapper:[60,89],write:[3,4,29,30,44,53,61,65,75,87,89,90],write_calibration_cach:69,writecalibrationcach:[3,4,44],wrote:75,www:[61,64,73,75,87,90],x64:63,x86:91,x86_64:[58,64],x9:54,x:[5,10,33,43,54,55,61,63,64,71,76,82,88],x_0:75,x_1:75,x_2:75,x_3:75,x_4:75,x_:75,x_lgamma:55,x_out:82,x_y_out:82,xavier:[45,93],xstr:[19,43,50],xx:87,xxx:89,y:[33,55,71,76,82],y_lgamma:55,y_out:82,yahoo:76,yaml:59,yet:[86,89],you:[0,1,2,29,30,46,48,49,52,53,54,55,57,58,60,61,62,64,65,70,71,73,75,76,77,81,85,86,87,88,89,90,91,92],your:[60,61,62,64,65,73,75,76,80,88,91,92],yourself:61,yy:87,z:76,zero_point:66,zip:[57,63,64,85],zisserman:90},titles:["Class DataType","Class Device::DeviceType","Class TensorFormat","Template Class Int8CacheCalibrator","Template Class Int8Calibrator","Define STR","Define TORCH_TENSORRT_PATCH_VERSION","Define TORCH_TENSORRT_MAJOR_VERSION","Define TORCH_TENSORRT_MINOR_VERSION","Define TORCHTRT_API","Define XSTR","Define TORCHTRT_HIDDEN","Define TORCH_TENSORRT_VERSION","Directory cpp","Directory include","Directory torch_tensorrt","Enum Level","Enum EngineCapability","File logging.h","File macros.h","File ptq.h","File torch_tensorrt.h","Function torch_tensorrt::logging::get_logging_prefix","Function torch_tensorrt::logging::get_reportable_log_level","Function torch_tensorrt::logging::get_is_colored_output_on","Function torch_tensorrt::logging::set_reportable_log_level","Function torch_tensorrt::logging::log","Function torch_tensorrt::logging::set_is_colored_output_on","Function torch_tensorrt::logging::set_logging_prefix","Template Function torch_tensorrt::ptq::make_int8_cache_calibrator","Template Function torch_tensorrt::ptq::make_int8_calibrator","Function torch_tensorrt::torchscript::check_method_operator_support","Function torch_tensorrt::torchscript::compile","Function torch_tensorrt::torchscript::embed_engine_in_new_module","Function torch_tensorrt::torchscript::convert_method_to_trt_engine","Function torch_tensorrt::get_build_info","Function torch_tensorrt::set_device","Function torch_tensorrt::dump_build_info","Namespace torch_tensorrt","Namespace torch_tensorrt::logging","Namespace torch_tensorrt::ptq","Namespace torch_tensorrt::torchscript","Program Listing for File logging.h","Program Listing for File macros.h","Program Listing for File ptq.h","Program Listing for File torch_tensorrt.h","Struct Device","Struct GraphInputs","Struct Input","Struct CompileSpec","Torch-TensorRT C++ API","Full API","torchtrtc","Conversion Phase","Lowering Phase","Partitioning Phase","Compiler Phases","Runtime Phase","System Overview","Useful Links for Torch-TensorRT Development","Writing Converters","Using Torch-TensorRT in C++","Using Torch-TensorRT in Python","Building Torch-TensorRT on Windows","Installation","Torch-TensorRT","Operators Supported","torch_tensorrt.fx","torch_tensorrt.logging","torch_tensorrt.ptq","torch_tensorrt","torch_tensorrt.ts","Changelog","Configuration","5. :mod:`test_py_module`","3. Paragraph Level Markup","4. Lists & Tables","1. Long Sticky Nav","1. Structural Elements","<no title>","Installation","Dynamo / torch.compile","Torch Compile Advanced Usage","Compiling ResNet using the Torch-TensorRT torch.compile Backend","Compiling a Transformer using torch.compile and TensorRT","Torch-TensorRT Tutorials","Example notebooks","Serving a Torch-TensorRT model with Triton","Creating a TorchScript Module","Torch-TensorRT (FX Frontend) User Guide","Post Training Quantization (PTQ)","Deploying Torch-TensorRT Programs","Using Torch-TensorRT Directly From PyTorch","DLA"],titleterms:{"1":[77,87],"10":77,"11":77,"12":77,"13":77,"14":77,"15":77,"16":77,"17":77,"18":77,"19":77,"2":[77,78,87],"20":77,"3":[77,87],"4":77,"5":77,"6":77,"7":77,"8":77,"9":77,"class":[0,1,2,3,4,20,21,38,40,41,50,67,69,70],"default":82,"enum":[16,17,38,39,50,69,70],"function":[22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,50,59,67,70,71],"import":[82,83,84],"long":[77,79],A:75,And:75,But:76,By:[18,19],Or:54,The:[61,75],To:54,With:63,aarch64:64,abi:[57,64],acc:89,acceler:86,add:89,addmm:54,admonit:75,advanc:82,advic:60,ahead:65,an:79,api:[50,51,59,64,65],applic:90,arg:[60,74],argument:[83,84],automat:55,avail:59,awar:[55,86],backend:83,background:[57,60],base:[3,4,48,73],bert:86,binari:64,block:75,branch:54,build:[63,64,73,87],bullet:76,c:[50,59,61,64,65,86,90],can:76,caption:[76,79],center:75,ch:75,changelog:72,check_method_operator_support:31,choos:64,citat:[75,90],citrinet:86,cleanup:[82,83,84],cli:[64,65],client:87,cmake:64,code:[54,63,75],compil:[32,56,58,61,63,64,65,81,82,83,84,85,86],compilespec:49,compound:75,configur:[63,73],construct:57,content:[18,19,20,21,38,39,40,41,73,74,75,76,77,78],context:[60,73],contigu:54,contract:60,contributor:65,convers:[53,56,58,60],convert:[53,60,61,66,89],convert_method_to_trt_engin:34,cpp:[13,18,19,20,21,55],creat:[88,90],creativ:75,cuda:[82,83,84],cudnn:64,current:66,custom:[61,82],cxx11:64,data:74,datatyp:0,dead:54,debug:64,deep:86,deeper:76,defin:[5,6,7,8,9,10,11,12,19,50],definit:[18,19,20,21,76,82,83,84],demo:79,depend:[55,64],deploi:[86,91],deseri:57,detect:86,develop:59,devic:[1,46],devicetyp:1,dimens:59,direct:75,directli:92,directori:[13,14,15,51],disk:88,distribut:64,dla:93,doctest:75,documen:65,document:[0,1,2,3,4,5,6,7,8,9,10,11,12,16,17,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,46,47,48,49,59,65,78,79],down:76,download:[75,80],driver:[82,83,84],dropout:54,dump_build_info:37,dynam:86,dynamo:[81,85],easier:59,efficentnet:86,element:78,elimin:54,eliminatecommonsubexpress:54,embed_engine_in_new_modul:33,emphas:75,engin:[57,89],enginecap:17,enumer:76,envior:64,error:[82,83,84],evalu:[53,66],exampl:[75,77,86],execept:54,executor:57,expect:59,face:86,fallback:[54,55],field:76,figur:75,file:[15,18,19,20,21,42,43,44,45,50,51],flatten:54,footnot:75,format:57,freez:54,from:[64,92],frontend:[86,89],full:[50,51],fuse:54,fx2trt:89,fx:[67,86,89],gaurd:54,gener:74,get:65,get_build_info:35,get_is_colored_output_on:24,get_logging_prefix:22,get_reportable_log_level:23,giant:76,git:80,glossari:75,gpu:65,graph:[54,57],graphinput:47,grid:76,guarante:60,guid:[65,89],h:[18,19,20,21,42,43,44,45,55],have:76,hierarchi:50,hlist:76,hole:76,hood:61,how:[73,89,90],html:73,hug:86,ien:75,imag:[75,76],includ:[14,18,19,20,21],incred:79,index:74,indic:65,infer:[83,84,87],inherit:[3,4,48],inlin:75,input:[48,83,84],instal:[63,64,80],int8:86,int8cachecalibr:3,int8calibr:4,ir:59,jetson:64,jit:65,languag:86,layer:59,learn:86,lenet:86,level:[16,73,75,76],librari:[64,91],libtorchtrt:91,like:76,line:75,linear:54,link:[59,75],list:[42,43,44,45,76],liter:75,local:64,log:[18,22,23,24,25,26,27,28,39,42,68],logsoftmax:54,loop:54,lower:[54,56,58],macro:[19,43],make_int8_cache_calibr:29,make_int8_calibr:30,markup:75,mask:86,math:75,menu:[77,79],meta:75,miss:89,mlm:86,mod:74,model:[82,83,84,86,87,89],modul:[54,61,88],namespac:[18,19,20,21,38,39,40,41,50],nativ:64,native_op:59,nav:77,nest:[1,46],node:53,note:[82,83,84],notebook:86,number:[75,76],nvidia:65,object:86,one:76,op:[57,89],oper:[61,66],optim:87,optimz:54,option:[73,74,76,83,84],other:60,overview:58,own:90,packag:[64,91],page:73,paragraph:[75,78],paramet:74,partit:[55,56,58],partitoninfo:55,pass:54,pattern:54,peephol:54,phase:[53,54,55,56,57,58],plugin:91,post:90,pre:64,precompil:64,prerequisit:64,program:[42,43,44,45,91],project:73,ptq:[20,29,30,40,44,69,90],python:[59,62,64,65,88,90],pytorch:[59,65,86,89,92],quantiz:[86,90],queri:87,quickstart:61,quot:75,rabbit:76,read:59,redund:54,refer:75,regist:61,relationship:[1,3,4,46,48],releas:64,remov:54,repeat:54,replac:[54,75],resnet50:86,resnet:83,respons:60,result:57,right:64,rubric:75,runtim:[56,57,58,91],save:88,second:76,section:78,segmentedblock:55,serial:57,serv:[86,87],server:87,set:[82,87],set_devic:36,set_is_colored_output_on:27,set_logging_prefix:28,set_reportable_log_level:25,setup:64,shape:86,shape_analysi:55,sidebar:75,so:91,sometim:59,sourc:64,ssd:86,start:65,step:87,sticki:77,str:5,struct:[46,47,48,49,50],structur:78,studio:63,subdirectori:[13,14],submenu:77,submodul:70,subsect:78,subsubmenu:77,subsubsect:78,support:66,system:58,tabl:[73,74,75,76,77,78],tarbal:64,target:75,templat:[3,4,29,30],tensorformat:2,tensorrt:[50,57,59,61,62,63,64,65,83,84,85,86,87,89,91,92],test_py_modul:74,text:75,theme:[73,79],thi:[76,79],through:66,tile:54,time:65,titl:75,toc:73,topic:75,torch:[50,59,61,62,63,65,81,82,83,84,85,86,87,89,91,92],torch_tensorrt:[15,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,45,67,68,69,70,71,83,84],torch_tensorrt_major_vers:7,torch_tensorrt_minor_vers:8,torch_tensorrt_patch_vers:6,torch_tensorrt_vers:12,torchscript:[31,32,33,34,41,61,65,88],torchtrt_api:9,torchtrt_hidden:11,torchtrtc:[52,61],tracer:89,train:[86,90],transform:[84,86],triton:87,ts:71,tupl:54,tutori:[65,85],type:[3,4,46,48],under:61,unpack:54,unrol:54,unsupport:61,up:87,us:[54,59,61,62,64,82,83,84,86,92],usag:82,user:[65,89],version:57,via:80,visual:63,wai:75,weight:60,what:60,wide:73,window:63,work:[61,88],write:60,xstr:10,your:[87,90]}}) \ No newline at end of file +Search.setIndex({docnames:["_cpp_api/classtorch__tensorrt_1_1DataType","_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType","_cpp_api/classtorch__tensorrt_1_1TensorFormat","_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator","_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator","_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502","_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268","_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e","_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827","_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b","_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da","_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59","_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883","_cpp_api/dir_cpp","_cpp_api/dir_cpp_include","_cpp_api/dir_cpp_include_torch_tensorrt","_cpp_api/enum_namespacetorch__tensorrt_1_1logging_1a130f65408ad8cbaee060f05e8db69558","_cpp_api/enum_namespacetorch__tensorrt_1a3fbe5d72e4fc624dbd038853079620eb","_cpp_api/file_cpp_include_torch_tensorrt_logging.h","_cpp_api/file_cpp_include_torch_tensorrt_macros.h","_cpp_api/file_cpp_include_torch_tensorrt_ptq.h","_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0593f776f469c20469e2f729fc7861a3","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0c012cb374addd90eb1f42eaec570650","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a56e110feaaba2c3fd44bd201fd21a76a","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a7cb50492421ea9de4e3db895819df6f2","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ac46ac0901cb97e3ae6e93b45f24e90b8","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ad2efd47b6c3689e58ccc595680579ae5","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1af8f3443813315af7901903d25dd495cc","_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a226e3c83379d1012cde8578c1c86b16c","_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a6186e305f47c1d94b6130ef6c7f7e178","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a5b405fd3bf3c8fc2e2a54cbbab979797","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a6e19490a08fb1553c9dd347a5ae79db9","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a81f9783517335dda877d8cfcf38987c9","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1ae8d56472106eeef37fbe51ff7f40c9b2","_cpp_api/function_namespacetorch__tensorrt_1ac4ab8313ae72c2c899ea31548b528528","_cpp_api/function_namespacetorch__tensorrt_1ad1acd06eaeaffbbcf6e7ebf426891384","_cpp_api/function_namespacetorch__tensorrt_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1","_cpp_api/namespace_torch_tensorrt","_cpp_api/namespace_torch_tensorrt__logging","_cpp_api/namespace_torch_tensorrt__ptq","_cpp_api/namespace_torch_tensorrt__torchscript","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h","_cpp_api/structtorch__tensorrt_1_1Device","_cpp_api/structtorch__tensorrt_1_1GraphInputs","_cpp_api/structtorch__tensorrt_1_1Input","_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec","_cpp_api/torch_tensort_cpp","_cpp_api/unabridged_orphan","cli/torchtrtc","contributors/conversion","contributors/fx_converters","contributors/lowering","contributors/partitioning","contributors/phases","contributors/runtime","contributors/system_overview","contributors/useful_links","contributors/writing_converters","contributors/writing_dynamo_aten_lowering_passes","getting_started/getting_started_with_cpp_api","getting_started/getting_started_with_python_api","getting_started/getting_started_with_windows","getting_started/installation","index","indices/supported_ops","py_api/fx","py_api/logging","py_api/ptq","py_api/torch_tensorrt","py_api/ts","src/pytorch-sphinx-theme/docs/changelog","src/pytorch-sphinx-theme/docs/configuring","src/pytorch-sphinx-theme/docs/demo/api","src/pytorch-sphinx-theme/docs/demo/demo","src/pytorch-sphinx-theme/docs/demo/lists_tables","src/pytorch-sphinx-theme/docs/demo/long","src/pytorch-sphinx-theme/docs/demo/structure","src/pytorch-sphinx-theme/docs/index","src/pytorch-sphinx-theme/docs/installing","tutorials/_rendered_examples/dynamo/index","tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage","tutorials/_rendered_examples/dynamo/torch_compile_resnet_example","tutorials/_rendered_examples/dynamo/torch_compile_transformers_example","tutorials/_rendered_examples/index","tutorials/notebooks","tutorials/serving_torch_tensorrt_with_triton","user_guide/creating_torchscript_module_in_python","user_guide/dynamic_shapes","user_guide/getting_started_with_fx_path","user_guide/ptq","user_guide/runtime","user_guide/saving_models","user_guide/use_from_pytorch","user_guide/using_dla"],envversion:{"sphinx.domains.c":2,"sphinx.domains.changeset":1,"sphinx.domains.citation":1,"sphinx.domains.cpp":5,"sphinx.domains.index":1,"sphinx.domains.javascript":2,"sphinx.domains.math":2,"sphinx.domains.python":3,"sphinx.domains.rst":2,"sphinx.domains.std":2,"sphinx.ext.intersphinx":1,"sphinx.ext.todo":2,"sphinx.ext.viewcode":1,nbsphinx:4,sphinx:56},filenames:["_cpp_api/classtorch__tensorrt_1_1DataType.rst","_cpp_api/classtorch__tensorrt_1_1Device_1_1DeviceType.rst","_cpp_api/classtorch__tensorrt_1_1TensorFormat.rst","_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8CacheCalibrator.rst","_cpp_api/classtorch__tensorrt_1_1ptq_1_1Int8Calibrator.rst","_cpp_api/define_macros_8h_1a18d295a837ac71add5578860b55e5502.rst","_cpp_api/define_macros_8h_1a282fd3c0b1c3a215148ae372070e1268.rst","_cpp_api/define_macros_8h_1a31398a6d4d27e28817afb0f0139e909e.rst","_cpp_api/define_macros_8h_1a35703561b26b1a9d2738ad7d58b27827.rst","_cpp_api/define_macros_8h_1abd1465eb38256d3f22cc1426b23d516b.rst","_cpp_api/define_macros_8h_1abe87b341f562fd1cf40b7672e4d759da.rst","_cpp_api/define_macros_8h_1ad19939408f7be171a74a89928b36eb59.rst","_cpp_api/define_macros_8h_1adad592a7b1b7eed529cdf6acd584c883.rst","_cpp_api/dir_cpp.rst","_cpp_api/dir_cpp_include.rst","_cpp_api/dir_cpp_include_torch_tensorrt.rst","_cpp_api/enum_namespacetorch__tensorrt_1_1logging_1a130f65408ad8cbaee060f05e8db69558.rst","_cpp_api/enum_namespacetorch__tensorrt_1a3fbe5d72e4fc624dbd038853079620eb.rst","_cpp_api/file_cpp_include_torch_tensorrt_logging.h.rst","_cpp_api/file_cpp_include_torch_tensorrt_macros.h.rst","_cpp_api/file_cpp_include_torch_tensorrt_ptq.h.rst","_cpp_api/file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0593f776f469c20469e2f729fc7861a3.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a0c012cb374addd90eb1f42eaec570650.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a56e110feaaba2c3fd44bd201fd21a76a.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1a7cb50492421ea9de4e3db895819df6f2.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ac46ac0901cb97e3ae6e93b45f24e90b8.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1ad2efd47b6c3689e58ccc595680579ae5.rst","_cpp_api/function_namespacetorch__tensorrt_1_1logging_1af8f3443813315af7901903d25dd495cc.rst","_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a226e3c83379d1012cde8578c1c86b16c.rst","_cpp_api/function_namespacetorch__tensorrt_1_1ptq_1a6186e305f47c1d94b6130ef6c7f7e178.rst","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a5b405fd3bf3c8fc2e2a54cbbab979797.rst","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a6e19490a08fb1553c9dd347a5ae79db9.rst","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1a81f9783517335dda877d8cfcf38987c9.rst","_cpp_api/function_namespacetorch__tensorrt_1_1torchscript_1ae8d56472106eeef37fbe51ff7f40c9b2.rst","_cpp_api/function_namespacetorch__tensorrt_1ac4ab8313ae72c2c899ea31548b528528.rst","_cpp_api/function_namespacetorch__tensorrt_1ad1acd06eaeaffbbcf6e7ebf426891384.rst","_cpp_api/function_namespacetorch__tensorrt_1ad6a4ee8ca6c8f6e5519eb1128ec7f4a1.rst","_cpp_api/namespace_torch_tensorrt.rst","_cpp_api/namespace_torch_tensorrt__logging.rst","_cpp_api/namespace_torch_tensorrt__ptq.rst","_cpp_api/namespace_torch_tensorrt__torchscript.rst","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_logging.h.rst","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_macros.h.rst","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_ptq.h.rst","_cpp_api/program_listing_file_cpp_include_torch_tensorrt_torch_tensorrt.h.rst","_cpp_api/structtorch__tensorrt_1_1Device.rst","_cpp_api/structtorch__tensorrt_1_1GraphInputs.rst","_cpp_api/structtorch__tensorrt_1_1Input.rst","_cpp_api/structtorch__tensorrt_1_1torchscript_1_1CompileSpec.rst","_cpp_api/torch_tensort_cpp.rst","_cpp_api/unabridged_orphan.rst","cli/torchtrtc.rst","contributors/conversion.rst","contributors/fx_converters.rst","contributors/lowering.rst","contributors/partitioning.rst","contributors/phases.rst","contributors/runtime.rst","contributors/system_overview.rst","contributors/useful_links.rst","contributors/writing_converters.rst","contributors/writing_dynamo_aten_lowering_passes.rst","getting_started/getting_started_with_cpp_api.rst","getting_started/getting_started_with_python_api.rst","getting_started/getting_started_with_windows.rst","getting_started/installation.rst","index.rst","indices/supported_ops.rst","py_api/fx.rst","py_api/logging.rst","py_api/ptq.rst","py_api/torch_tensorrt.rst","py_api/ts.rst","src/pytorch-sphinx-theme/docs/changelog.rst","src/pytorch-sphinx-theme/docs/configuring.rst","src/pytorch-sphinx-theme/docs/demo/api.rst","src/pytorch-sphinx-theme/docs/demo/demo.rst","src/pytorch-sphinx-theme/docs/demo/lists_tables.rst","src/pytorch-sphinx-theme/docs/demo/long.rst","src/pytorch-sphinx-theme/docs/demo/structure.rst","src/pytorch-sphinx-theme/docs/index.rst","src/pytorch-sphinx-theme/docs/installing.rst","tutorials/_rendered_examples/dynamo/index.rst","tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.rst","tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.rst","tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.rst","tutorials/_rendered_examples/index.rst","tutorials/notebooks.rst","tutorials/serving_torch_tensorrt_with_triton.rst","user_guide/creating_torchscript_module_in_python.rst","user_guide/dynamic_shapes.rst","user_guide/getting_started_with_fx_path.rst","user_guide/ptq.rst","user_guide/runtime.rst","user_guide/saving_models.rst","user_guide/use_from_pytorch.rst","user_guide/using_dla.rst"],objects:{"":[[5,0,1,"c.STR","STR"],[9,0,1,"c.TORCHTRT_API","TORCHTRT_API"],[11,0,1,"c.TORCHTRT_HIDDEN","TORCHTRT_HIDDEN"],[7,0,1,"c.TORCH_TENSORRT_MAJOR_VERSION","TORCH_TENSORRT_MAJOR_VERSION"],[8,0,1,"c.TORCH_TENSORRT_MINOR_VERSION","TORCH_TENSORRT_MINOR_VERSION"],[6,0,1,"c.TORCH_TENSORRT_PATCH_VERSION","TORCH_TENSORRT_PATCH_VERSION"],[12,0,1,"c.TORCH_TENSORRT_VERSION","TORCH_TENSORRT_VERSION"],[10,0,1,"c.XSTR","XSTR"],[0,1,1,"_CPPv4N14torch_tensorrt8DataTypeE","torch_tensorrt::DataType"],[0,2,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeE5Value","torch_tensorrt::DataType::DataType"],[0,2,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeEN3c1010ScalarTypeE","torch_tensorrt::DataType::DataType"],[0,2,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeEv","torch_tensorrt::DataType::DataType"],[0,3,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeE5Value","torch_tensorrt::DataType::DataType::t"],[0,3,1,"_CPPv4N14torch_tensorrt8DataType8DataTypeEN3c1010ScalarTypeE","torch_tensorrt::DataType::DataType::t"],[0,4,1,"_CPPv4N14torch_tensorrt8DataType5ValueE","torch_tensorrt::DataType::Value"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kBoolE","torch_tensorrt::DataType::Value::kBool"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kCharE","torch_tensorrt::DataType::Value::kChar"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value7kDoubleE","torch_tensorrt::DataType::Value::kDouble"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value6kFloatE","torch_tensorrt::DataType::Value::kFloat"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kHalfE","torch_tensorrt::DataType::Value::kHalf"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value4kIntE","torch_tensorrt::DataType::Value::kInt"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kLongE","torch_tensorrt::DataType::Value::kLong"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value8kUnknownE","torch_tensorrt::DataType::Value::kUnknown"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kBoolE","torch_tensorrt::DataType::kBool"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kCharE","torch_tensorrt::DataType::kChar"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value7kDoubleE","torch_tensorrt::DataType::kDouble"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value6kFloatE","torch_tensorrt::DataType::kFloat"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kHalfE","torch_tensorrt::DataType::kHalf"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value4kIntE","torch_tensorrt::DataType::kInt"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value5kLongE","torch_tensorrt::DataType::kLong"],[0,5,1,"_CPPv4N14torch_tensorrt8DataType5Value8kUnknownE","torch_tensorrt::DataType::kUnknown"],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypecv5ValueEv","torch_tensorrt::DataType::operator Value"],[0,2,1,"_CPPv4N14torch_tensorrt8DataTypecvbEv","torch_tensorrt::DataType::operator bool"],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypeneE8DataType","torch_tensorrt::DataType::operator!="],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypeneEN8DataType5ValueE","torch_tensorrt::DataType::operator!="],[0,3,1,"_CPPv4NK14torch_tensorrt8DataTypeneE8DataType","torch_tensorrt::DataType::operator!=::other"],[0,3,1,"_CPPv4NK14torch_tensorrt8DataTypeneEN8DataType5ValueE","torch_tensorrt::DataType::operator!=::other"],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypeeqE8DataType","torch_tensorrt::DataType::operator=="],[0,2,1,"_CPPv4NK14torch_tensorrt8DataTypeeqEN8DataType5ValueE","torch_tensorrt::DataType::operator=="],[0,3,1,"_CPPv4NK14torch_tensorrt8DataTypeeqE8DataType","torch_tensorrt::DataType::operator==::other"],[0,3,1,"_CPPv4NK14torch_tensorrt8DataTypeeqEN8DataType5ValueE","torch_tensorrt::DataType::operator==::other"],[46,1,1,"_CPPv4N14torch_tensorrt6DeviceE","torch_tensorrt::Device"],[46,2,1,"_CPPv4N14torch_tensorrt6Device6DeviceEv","torch_tensorrt::Device::Device"],[1,1,1,"_CPPv4N14torch_tensorrt6Device10DeviceTypeE","torch_tensorrt::Device::DeviceType"],[46,1,1,"_CPPv4N14torch_tensorrt6Device10DeviceTypeE","torch_tensorrt::Device::DeviceType"],[1,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value","torch_tensorrt::Device::DeviceType::DeviceType"],[1,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE","torch_tensorrt::Device::DeviceType::DeviceType"],[1,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEv","torch_tensorrt::Device::DeviceType::DeviceType"],[46,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value","torch_tensorrt::Device::DeviceType::DeviceType"],[46,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE","torch_tensorrt::Device::DeviceType::DeviceType"],[46,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEv","torch_tensorrt::Device::DeviceType::DeviceType"],[1,3,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value","torch_tensorrt::Device::DeviceType::DeviceType::t"],[1,3,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE","torch_tensorrt::Device::DeviceType::DeviceType::t"],[46,3,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeE5Value","torch_tensorrt::Device::DeviceType::DeviceType::t"],[46,3,1,"_CPPv4N14torch_tensorrt6Device10DeviceType10DeviceTypeEN3c1010DeviceTypeE","torch_tensorrt::Device::DeviceType::DeviceType::t"],[1,4,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5ValueE","torch_tensorrt::Device::DeviceType::Value"],[46,4,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5ValueE","torch_tensorrt::Device::DeviceType::Value"],[1,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE","torch_tensorrt::Device::DeviceType::Value::kDLA"],[46,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE","torch_tensorrt::Device::DeviceType::Value::kDLA"],[1,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE","torch_tensorrt::Device::DeviceType::Value::kGPU"],[46,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE","torch_tensorrt::Device::DeviceType::Value::kGPU"],[1,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kDLAE","torch_tensorrt::Device::DeviceType::kDLA"],[1,5,1,"_CPPv4N14torch_tensorrt6Device10DeviceType5Value4kGPUE","torch_tensorrt::Device::DeviceType::kGPU"],[1,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypecv5ValueEv","torch_tensorrt::Device::DeviceType::operator Value"],[46,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypecv5ValueEv","torch_tensorrt::Device::DeviceType::operator Value"],[1,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceTypecvbEv","torch_tensorrt::Device::DeviceType::operator bool"],[46,2,1,"_CPPv4N14torch_tensorrt6Device10DeviceTypecvbEv","torch_tensorrt::Device::DeviceType::operator bool"],[1,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType","torch_tensorrt::Device::DeviceType::operator!="],[46,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType","torch_tensorrt::Device::DeviceType::operator!="],[1,3,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType","torch_tensorrt::Device::DeviceType::operator!=::other"],[46,3,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeneE10DeviceType","torch_tensorrt::Device::DeviceType::operator!=::other"],[1,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType","torch_tensorrt::Device::DeviceType::operator=="],[46,2,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType","torch_tensorrt::Device::DeviceType::operator=="],[1,3,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType","torch_tensorrt::Device::DeviceType::operator==::other"],[46,3,1,"_CPPv4NK14torch_tensorrt6Device10DeviceTypeeqE10DeviceType","torch_tensorrt::Device::DeviceType::operator==::other"],[46,6,1,"_CPPv4N14torch_tensorrt6Device18allow_gpu_fallbackE","torch_tensorrt::Device::allow_gpu_fallback"],[46,6,1,"_CPPv4N14torch_tensorrt6Device11device_typeE","torch_tensorrt::Device::device_type"],[46,6,1,"_CPPv4N14torch_tensorrt6Device8dla_coreE","torch_tensorrt::Device::dla_core"],[46,6,1,"_CPPv4N14torch_tensorrt6Device6gpu_idE","torch_tensorrt::Device::gpu_id"],[17,4,1,"_CPPv4N14torch_tensorrt16EngineCapabilityE","torch_tensorrt::EngineCapability"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability15kDLA_STANDALONEE","torch_tensorrt::EngineCapability::kDLA_STANDALONE"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability7kSAFETYE","torch_tensorrt::EngineCapability::kSAFETY"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability9kSTANDARDE","torch_tensorrt::EngineCapability::kSTANDARD"],[47,1,1,"_CPPv4N14torch_tensorrt11GraphInputsE","torch_tensorrt::GraphInputs"],[47,6,1,"_CPPv4N14torch_tensorrt11GraphInputs15input_signatureE","torch_tensorrt::GraphInputs::input_signature"],[47,6,1,"_CPPv4N14torch_tensorrt11GraphInputs6inputsE","torch_tensorrt::GraphInputs::inputs"],[48,1,1,"_CPPv4N14torch_tensorrt5InputE","torch_tensorrt::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN2at6TensorE","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input"],[48,2,1,"_CPPv4N14torch_tensorrt5Input5InputEv","torch_tensorrt::Input::Input"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::dtype"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::format"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::max_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::min_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::opt_shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataType12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::shape"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN2at6TensorE","torch_tensorrt::Input::Input::tensor"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputEN3c108ArrayRefI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEE8DataTypeNSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,3,1,"_CPPv4N14torch_tensorrt5Input5InputENSt6vectorI7int64_tEENSt6vectorIdEE12TensorFormat","torch_tensorrt::Input::Input::tensor_domain"],[48,6,1,"_CPPv4N14torch_tensorrt5Input5dtypeE","torch_tensorrt::Input::dtype"],[48,6,1,"_CPPv4N14torch_tensorrt5Input6formatE","torch_tensorrt::Input::format"],[48,6,1,"_CPPv4N14torch_tensorrt5Input9max_shapeE","torch_tensorrt::Input::max_shape"],[48,6,1,"_CPPv4N14torch_tensorrt5Input9min_shapeE","torch_tensorrt::Input::min_shape"],[48,6,1,"_CPPv4N14torch_tensorrt5Input9opt_shapeE","torch_tensorrt::Input::opt_shape"],[48,6,1,"_CPPv4N14torch_tensorrt5Input5shapeE","torch_tensorrt::Input::shape"],[48,6,1,"_CPPv4N14torch_tensorrt5Input13tensor_domainE","torch_tensorrt::Input::tensor_domain"],[2,1,1,"_CPPv4N14torch_tensorrt12TensorFormatE","torch_tensorrt::TensorFormat"],[2,2,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatE5Value","torch_tensorrt::TensorFormat::TensorFormat"],[2,2,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEN2at12MemoryFormatE","torch_tensorrt::TensorFormat::TensorFormat"],[2,2,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEv","torch_tensorrt::TensorFormat::TensorFormat"],[2,3,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatE5Value","torch_tensorrt::TensorFormat::TensorFormat::t"],[2,3,1,"_CPPv4N14torch_tensorrt12TensorFormat12TensorFormatEN2at12MemoryFormatE","torch_tensorrt::TensorFormat::TensorFormat::t"],[2,4,1,"_CPPv4N14torch_tensorrt12TensorFormat5ValueE","torch_tensorrt::TensorFormat::Value"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value13kChannelsLastE","torch_tensorrt::TensorFormat::Value::kChannelsLast"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value11kContiguousE","torch_tensorrt::TensorFormat::Value::kContiguous"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value8kUnknownE","torch_tensorrt::TensorFormat::Value::kUnknown"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value13kChannelsLastE","torch_tensorrt::TensorFormat::kChannelsLast"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value11kContiguousE","torch_tensorrt::TensorFormat::kContiguous"],[2,5,1,"_CPPv4N14torch_tensorrt12TensorFormat5Value8kUnknownE","torch_tensorrt::TensorFormat::kUnknown"],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormatcv5ValueEv","torch_tensorrt::TensorFormat::operator Value"],[2,2,1,"_CPPv4N14torch_tensorrt12TensorFormatcvbEv","torch_tensorrt::TensorFormat::operator bool"],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormatneE12TensorFormat","torch_tensorrt::TensorFormat::operator!="],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormatneEN12TensorFormat5ValueE","torch_tensorrt::TensorFormat::operator!="],[2,3,1,"_CPPv4NK14torch_tensorrt12TensorFormatneE12TensorFormat","torch_tensorrt::TensorFormat::operator!=::other"],[2,3,1,"_CPPv4NK14torch_tensorrt12TensorFormatneEN12TensorFormat5ValueE","torch_tensorrt::TensorFormat::operator!=::other"],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormateqE12TensorFormat","torch_tensorrt::TensorFormat::operator=="],[2,2,1,"_CPPv4NK14torch_tensorrt12TensorFormateqEN12TensorFormat5ValueE","torch_tensorrt::TensorFormat::operator=="],[2,3,1,"_CPPv4NK14torch_tensorrt12TensorFormateqE12TensorFormat","torch_tensorrt::TensorFormat::operator==::other"],[2,3,1,"_CPPv4NK14torch_tensorrt12TensorFormateqEN12TensorFormat5ValueE","torch_tensorrt::TensorFormat::operator==::other"],[37,2,1,"_CPPv4N14torch_tensorrt15dump_build_infoEv","torch_tensorrt::dump_build_info"],[35,2,1,"_CPPv4N14torch_tensorrt14get_build_infoEv","torch_tensorrt::get_build_info"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability15kDLA_STANDALONEE","torch_tensorrt::kDLA_STANDALONE"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability7kSAFETYE","torch_tensorrt::kSAFETY"],[17,5,1,"_CPPv4N14torch_tensorrt16EngineCapability9kSTANDARDE","torch_tensorrt::kSTANDARD"],[16,4,1,"_CPPv4N14torch_tensorrt7logging5LevelE","torch_tensorrt::logging::Level"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kDEBUGE","torch_tensorrt::logging::Level::kDEBUG"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kERRORE","torch_tensorrt::logging::Level::kERROR"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kGRAPHE","torch_tensorrt::logging::Level::kGRAPH"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level5kINFOE","torch_tensorrt::logging::Level::kINFO"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level15kINTERNAL_ERRORE","torch_tensorrt::logging::Level::kINTERNAL_ERROR"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level8kWARNINGE","torch_tensorrt::logging::Level::kWARNING"],[24,2,1,"_CPPv4N14torch_tensorrt7logging24get_is_colored_output_onEv","torch_tensorrt::logging::get_is_colored_output_on"],[22,2,1,"_CPPv4N14torch_tensorrt7logging18get_logging_prefixEv","torch_tensorrt::logging::get_logging_prefix"],[23,2,1,"_CPPv4N14torch_tensorrt7logging24get_reportable_log_levelEv","torch_tensorrt::logging::get_reportable_log_level"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kDEBUGE","torch_tensorrt::logging::kDEBUG"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kERRORE","torch_tensorrt::logging::kERROR"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level6kGRAPHE","torch_tensorrt::logging::kGRAPH"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level5kINFOE","torch_tensorrt::logging::kINFO"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level15kINTERNAL_ERRORE","torch_tensorrt::logging::kINTERNAL_ERROR"],[16,5,1,"_CPPv4N14torch_tensorrt7logging5Level8kWARNINGE","torch_tensorrt::logging::kWARNING"],[26,2,1,"_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE","torch_tensorrt::logging::log"],[26,3,1,"_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE","torch_tensorrt::logging::log::lvl"],[26,3,1,"_CPPv4N14torch_tensorrt7logging3logE5LevelNSt6stringE","torch_tensorrt::logging::log::msg"],[27,2,1,"_CPPv4N14torch_tensorrt7logging24set_is_colored_output_onEb","torch_tensorrt::logging::set_is_colored_output_on"],[27,3,1,"_CPPv4N14torch_tensorrt7logging24set_is_colored_output_onEb","torch_tensorrt::logging::set_is_colored_output_on::colored_output_on"],[28,2,1,"_CPPv4N14torch_tensorrt7logging18set_logging_prefixENSt6stringE","torch_tensorrt::logging::set_logging_prefix"],[28,3,1,"_CPPv4N14torch_tensorrt7logging18set_logging_prefixENSt6stringE","torch_tensorrt::logging::set_logging_prefix::prefix"],[25,2,1,"_CPPv4N14torch_tensorrt7logging24set_reportable_log_levelE5Level","torch_tensorrt::logging::set_reportable_log_level"],[25,3,1,"_CPPv4N14torch_tensorrt7logging24set_reportable_log_levelE5Level","torch_tensorrt::logging::set_reportable_log_level::lvl"],[3,1,1,"_CPPv4I0EN14torch_tensorrt3ptq19Int8CacheCalibratorE","torch_tensorrt::ptq::Int8CacheCalibrator"],[3,7,1,"_CPPv4I0EN14torch_tensorrt3ptq19Int8CacheCalibratorE","torch_tensorrt::ptq::Int8CacheCalibrator::Algorithm"],[3,2,1,"_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator19Int8CacheCalibratorERKNSt6stringE","torch_tensorrt::ptq::Int8CacheCalibrator::Int8CacheCalibrator"],[3,3,1,"_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibrator19Int8CacheCalibratorERKNSt6stringE","torch_tensorrt::ptq::Int8CacheCalibrator::Int8CacheCalibrator::cache_file_path"],[3,2,1,"_CPPv4N14torch_tensorrt3ptq19Int8CacheCalibratorcvPN8nvinfer115IInt8CalibratorEEv","torch_tensorrt::ptq::Int8CacheCalibrator::operator nvinfer1::IInt8Calibrator*"],[4,1,1,"_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE","torch_tensorrt::ptq::Int8Calibrator"],[4,7,1,"_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE","torch_tensorrt::ptq::Int8Calibrator::Algorithm"],[4,7,1,"_CPPv4I00EN14torch_tensorrt3ptq14Int8CalibratorE","torch_tensorrt::ptq::Int8Calibrator::DataLoaderUniquePtr"],[4,2,1,"_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb","torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator"],[4,3,1,"_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb","torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::cache_file_path"],[4,3,1,"_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb","torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::dataloader"],[4,3,1,"_CPPv4N14torch_tensorrt3ptq14Int8Calibrator14Int8CalibratorE19DataLoaderUniquePtrRKNSt6stringEb","torch_tensorrt::ptq::Int8Calibrator::Int8Calibrator::use_cache"],[4,2,1,"_CPPv4N14torch_tensorrt3ptq14Int8CalibratorcvPN8nvinfer115IInt8CalibratorEEv","torch_tensorrt::ptq::Int8Calibrator::operator nvinfer1::IInt8Calibrator*"],[29,2,1,"_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE","torch_tensorrt::ptq::make_int8_cache_calibrator"],[29,7,1,"_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE","torch_tensorrt::ptq::make_int8_cache_calibrator::Algorithm"],[29,3,1,"_CPPv4I0EN14torch_tensorrt3ptq26make_int8_cache_calibratorE19Int8CacheCalibratorI9AlgorithmERKNSt6stringE","torch_tensorrt::ptq::make_int8_cache_calibrator::cache_file_path"],[30,2,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator"],[30,7,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::Algorithm"],[30,7,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::DataLoader"],[30,3,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::cache_file_path"],[30,3,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::dataloader"],[30,3,1,"_CPPv4I00EN14torch_tensorrt3ptq20make_int8_calibratorE14Int8CalibratorI9Algorithm10DataLoaderE10DataLoaderRKNSt6stringEb","torch_tensorrt::ptq::make_int8_calibrator::use_cache"],[36,2,1,"_CPPv4N14torch_tensorrt10set_deviceEKi","torch_tensorrt::set_device"],[36,3,1,"_CPPv4N14torch_tensorrt10set_deviceEKi","torch_tensorrt::set_device::gpu_id"],[49,1,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpecE","torch_tensorrt::torchscript::CompileSpec"],[49,2,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecEN5torch3jit6IValueE","torch_tensorrt::torchscript::CompileSpec::CompileSpec"],[49,2,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorI5InputEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec"],[49,2,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorIN3c108ArrayRefI7int64_tEEEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec"],[49,2,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorINSt6vectorI7int64_tEEEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec"],[49,3,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorIN3c108ArrayRefI7int64_tEEEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec::fixed_sizes"],[49,3,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorINSt6vectorI7int64_tEEEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec::fixed_sizes"],[49,3,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecEN5torch3jit6IValueE","torch_tensorrt::torchscript::CompileSpec::CompileSpec::input_signature"],[49,3,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec11CompileSpecENSt6vectorI5InputEE","torch_tensorrt::torchscript::CompileSpec::CompileSpec::inputs"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec19allow_shape_tensorsE","torch_tensorrt::torchscript::CompileSpec::allow_shape_tensors"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec10capabilityE","torch_tensorrt::torchscript::CompileSpec::capability"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec5debugE","torch_tensorrt::torchscript::CompileSpec::debug"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec6deviceE","torch_tensorrt::torchscript::CompileSpec::device"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec12disable_tf32E","torch_tensorrt::torchscript::CompileSpec::disable_tf32"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec20dla_global_dram_sizeE","torch_tensorrt::torchscript::CompileSpec::dla_global_dram_size"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec19dla_local_dram_sizeE","torch_tensorrt::torchscript::CompileSpec::dla_local_dram_size"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec13dla_sram_sizeE","torch_tensorrt::torchscript::CompileSpec::dla_sram_size"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec18enabled_precisionsE","torch_tensorrt::torchscript::CompileSpec::enabled_precisions"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec12graph_inputsE","torch_tensorrt::torchscript::CompileSpec::graph_inputs"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec14min_block_sizeE","torch_tensorrt::torchscript::CompileSpec::min_block_size"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec20num_avg_timing_itersE","torch_tensorrt::torchscript::CompileSpec::num_avg_timing_iters"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec14ptq_calibratorE","torch_tensorrt::torchscript::CompileSpec::ptq_calibrator"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec5refitE","torch_tensorrt::torchscript::CompileSpec::refit"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec24require_full_compilationE","torch_tensorrt::torchscript::CompileSpec::require_full_compilation"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec14sparse_weightsE","torch_tensorrt::torchscript::CompileSpec::sparse_weights"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec22torch_executed_modulesE","torch_tensorrt::torchscript::CompileSpec::torch_executed_modules"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec18torch_executed_opsE","torch_tensorrt::torchscript::CompileSpec::torch_executed_ops"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec24truncate_long_and_doubleE","torch_tensorrt::torchscript::CompileSpec::truncate_long_and_double"],[49,6,1,"_CPPv4N14torch_tensorrt11torchscript11CompileSpec14workspace_sizeE","torch_tensorrt::torchscript::CompileSpec::workspace_size"],[31,2,1,"_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE","torch_tensorrt::torchscript::check_method_operator_support"],[31,3,1,"_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE","torch_tensorrt::torchscript::check_method_operator_support::method_name"],[31,3,1,"_CPPv4N14torch_tensorrt11torchscript29check_method_operator_supportERKN5torch3jit6ModuleENSt6stringE","torch_tensorrt::torchscript::check_method_operator_support::module"],[32,2,1,"_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec","torch_tensorrt::torchscript::compile"],[32,3,1,"_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec","torch_tensorrt::torchscript::compile::info"],[32,3,1,"_CPPv4N14torch_tensorrt11torchscript7compileERKN5torch3jit6ModuleE11CompileSpec","torch_tensorrt::torchscript::compile::module"],[34,2,1,"_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec","torch_tensorrt::torchscript::convert_method_to_trt_engine"],[34,3,1,"_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec","torch_tensorrt::torchscript::convert_method_to_trt_engine::info"],[34,3,1,"_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec","torch_tensorrt::torchscript::convert_method_to_trt_engine::method_name"],[34,3,1,"_CPPv4N14torch_tensorrt11torchscript28convert_method_to_trt_engineERKN5torch3jit6ModuleENSt6stringE11CompileSpec","torch_tensorrt::torchscript::convert_method_to_trt_engine::module"],[33,2,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module"],[33,3,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module::device"],[33,3,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module::engine"],[33,3,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module::input_binding_names"],[33,3,1,"_CPPv4N14torch_tensorrt11torchscript26embed_engine_in_new_moduleERKNSt6stringE6DeviceRKNSt6vectorINSt6stringEEERKNSt6vectorINSt6stringEEE","torch_tensorrt::torchscript::embed_engine_in_new_module::output_binding_names"],[72,8,0,"-","torch_tensorrt"]],"torch_tensorrt.Device":[[72,10,1,"","__init__"],[72,11,1,"","allow_gpu_fallback"],[72,11,1,"","device_type"],[72,11,1,"","dla_core"],[72,11,1,"","gpu_id"]],"torch_tensorrt.Input":[[72,10,1,"","__init__"],[72,11,1,"","dtype"],[72,10,1,"","example_tensor"],[72,11,1,"","format"],[72,10,1,"","from_tensor"],[72,10,1,"","from_tensors"],[72,11,1,"","shape"],[72,11,1,"","shape_mode"]],"torch_tensorrt.fx":[[69,9,1,"","InputTensorSpec"],[69,9,1,"","TRTInterpreter"],[69,9,1,"","TRTInterpreterResult"],[69,9,1,"","TRTModule"],[69,12,1,"","compile"]],"torch_tensorrt.logging":[[70,9,1,"","Level"],[70,9,1,"","debug"],[70,9,1,"","errors"],[70,12,1,"","get_is_colored_output_on"],[70,12,1,"","get_logging_prefix"],[70,12,1,"","get_reportable_log_level"],[70,9,1,"","graphs"],[70,9,1,"","info"],[70,9,1,"","internal_errors"],[70,12,1,"","log"],[70,12,1,"","set_is_colored_output_on"],[70,12,1,"","set_logging_prefix"],[70,12,1,"","set_reportable_log_level"],[70,9,1,"","warnings"]],"torch_tensorrt.logging.Level":[[70,11,1,"","Debug"],[70,11,1,"","Error"],[70,11,1,"","Graph"],[70,11,1,"","Info"],[70,11,1,"","InternalError"],[70,11,1,"","Warning"]],"torch_tensorrt.ptq":[[71,9,1,"id1","CacheCalibrator"],[71,9,1,"id2","CalibrationAlgo"],[71,9,1,"id0","DataLoaderCalibrator"],[71,12,1,"","get_batch"],[71,12,1,"","get_batch_size"],[71,12,1,"","get_cache_mode_batch"],[71,12,1,"","read_calibration_cache"],[71,12,1,"","write_calibration_cache"]],"torch_tensorrt.ptq.CacheCalibrator":[[71,10,1,"","__init__"]],"torch_tensorrt.ptq.CalibrationAlgo":[[71,11,1,"","ENTROPY_CALIBRATION"],[71,11,1,"","ENTROPY_CALIBRATION_2"],[71,11,1,"","LEGACY_CALIBRATION"],[71,11,1,"","MINMAX_CALIBRATION"]],"torch_tensorrt.ptq.DataLoaderCalibrator":[[71,10,1,"","__init__"]],"torch_tensorrt.ts":[[73,12,1,"","TensorRTCompileSpec"],[73,12,1,"","check_method_op_support"],[73,12,1,"","compile"],[73,12,1,"","convert_method_to_trt_engine"],[73,12,1,"","embed_engine_in_new_module"]],torch_tensorrt:[[72,9,1,"","Device"],[72,9,1,"","DeviceType"],[72,9,1,"","EngineCapability"],[72,9,1,"","Input"],[72,9,1,"","TensorFormat"],[72,12,1,"","compile"],[72,12,1,"","convert_method_to_trt_engine"],[72,9,1,"","dtype"],[72,12,1,"","dump_build_info"],[69,8,0,"-","fx"],[72,12,1,"","get_build_info"],[70,8,0,"-","logging"],[71,8,0,"-","ptq"],[72,12,1,"","set_device"],[73,8,0,"-","ts"]]},objnames:{"0":["c","macro","C macro"],"1":["cpp","class","C++ class"],"10":["py","method","Python method"],"11":["py","attribute","Python attribute"],"12":["py","function","Python function"],"2":["cpp","function","C++ function"],"3":["cpp","functionParam","C++ function parameter"],"4":["cpp","enum","C++ enum"],"5":["cpp","enumerator","C++ enumerator"],"6":["cpp","member","C++ member"],"7":["cpp","templateParam","C++ template parameter"],"8":["py","module","Python module"],"9":["py","class","Python class"]},objtypes:{"0":"c:macro","1":"cpp:class","10":"py:method","11":"py:attribute","12":"py:function","2":"cpp:function","3":"cpp:functionParam","4":"cpp:enum","5":"cpp:enumerator","6":"cpp:member","7":"cpp:templateParam","8":"py:module","9":"py:class"},terms:{"0":[33,43,44,45,49,52,54,56,59,61,62,63,65,66,68,69,70,71,72,73,74,76,77,83,84,85,86,87,89,91,92,93,96,97],"000":[84,85,86],"0000":78,"01":[63,68,78],"0208":63,"03":78,"0358":63,"0383":63,"04":[63,89],"0435":63,"0464":63,"0530":63,"0678":63,"0805":63,"0818":63,"0932":63,"0x7f6f6edc45b0":73,"1":[3,4,33,44,45,48,49,52,54,55,56,58,61,62,63,64,65,66,68,69,70,71,72,73,74,75,77,78,81,85,86,88,90,91,92,93,95,96,97],"10":[49,63,66,69,73,81,88,89,90,93],"100":[69,92],"1000":89,"1012":55,"1013":55,"1024":[52,72,73,88],"1045":63,"1048576":[45,49,73],"1056":63,"1063":63,"1073741824":[45,49,73],"109":63,"11":[55,63,65,66,77,81,89],"119":90,"12":[55,56,63,77,81,89,90],"120":[63,90],"123":78,"129":90,"13":[77,81],"136":89,"137":90,"138":90,"14":[81,86,89,91],"1409":93,"15":[77,81],"1502":63,"1549":63,"1556":93,"16":[63,64,72,81,90],"1691":63,"17":81,"18":[63,81],"19":[78,81],"1994":93,"1b":65,"1d":55,"1e":52,"2":[33,43,56,61,63,66,68,70,71,72,73,75,77,78,81,83,84,86,87,90,91,92,93,95],"20":[56,81,85,86,91],"2009":93,"2010":93,"2012":78,"2014":93,"2015":65,"2017":65,"2019":65,"2020":[63,67],"2022":65,"2023":93,"2048":[69,92],"2052":[84,85,86],"22":89,"224":[56,69,72,73,85,88,89,91,95],"225":[69,89],"229":89,"23":[49,55,73,78],"2341":95,"234375":89,"24":55,"244":[72,73],"248":55,"249":55,"25":[63,69,92],"256":89,"258":77,"27":[56,63],"28":63,"2802":63,"2822":77,"287":77,"29":63,"2c3":78,"3":[45,49,52,55,56,58,63,66,68,70,71,72,73,77,78,81,85,88,90,91,92,93,95,96,97],"30":[85,86],"300":[52,96],"31":63,"32":[52,63,64,72,90,93,97],"320":93,"32bit":52,"33":63,"33554432":[69,92],"346":63,"35":63,"36":63,"3677":55,"37":63,"38":90,"39":90,"3d":92,"4":[56,58,63,66,68,70,75,77,78,81,84,86,91,92],"406":89,"429688":89,"4465":93,"456":89,"468750":89,"4822":93,"485":89,"4914":93,"5":[52,56,58,59,63,65,66,70,77,78,81,84,89,90,91,92],"50":88,"50ab2c1":77,"512":[52,72,73,88,91],"523438":89,"53":78,"536870912":[45,49,73],"539":63,"56":63,"576":63,"6":[55,56,58,63,66,68,81,90,91],"622":55,"64":[64,72,92],"64bit":52,"664062":89,"7":[56,58,59,63,65,72,81,84,85,86],"72048":66,"7302":78,"8":[3,52,55,63,65,66,72,77,78,81,85,89,91],"8000":89,"8001":89,"8002":89,"84":[63,90],"9":[63,81,89],"90":89,"92":89,"9223372036854775807":68,"96":65,"abstract":[58,61,78],"boolean":[72,92],"break":[77,92],"byte":[71,72,73,88],"case":[0,1,2,46,49,53,54,56,58,61,62,66,72,91,92,93,94],"catch":[55,63],"char":[3,4,44,52,63],"class":[29,30,44,45,46,51,58,61,63,64,70,73,77,78,84,88,90,91,92,93],"const":[0,1,2,3,4,29,30,31,32,33,34,36,44,45,46,55,61,63,68,93],"default":[0,1,2,3,4,16,29,30,33,43,45,46,48,49,52,54,56,62,63,64,66,69,72,73,75,76,77,91,92,93,95,96],"do":[53,54,55,56,61,63,64,76,78,90,92,93,97],"enum":[0,1,2,42,45,46,54,70,73,93],"export":[54,66,91,95],"final":[53,56,57,59,66,84,85,86,88],"float":[49,52,63,64,68,72,84,86,90,93,96],"function":[0,1,2,3,4,46,48,49,54,55,56,58,61,62,63,66,84,85,86,88,89,90,91,92,93,96,97],"import":[52,55,56,63,64,66,75,77,89,90,91,92,94,95,96],"int":[0,3,4,36,44,45,49,52,56,63,68,69,71,72,73,75,91],"long":[49,52,53,72,77,78],"new":[0,1,2,3,4,32,33,46,48,49,54,56,58,59,61,62,63,70,73,77,83,85,86,87,89,91,92,95],"public":[0,1,2,3,4,44,45,46,47,48,49,78,93],"return":[0,1,2,3,4,23,24,29,30,31,32,33,34,35,42,43,44,45,46,54,55,56,57,58,59,61,62,63,64,69,70,72,73,84,89,90,91,92,93],"short":[55,77,78,91],"static":[48,49,53,61,63,72,73,75],"super":[44,84,90,91],"switch":95,"throw":[52,55,63],"true":[0,1,2,4,46,49,54,55,56,61,62,63,68,69,72,73,75,78,84,85,86,89,91,92,93,96,97],"try":[59,63,77,78,96],"var":68,"void":[3,4,25,26,27,28,36,37,42,44,45],"while":[54,56,66,88,89,93],A:[4,29,30,32,33,47,48,54,55,56,61,66,69,72,73,78,89,92,93],And:63,As:[54,56,63,92],At:76,But:[54,63,77],By:[29,30,51,56,75,90,91],For:[53,54,56,62,63,66,69,75,77,78,84,88,89,90,91,92,93,94,96],If:[27,33,53,54,55,56,62,63,64,66,69,70,72,75,77,84,89,91,92,93,94,97],In:[0,1,2,46,53,54,56,57,58,59,61,64,66,67,77,78,80,83,87,88,89,91,92,93,94,95],Is:[24,72],It:[52,54,55,56,57,59,61,66,75,77,88,92],Its:[61,77],Not:3,On:56,One:[54,63,77,78,88,92],Or:77,Such:54,THE:77,TO:63,That:77,Thats:63,The:[1,46,48,49,52,53,54,55,56,57,58,59,61,62,64,66,70,72,73,75,78,87,88,89,90,91,92,93,95,96],Then:[56,66,93,96],There:[4,53,54,59,61,62,65,66,78,88,89,90,91,92,93,94,95],These:[53,54,56,58,62,75,77,89,93],To:[1,46,52,56,63,64,66,75,89,90,96],Will:31,With:[63,75,77,89,93],_:[71,77,92],___torch_mangle_10:90,___torch_mangle_4847:58,___torch_mangle_5:90,___torch_mangle_9:90,__and__:68,__attribute__:43,__derive_index:68,__getitem__:68,__gnuc__:43,__init__:[62,71,72,77,84,90,91],__is__:68,__isnot__:68,__main__:[84,85,86],__name__:[84,85,86],__not__:68,__or__:68,__range_length:68,__round_to_zero_floordiv:68,__torch__:[58,63,90],__torch___pytorch_detection_ssd_src_model_ssd300_trt_engin:58,__torch___torchvision_models_resnet____torch_mangle_4847_resnet_trt_engin:58,__visibility__:43,__xor__:68,_all_:55,_aten_lowering_pass:62,_c:[72,73,96],_convolut:[63,68],_decomposit:54,_decomposition_group:54,_devic:73,_dynamo:[84,85,86],_enum:72,_export:[91,95],_input:[72,73],_jit_to_backend:96,_jit_to_tensorrt:73,_leaky_relu:54,_remove_lowering_pass:62,_rendered_examples_jupyt:87,_rendered_examples_python:87,_script:[72,73],_set:84,_shapemod:72,_theme:82,_validate_not_a_forked_repo:89,a1b:78,aarch64:59,ab:68,abi:94,abl:[53,55,61,62,67,92,93,96],about:[52,53,58,61,63,66,72,75,89,91],abov:[25,54,56,62,63,66,70,76,77,85,86,91,92],absolut:52,ac:80,acc_mod:92,acc_norm:92,acc_op:92,acc_op_convert:92,acc_ops_convert:54,acc_ops_sigmoid:92,acc_trac:92,acceler:[69,83,87,97],accept:[48,52,58,61,63,64,72,84],access:[55,61,63,67,75,92,96],accord:[54,61,73],accordingli:[75,91,92],account:[54,89],accumsan:80,accumul:[49,73],accuraci:[88,93],achiev:[56,88],aco:68,acosh:68,acoust:88,acquir:63,across:[49,52,54,55,56,73,75,91],acthardtanh:61,action:[77,92],activ:[54,63,73,77,88,92,93,97],activationtyp:[61,92],actual:[55,58,61,63,70,90,92],ad:[25,52,53,56,62,91,92],adaptive_avg_pool1d:68,adaptive_avg_pool2d:68,adaptive_avg_pool3d:68,adaptive_max_pool1d:68,adaptive_max_pool2d:68,adaptive_max_pool3d:68,adaptiveavgpool2d:91,add:[26,53,54,55,56,61,63,64,65,66,68,70,75,77,82,91],add_:[55,63,68],add_activ:[54,92],addactiv:61,addit:[54,55,63,65,72,88,91,92,95],addition:54,addlay:63,addmm:54,addmm_replac:54,address:[78,91],addshuffl:63,adipisc:[78,80],adjac:[56,77],adjust:77,adopt:88,advanc:[78,83,87,93],advis:77,aenean:80,afford:92,aforement:89,after:[52,53,55,56,62,63,64,65,67,84,85,86,89,90,92,94,95],again:[44,58,61,77],against:[52,63],agx:45,ahead:63,aim:55,algo_typ:[71,93],algorithm:[3,4,29,30,44,71,92,93],algorithm_selector:92,alias:43,align:77,align_corn:68,aliquam:80,aliquet:[78,80],all:[16,42,43,44,45,49,52,54,55,56,58,62,63,64,65,66,70,72,77,78,87,88,89,90,92,93,94,95],alloc:61,allow:[48,49,52,53,55,56,62,65,72,73,75,85,86,92],allow_gpu_fallback:[45,46,72,73,93,96,97],allow_shape_tensor:[45,49,73],allow_tf32:68,almost:63,along:[91,95],alpha:[54,68,78,92],alreadi:[52,53,54,55,63,93],also:[29,53,61,62,63,64,65,66,67,75,77,78,87,88,93],altern:[48,54,56,62,64,88],although:[54,77],altogeth:[56,75],alwai:[3,4,27,52,54,77],amet:[78,80],an:[2,3,4,48,49,52,53,54,55,56,57,58,59,61,62,63,64,65,66,67,69,71,72,73,75,77,78,84,88,89,90,91,92,93,94,95],analogu:61,analysi:[56,91],analyt:75,analytics_id:75,ancient:77,ani:[48,52,53,54,61,62,63,64,66,68,71,72,73,75,77,91,92,93],ann:77,annot:[61,63],anonym:77,anoth:[64,77,78,90],ant:80,anyon:78,anyth:[77,78,94],aot:[54,63,67,91],api:[54,56,59,61,62,63,64,72,73,76,83,84,87,88,89,91,92,93,94,96],appear:[56,77],append:[68,91],appli:[62,93],applic:[1,29,46,52,55,59,63,64,94,96,97],apply_lowering_pass:[62,91],approach:[56,95],appropri:[54,65],approri:54,apr:63,ar:[42,46,49,52,53,54,55,56,58,59,61,62,63,65,66,67,72,73,75,77,78,79,88,89,90,91,92,93,94,95,96],arab:78,arang:68,arbitrari:62,architectur:[66,67,88],archiv:66,arcu:[78,80],area:79,aren:63,arg0_1:91,arg:[53,54,62,63,71,72,81,88,91,92],arg_replacement_tupl:92,argc:63,argmax:68,argmin:68,argument:[48,52,54,55,58,61,62,63,64,72,73,77,78,91,92],argv:63,arithmet:56,around:[55,58,61,77,80,90],arrai:[3,4,33,53,73],arrayref:[45,48,49],artifact:65,arxiv:93,as_numpi:89,asin:68,asinh:68,aspect:52,assembl:[53,62,63],assign:[3,4,76],associ:[53,61,63],associatevalueandivalu:61,associatevalueandtensor:[61,63],assum:[33,91,96],atan:68,atanh:68,aten:[49,54,55,56,60,61,63,67,68,73,84,91],aten_:54,aten_op:54,aten_ops_convert:54,aten_ops_leaky_relu:54,aten_trac:[54,91],atol:52,attention_mask:91,attrdict:89,attribut:[54,55,56,58,63,77,92],auctor:80,audio:88,augu:80,author:78,auto:[44,56,61,63,77,78,93,97],autodoc:[77,78],autograd:54,automat:[54,63,77,91],avail:[52,61,62,66,75,92,97],averag:[49,52,73],avg:52,avg_pool1d:68,avg_pool2d:68,avg_pool3d:68,avgpool:91,awai:77,awaken:77,axi:68,b0:88,b:[65,66,68,78,89,95],b_hh:68,b_ih:68,back:[55,56,58,59,63,72,77,90],back_insert:44,backend:[54,62,73,76,83,84,86,87,91,96],backend_kwarg:84,background:[77,90],backlink:77,backward:92,bar:[75,77],base:[37,50,54,58,64,66,69,70,71,72,77,86,88,90,91,93,95],bash:66,basi:77,basic:[52,56,78,87,89,92],batch:[3,4,44,69,85,86,89,91,92,93,97],batch_norm:[54,61,68],batch_siz:[44,93],batched_data_:44,batchnorm:55,batchtyp:44,bathroom:77,bazel:[59,66],bazel_vers:66,bazelbuild:66,bazelisk:66,bazelvers:66,bdist_wheel:66,beat:78,becaus:[61,63,66,69,90,92],becom:61,bee:77,been:[53,61,63,78,95],befor:[49,55,56,59,61,63,66,67,73,89,92],beforehand:63,begin:[44,66,77,84,92],beginn:90,begun:77,behav:79,behavior:[49,56,72,73,91,92,95],behind:77,being:[63,92],belong:[54,77],below:[33,54,56,61,62,63,64,66,77,89,92],benchmark:68,benefit:[61,63],bert:86,bertmodel:[86,91],besid:77,best:[66,77,92],beta:[54,68,73,92],better:[88,90],between:[55,56,61,66,77,78,93],bia:[55,63,68],bibendum:80,bibliograph:78,bigger:77,bin:66,binari:[44,93],binary_data:89,bind:[3,4,33,44,73,77],bird:89,bit:[49,61,63,72,73,92],bitbucket:75,bitbucket_url:75,bitwise_not:68,blandit:80,blank:77,blob:[60,75,93],block0:55,block1:55,block:[52,53,55,56,81],blue:77,bmm:68,bodi:[77,78],bold:77,bool:[0,1,2,3,4,24,27,30,31,42,44,45,46,49,55,61,63,68,69,70,72,73,75,93],border:77,both:[54,56,66,69,75,77,90,93],bottom:75,bound:72,boundari:[56,70,71],box:[77,91],bracket:77,branch:66,bread:77,brief:56,briefli:90,brontosaurus:77,browser:77,bsd:[42,43,44,45],buffer:[3,4,92],bug:66,bui:78,build:[29,30,35,49,52,53,57,59,61,63,72,76,81,85,86,92,93],build_fil:66,build_model:92,buildcommandarg:65,builddirectori:65,builder:92,builderconfig:45,buildroot:65,built:[33,52,58,59,66,73],builtin:92,button:[75,77],bytearrai:[73,92],c10:[0,1,45,46,48,49,63,93],c:[42,43,44,45,52,59,64,65,68,69,78,89,94,97],c_api:60,c_str:[61,63],cach:[3,4,29,30,44,52,54,63,69,71,92,93],cache_:44,cache_fil:[44,71,93],cache_file_path:[3,4,29,30,44],cache_file_path_:44,cache_size_:44,cachecalibr:[71,93],cackl:78,calcul:[48,53,56,63],calibr:[3,4,29,30,44,49,52,63,71,73,93],calibration_cache_fil:[29,30,93],calibration_dataload:[30,93],calibration_dataset:93,calibrationalgo:[71,93],call:[29,30,32,49,54,55,58,61,63,69,73,77,84,85,86,88,90,91,92,96],call_funct:[54,62,91,92],call_modul:[54,95],call_spec:95,callabl:72,caller:62,callmethod:90,can:[0,1,4,29,30,34,46,47,48,49,52,53,54,55,56,57,58,59,61,62,63,64,66,72,73,75,77,83,84,85,86,87,88,89,90,91,92,93,94,95,96],canada:78,cannot:[48,55,56,66,72,73,76,90,91,92,95],canon:75,canonical_url:75,capability_valid:54,capabl:[17,45,49,52,58,72,73,96],capbility_valid:54,capit:77,caption:[77,80],captur:91,care:54,cast:[3,4,55],cat:[56,66,68],categor:54,categori:54,caught:55,caus:[61,66,75,84,85,86],cd:[66,89],cdll:63,ceil:68,ceil_mod:68,cell:78,centercrop:89,cerr:63,certain:[54,66,84,92],cfg:56,chain:61,challeng:89,chanc:61,chang:[29,55,56,59,62,65,73,75,89,92,93],changelog:81,channel:[2,72,76],channel_last:[72,73,88],channels_last:72,charact:77,check:[0,1,31,46,52,55,61,63,66,73,89,92,94],check_method_op_support:73,check_method_operator_support:[41,45,50],checkmethodoperatorsupport:63,child:78,children:92,choic:[66,71],choos:[54,90,92],cifar10:93,cifar:93,clamp:68,clamp_max:68,clamp_min:68,class_count:89,classif:[63,88,90],classifi:[78,88],classification_index:89,classmethod:72,clean:[56,62,65,77,84,85,86],cleanli:56,clear:44,cli:[52,64],click:65,clickabl:77,clone:[62,65,68],cloned_placehold:62,close:63,closer:55,closet:77,cmake:65,cmake_build_typ:65,cmake_cuda_flag:65,cmake_cxx_flag:65,cmake_module_path:65,cmakecommandarg:65,cmakeset:65,cnn:88,co:[68,78,88],coalesc:62,code:[54,56,59,62,63,67,76,78,84,85,86,87,90,91,92,93,95],coeffici:88,collapse_navig:75,collat:78,collect:[56,63,64,73],colon:77,color:[24,27,70,77],colored_output_on:[27,42,70],column:78,com:[60,63,66,84,85,86,89,93,94,95],combin:[56,92],come:[66,76,89,92],command:[52,63,66,77,78,89,90],comment:[66,77],commodo:80,common:[53,54,55,69,77,92],common_subexpression_elimin:55,commonli:78,commun:[49,52,63,65,73],compar:[54,64,92],comparis:[0,2],comparison:[1,46],compat:[0,1,46,55,58,65,66,73,92],compil:[31,34,41,45,49,50,52,54,55,56,58,61,62,64,69,70,72,73,75,89,90,92,93,94,96,97],compilation_kwarg:86,compilationset:84,compile_engine_and_inf:[84,85,86],compile_spec:[91,93,97],compilegraph:[63,93],compilesepc:33,compilespec:[3,4,21,32,34,41,45,50,56,63,73,93,97],compilespecstruct:50,complet:[56,63,90],complex:[47,49,64,66,90],complianc:52,compliat:93,complic:66,compon:[57,59,90,94],compos:[89,90,92,93],composit:63,compound:88,comput:[49,77,88,92,93],concaten:54,conceiv:77,concept:87,concorr:89,condimentum:80,condit:[54,56,77],conf:[75,82],confidence_scor:89,config:[66,69,89,92],configur:[32,34,48,62,63,66,67,72,73,81,89,91,93],configurationtyp:65,configureset:65,congu:80,connect:[55,73,77,89,97],consectetur:[78,80],consecut:56,consid:[56,63,73],consider:89,consist:[55,77,92],consol:52,consolid:[56,90],constant:[53,55,56,63],constant_pad_nd:68,constexpr:[0,1,2,45,46],constraint_dim:91,construct:[0,1,2,3,4,46,48,49,53,55,57,59,61,63,71,72,77,78,92,93],constructor:[0,2,46,48,49,58,90],consult:76,consum:[4,53,90],contact:78,contain:[30,31,52,53,54,55,56,61,63,66,69,72,77,78,89,90,92,93,94],content:[81,89,93],context:[53,57,58,59,70],contextnet:88,contigu:[2,48,49,52,72,73],continu:[77,92,94],contributor:63,control:[62,90,92],conv1:[63,90],conv2:[63,90],conv2d:90,conv:[49,52,63],conval:80,convect:48,conveni:[62,86,88,93],convent:33,converison:92,convers:[54,55,56,58,63,72,73,91,92],conversionctx:[61,63],convert:[3,4,31,32,34,52,55,56,57,59,64,67,72,73,85,86,88,94,95,96],convert_activ:54,convert_method_to_trt_engin:[41,45,50,72,73,96],converter_implement:54,converter_util:54,convertersupport:54,convertgraphtotrtengin:63,convien:49,convienc:[3,4,49],convolut:[73,93,97],convtert:92,coordin:59,copi:[44,61,68,71,78,89,92],copy_:68,copyright:[42,43,44,45,63,78],core:[45,52,55,56,59,63,72,97],core_id:72,corpor:[42,43,44,45],corpu:88,correct:[58,66,75],correctli:66,correctness_atol:69,correctness_rtol:69,correspond:[54,61,66,92,95],cosh:68,could:[56,85,86,92],count_include_pad:68,coupl:[53,59,92,94],cout:63,cover:[87,88],cp:66,cpp:[14,15,42,43,44,45,51,55,59,63,66,93],cpp_frontend:93,cppdirectori:50,cppdoc:63,cpu:69,cra:80,creat:[29,30,33,52,53,54,56,58,61,63,67,73,77,89,91,92,95],create_exported_program:95,credit:63,criteria:[56,57,59],cross:77,cs:93,csrc:[55,60],cstddef:93,ctestcommandarg:65,ctrl:65,ctx:[61,63],ctype:63,cuda113:66,cuda:[49,58,63,64,65,66,69,72,89,91,92,93,95,96],cuda_graph_batch_s:[69,92],cuda_runtim:[21,45],cudafloattyp:63,cudasetdevic:36,cudnn:65,cudnn_en:68,cudnn_root_dir:65,cumsum:68,curabitur:80,curl:[66,77],current:[23,54,56,58,61,62,65,66,69,73,75,91,92],cursu:80,custom:[52,62,66,83,87,92],custom_class:[21,45],custom_mapp:92,customclasshold:[45,48],cut:77,cxx11:94,d:[52,77,78,97],d_silence_experimental_filesystem_deprecation_warn:65,dapibu:80,data:[0,2,3,4,29,30,44,46,48,49,52,53,56,57,59,61,64,68,69,71,72,73,77,81,88,92,93],data_dir:93,data_item_1:76,data_typ:89,dataclass:[84,92],dataflow:[61,63],dataload:[4,29,30,44,49,71,93],dataloader_:44,dataloadercalibr:[71,93],dataloaderopt:93,dataloaderuniqueptr:[4,44],dataset:[29,71,88,93],datatyp:[1,21,38,45,46,48,49,50,64,72,73,89],datatypeclass:50,date:[62,78],david:78,dbg:66,dcmake_build_typ:66,dcmake_module_path:66,dead_code_elimin:55,deal:61,debug:[16,27,45,49,52,61,62,65,70,73,84,85,86,96],debugg:[52,73],decid:72,declar:66,decomp_t:91,decompos:54,decomposit:54,deconvolut:97,decor:[54,62,92],dedic:[55,78],deep:[61,67,75,93,97],deeplearn:[60,92],def:[54,62,64,77,84,89,90,91,92],defin:[0,1,2,3,4,33,43,46,47,48,49,51,52,54,63,64,72,75,84,86,88,90,92,93],definit:[51,61,77],deiti:77,delet:[0,1,2,45,46,55],delimit:55,demo:[77,93],demonstr:[77,78,79,88,89,93],demostr:88,denot:77,dep:[65,66],depend:[29,35,53,54,59,63,64,65,89,92,94],depickl:58,deploi:[63,67,89,93],deploy:[52,63,64,88,89,93,94,97],deprec:[54,68,92],depth:[75,88],descclassnam:77,descnam:77,describ:[49,56,61,83,87,89,90,91,96],descript:[56,78],deseri:[63,72,73,95],design:[88,92,97],desir:[62,78,93],desktop:65,destini:78,destroi:[61,78],destructor:61,detail:[54,63,89,90,91,92,94],detect:[48,58],determin:[55,91,92],determinist:68,dev0:77,develop:[63,65,66,67,77,78,92],deviat:52,devic:[21,33,36,38,45,49,50,52,58,64,68,69,71,72,73,88,93,96,97],device_typ:[45,46,72,93,96,97],deviceclass:50,devicetyp:[21,38,45,46,50,72,73,93,96,97],devicetypestruct:50,diam:80,dict:[54,72,73],dictionari:[54,72,73,84,96],dictioneri:54,dictum:80,dictumst:80,did:77,didn:77,differ:[29,54,55,56,59,66,67,75,90,91,92],dignissim:80,dilat:68,dim0:68,dim1:68,dim:[68,69,89,91,92],dim_int:68,dim_intlist:68,dimens:[48,55,69,88,91,92],direct:[62,81,94],direct_output:62,directli:[54,61,62,65,66,67,71,83,84,87,91,93,95],directori:[18,19,20,21,42,43,44,45,50,54,65,66,93],disabl:[52,54,70,75,76],disable_memory_format_check:72,disable_pass:54,disable_tf32:[45,49,73,93],disallow:62,disclos:66,disconnect:77,discret:77,discuss:[54,89],disk:95,displai:[52,62,70,75],display_github:75,display_gitlab:75,display_vers:75,dist:66,distdir:66,distribut:[63,72,93,94],div:[56,68],div_:68,div_lgamma:56,divid:54,divisor_overrid:68,django:76,dl:77,dl_open:94,dla:[1,45,46,49,52,67,72,73],dla_cor:[45,46,52,72,93,96,97],dla_global_dram_s:[45,49,52,73],dla_local_dram_s:[45,49,52,73],dla_sram_s:[45,49,52,73],dla_standalon:52,dlacor:52,dll:52,do_not_merg:56,doc:[59,60,66,75,76,77,82],docker:89,docsrc:59,docstr:[64,77,78,91],document:[42,43,44,45,50,54,59,63,75,77,78,82,89,90,93,94,96],docutil:[77,78],doe:[43,44,55,56,61,62,77,85,86,92,93],doesn:[63,66,77,90],dolor:[78,80],domain:[48,72,78,93],don:[61,75,77,78,89,91,92,93],done:[53,56,59,89],donec:[78,80],dont:42,dothismethod:77,dotpai:76,dotpayprovid:76,doubl:[45,48,49,52,72,73,77],down:[66,75,92],download:[66,81,84,85,86,87,89,93],downstream:88,doxygen_should_skip_thi:[44,45],dpython:[72,73],dram:52,dream:78,driver:66,drop:[66,75],dt:77,dtensorrt_root:66,dtorch_dir:66,dtyep:69,dtype:[45,48,49,52,64,68,69,72,73,86,88,91,92],dual:77,due:[3,4,66,76,77],dui:[78,80],dummi:91,dump:[37,52,66],dump_build_info:[38,45,50,72],dump_lowering_pass:62,durat:77,dure:[49,52,56,61,71,88,91,93,94],dyn_range_fn:54,dynam:[48,49,54,69,72,73,92],dynamic_batch:[69,92],dynamic_dim:91,dynamic_input:91,dynamic_shap:67,dynamo:[67,84,85,86,91,95],dynamo_convert:54,dynamo_tensorrt_convert:54,e:[29,30,52,55,61,63,65,66,69,72,90,92,93],each:[3,4,49,53,54,55,56,58,61,63,66,69,75,77,92],eagerli:91,ear:77,earli:92,earlier:54,eas:43,easi:[52,53,55,63,93],easier:[57,59,61,63,92,93],easiest:66,easili:[3,4],echo:77,edg:77,edit:[65,75],edu:93,effect:[55,63,75,88,92,93],effici:61,efficientnet:88,efficitur:80,eg:[54,89],egesta:80,eget:80,either:[47,48,52,61,62,63,64,66,72,73,75,77,90],el:68,eleifend:78,element:[58,77,78,81,92],element_typ:44,elementum:80,elementwis:54,eliminate_dead_cod:62,elit:[78,80],elk:77,els:[43,44,48,73,77,78],elu:[54,68],emb:[33,52,73,78],embed:[52,54,58,68,73,77,97],embed_engine_in_new_modul:[41,45,50,73],embedding_param_valid:54,emit:53,emphasi:77,empti:[49,69,73,78,90],emum:[16,17],en:75,enabl:[3,4,24,49,52,54,56,57,59,69,70,71,73,75,85,86,92],enable_precis:63,enabled_precis:[45,49,63,64,72,73,84,85,86,89,93,96,97],enalbed_precis:97,encod:[58,88],encompass:73,encount:[56,66,84,85,86,91],encourag:89,end:[44,52,61,62,63,68,73,77,84,85,86,93],end_dim:[63,68],endif:[43,44,45],energi:77,enforc:63,engin:[0,1,17,32,33,34,45,46,48,49,52,53,56,57,59,62,63,64,67,69,72,73,75,85,86,93,94,96,97],engine_converted_from_jit:63,enginecap:[38,45,49,50,72,73,96],english:88,enhanc:77,enim:80,ensur:[29,55,56,62,65],enter:[53,72],entir:77,entiti:77,entri:[49,61],entropi:[29,30,93],entropy_calibr:71,entropy_calibration_2:[71,93],enumer:[0,1,2,16,17,46],environ:[89,92],ep:[68,95],eq:[68,77],equat:77,equival:[32,57,59,61,63,73,85,86,90,93],equivil:34,erat:80,erf:68,eric:77,ero:80,error:[16,49,52,53,55,59,63,66,70,73,77,91,92],essenc:77,essenti:92,est:80,et:80,etc:[75,77,92,97],etiam:80,eu:80,euismod:80,eval:[63,64,84,85,86,89,91,95],evalu:[54,57,58,59],evaluated_value_map:[53,61],even:63,event:48,everi:[56,63,69],everyth:16,evolv:62,ex:[0,1,2,33,46,73,78,80],exact:89,exactli:91,examin:92,exampl:[48,54,56,58,59,61,63,64,65,67,70,72,73,75,76,78,81,83,84,85,86,87,89,90,92,93,94],example_tensor:72,exceedingli:77,except:92,exception_elimin:55,excerpt:78,excit:88,execpt:55,execut:[33,49,52,55,57,58,59,63,66,69,72,73,89,90,91,92,93],execute_engin:[58,63],exert:77,exeuct:58,exhaust:63,exist:[4,31,32,34,54,66,71,72,73,88,92,93],exit:[84,85,86,89],exp:68,expand:[55,68],expand_a:68,expanded_asset:66,expect:[48,55,61,63,64,72,88],expected_op:54,experiment:[73,92,95],experimental_decomposit:91,explain:92,explan:92,explic:44,explicit:[0,1,2,3,4,45,46,55,67,69,77,92,93],explicit_batch_dimens:[69,92],explicit_precis:69,explicitli:[56,57,59,64,73,93,96],explict:44,explictli:0,explor:[87,91],expon:68,exportedprogram:95,expos:93,express:77,ext:[77,78],extend:[57,59,61,63,65,68,88],extens:65,extent:[63,67],extern:[75,77],extra:63,extract:[54,62,63,88],extrem:77,ey:77,f16:[52,63,97],f32:52,f:[62,66,77,90,92],facilisi:80,fact:66,facto:77,factori:[4,29,30,93],fail:[54,63,97],fake_quantize_per_channel_affin:68,fake_quantize_per_tensor_affin:68,fall:[54,72],fallback:[52,57,59,61,97],fals:[0,1,2,3,4,44,45,46,49,62,63,68,69,72,73,75,76,77,78,84,92,93,96],fame:80,familiar:89,far:[77,92],fashion:[63,88],fast:[49,52,73],faster:88,faucibu:80,fc1:[63,90],fc2:[63,90],fc3:[63,90],fc:[49,52,55],feat:[63,90],featur:[52,56,63,88,92,93,96],fed:[3,4,48],feed:[29,30,63],feedforward:88,feel:67,feli:80,feugiat:[78,80],few:[66,72,91,92],field:[3,4,69,72,93],fifth:78,figur:[56,78,80],file:[0,1,2,3,4,5,6,7,8,9,10,11,12,46,47,48,49,52,54,56,58,59,63,65,66,69,71,72,73,75,76,78,82,89,91,92,93],file_path:52,filepath:65,find:[4,63,66,92],finder:66,finibu:80,finish:92,first:[48,53,55,63,64,77,78,84,89,91,92,93],firstli:89,fit:77,fix:[49,77,92,97],fixed_s:[45,49],flag:[52,56,57,59,64,66,71,94],flaten:49,flatten:[45,47,63,68,90,91],flatten_convert:63,flesh:89,float16:[52,72],float32:[48,49,52,72,73,91,92],float64:[72,73],float_int:68,floor:68,floor_divid:68,floordiv:68,flow:[61,77,88,90,92],flox:77,flush:77,fly:90,fmod:54,focus:54,fold:78,folder:[65,92],follow:[33,52,54,56,58,62,63,65,66,73,75,77,78,82,83,87,88,89,90,91,92,93,94,95],foo:[77,78,92],foo_kwarg:92,foo_nod:92,forc:[52,73,75,92],force_fp32_output:92,forced_fallback_op:56,form:[53,54,64,72,77,89],format:[33,45,48,49,52,64,68,72,73,77,78,88,89,95],forth:78,forum:66,forward:[29,30,32,33,56,58,61,63,64,72,73,84,90,91,93,96],found:[42,43,44,45,63,66,77,93,94],four:[77,78],fp16:[0,48,49,52,63,64,67,69,92,97],fp32:[0,48,49,52,67,73,88,89,92,93],fp64:0,frac:77,freed:61,freeze_modul:55,friend:45,fringilla:80,from:[0,1,2,3,4,29,30,44,46,48,49,52,53,54,55,56,57,58,59,61,63,65,67,69,73,75,76,77,78,86,88,89,90,91,92,93,95],from_pretrain:[86,91],from_tensor:[72,92],front:62,frontend:[64,67,83,85,86,87],fssl:66,fstream:[20,44],full:[45,49,52,61,63,70,84,85,86,89,93,94,97],fulli:[31,52,55,63,73,93,97],further:[56,92],fusc:80,fuse_addmm_branch:55,fuse_flatten_linear:55,fuse_linear:55,fusion:[61,62,92],futur:[73,91,92],fx2trt:69,fx2trt_exampl:92,fx:[54,62,64,67,72,91,95],g:[29,30,52,55,65,66,69,72,77,92,93],g_:77,galleri:[84,85,86,87],gamma:68,gatewai:76,gather:56,gaurd:43,gcc:[59,63],ge:68,gear:93,gener:[3,4,29,52,54,55,58,59,61,62,63,65,66,69,75,77,78,81,84,85,86,87,90,91,92,93],get:[0,1,2,3,4,23,35,44,46,55,56,61,62,63,66,70,72,88,89,92,93],get_batch:71,get_batch_impl:44,get_batch_s:71,get_build_info:[38,45,50,72],get_cache_mode_batch:71,get_decomposit:91,get_is_colored_output_on:[39,42,50,70],get_logging_prefix:[39,42,50,70],get_output:92,get_reportable_log_level:[39,42,50,70],getattr:[55,58,63,90],getbatch:[3,4,44],getbatchs:[3,4,44],getdimens:[61,63],getitem:54,getoutput:[61,63],git:81,github:[60,63,66,75,84,85,86,89,93,94,95],github_url:75,gitlab:75,gitlab_url:75,give:[75,77,92],given:[48,49,52,54,55,63,64,69,71,72,73,90,91,92,96],global:[26,52,63],gm:62,gnu:66,go:[44,55,56,63,67,84,85,86,88,89,90,92],goal:61,goe:[77,92],good:[44,61,77,92],goodger:78,googl:75,got:[63,77],gpu:[1,32,34,36,45,46,52,63,72,73,89,92,93,96,97],gpu_id:[36,45,46,52,72,73,93,96,97],graph:[16,31,32,34,45,49,52,53,54,56,57,59,61,62,63,67,69,70,73,85,86,88,90,91,92],graph_input:[45,49],graph_modul:[62,69,72,91],graphinput:[21,38,45,49,50],graphinputsstruct:50,graphmodul:[62,64,69,72,91,95],gravida:80,great:[63,77],greater:70,greedi:56,group:[68,77,78],grpc:89,gru_cel:68,gt:68,gtc:67,guangzhou:78,guarante:56,guard:55,guard_elimin:55,gui:77,guid:[76,87],gulf:89,gz:[77,78,93],h:[0,1,2,3,4,5,6,7,8,9,10,11,12,15,46,47,48,49,50,51,52,55,63,93],ha:[49,53,54,55,56,57,59,61,62,63,65,69,77,78,88,90,91,92,93,95],habit:80,habitass:80,hac:80,hack:44,had:54,hakaimagazin:89,half:[52,63,64,72,77,84,85,89,93,96,97],hand:89,handl:[55,56,58,91,92,95],happen:[90,91,92],hardtanh:[54,61,68],hardtanh_:68,hardwar:97,has_batch_dim:69,hash:66,have:[29,33,44,52,53,54,55,56,61,62,63,64,66,67,69,72,73,77,85,86,88,89,90,91,92,93],haven:63,header:[63,75,77,78,89],heart:78,heaven:77,heck:77,heh:78,hehe:78,height:77,help:[27,52,53,54,61,63,88,92,94],helper:61,henc:[54,95],hendrerit:80,here:[44,53,56,58,63,66,75,77,78,89,90,91,92,93,94,95],hermet:66,hexagram:77,hfile:50,hi:[68,77,78],hidden:[43,75],high:[48,55,56,75],higher:[55,75,77,90],highli:[88,89],highlight:77,hinton:93,hit:56,hold:[46,47,48,53,61,93],holder:[58,79],holi:77,home:66,hood:59,hope:78,host:[49,52,66,73,89],how:[3,4,66,77,79,81,84,88,89,90,91,94,96],howev:[29,66,75,76,89,91],html:[60,66,77,90,93],html_theme:82,html_theme_opt:75,html_theme_path:82,http:[60,63,66,75,77,84,85,86,88,89,90,93,94,95],http_archiv:66,httpclient:89,hub:89,huggingfac:88,human:77,humankind:78,hx:68,hybrid:73,hyperlink:77,hyphen:77,i8:52,i:[52,55,61,63,68,77,78,90,93],iaculi:80,icon:[75,77],id:[36,45,52,72,75,76,80,97],idea:[55,77],ident:[52,62],identifi:[54,56],idx:68,ifndef:[44,45],ifstream:44,ignor:72,iii:78,iint8calibr:[3,4,29,30,44,45,49,73,93],iint8entropycalibrator2:[3,4,29,30,44,93],iint8minmaxcalibr:[29,30,93],ilay:61,illustr:[54,88,92,95],imag:[89,93],imagenet:88,imagenett:88,images_:93,img1:89,img:89,img_path:89,imperdiet:80,impl:54,implement:[3,4,55,56,58,63,76,91,92,93,94],impli:54,implic:55,implicit:[68,69,77,92],implicitli:72,implictli:72,improv:78,in_shap:63,in_tensor:90,incas:[44,91],includ:[13,15,16,35,37,42,43,44,45,51,52,54,56,57,58,59,62,63,66,69,75,77,83,87,90,92,93],includedirectori:50,includehidden:75,incompat:66,incorpor:78,indent:77,index:[33,60,62,67,68,73,75,81,93],indic:[68,75,77],indirect:77,individu:[54,64],inetworkdefinit:53,infer:[55,63,72,73,83,84,87,88,91,92,93,95],inference_output:89,inferenceservercli:89,inferinput:89,inferrequestedoutput:89,info:[16,32,34,45,52,61,63,70,72],inform:[25,33,35,37,48,52,53,56,58,62,63,66,67,69,70,72,77,90,91,92,93,96],infrastructur:[89,93],ingest:59,inherit:[50,92,93],inheritenviron:65,initi:[77,84,85,86],injuri:77,inlin:[0,1,2,3,4,29,30,44,46,48,55,63,78,81,95],inner:[49,78,88],input0:[63,64],input1:[63,64,91],input2:[63,91],input:[3,4,21,29,33,38,44,45,47,49,50,52,53,55,56,58,61,62,63,64,68,69,70,72,73,78,84,88,89,90,91,92,93,95,96,97],input_0:[58,63],input_:54,input__0:89,input_binding_nam:[33,45,73],input_data:[64,90],input_file_path:[52,97],input_id:91,input_is_dynam:45,input_nam:[69,91,92],input_s:[56,63],input_scal:68,input_shap:[93,97],input_signatur:[45,47,49,64,73],input_spec:[52,69,92],input_tensor_spec:[69,72,92],input_v:[54,92],inputclass:50,inputrang:[56,63],inputs_bs2:91,inputtensorspec:[69,72,92],insert:[62,63,93],inserting_aft:62,inserting_befor:92,insid:[77,89],inspect:[61,63,90],instal:[63,67,81,89,94],installroot:65,instanc:[55,62,63,71,88,90],instance_norm:68,instanti:[57,58,59,61,63],instatin:[0,1,2,46],instead:[49,52,53,55,63,66,94],instnanti:58,instruct:[56,57,59,63,66,89,91,92],insur:66,int32:[72,73,86,88,91],int64:[0,72,73],int64_t:[45,46,48,49,93,97],int8:[0,44,48,49,52,67,72,73,93,97],int8_t:45,int8cachecalibr:[20,29,40,44,50],int8cachecalibratortempl:50,int8calibr:[3,20,30,40,44,50],int8calibratornamespac:50,int_float:68,integ:[72,80],integr:[67,84],intend:[66,84,85,86],intent:[55,77],interact:[77,84,85,86],interdum:80,interest:[55,77],interfac:[0,1,2,46,58,59,61,93],interfer:77,intermedi:[16,49,52,70,73,90,91],intern:[1,16,46,61,63,70,77],internal_error:70,internalerror:70,interpol:77,interpret:[58,77,92],interv:72,intro_to_torchscript_tutori:90,introduc:[88,91,92,95],invoc:84,invok:[62,63,90,92],involv:91,io:[44,89],iostream:[20,21,44,45,63],ipso:77,ipsum:[78,80],ipynb:[84,85,86],ir:[54,57,59,61,64,72,84,85,86,90,95],is_aten:69,is_floating_point:68,is_train:93,iscustomclass:61,ishap:49,ishapelay:73,isinst:[62,92],isn:[75,77],issu:[3,4,63,66,84,85,86,91,95],issubclass:62,istensor:61,istream_iter:44,it_:44,ital:77,item:[76,78],itensor:[53,61,63,92],iter:[20,44,49,52,53,71,72,73],its:[29,53,54,56,58,61,66,77],itself:[0,1,2,46,52,55,66,89,96],iv:78,ivalu:[45,47,49,53,58,61,63],jan:78,jetpack:66,jetpack_5:66,jetpack_x:66,jetson:88,jit:[31,32,33,34,45,47,49,52,53,55,56,57,58,59,60,61,63,64,72,73,89,90,95,96],jp_workspac:66,jpg:89,json:65,jump:89,jupyt:[84,85,86,87],just:[44,45,54,55,56,63,64,67,70,77,79,88,90,92,94,96],justo:[78,80],k:[68,93],kbool:[0,45],kchannelslast:[2,45],kchar:[0,45],kclip:61,kcontigu:[2,45,48],kcpu:[1,46],kcuda:[1,46,56,63],kdebug:[16,42,44],kdla:[1,45,46,97],kdla_standalon:[17,45],kdoubl:[0,45],keepdim:68,kei:[54,77,89,90],kept:78,kernel:[48,49,52,61,72,73,92],kernel_s:68,kerror:[16,42],keyboard:77,keyword:[62,72,73,84,86],kf16:[93,97],kfloat:[0,45,49],kgpu:[1,45,46],kgraph:[16,42,55],khalf:[0,45,63],ki8:93,kind:[53,54,92],kinfo:[16,42,44],kint:[0,45],kinternal_error:[16,42],klong:[0,45],know:[42,61,75,77],knowledg:77,known:95,kriz:93,krizhevski:93,ksafeti:[17,45],kstandard:[17,45,49],ktest:93,ktrain:93,kunknown:[0,2,45],kwarg:[54,71,72,88,91,92],kwarn:[16,42],l:68,label:[77,88,89,93],lacinia:80,lack:[56,57,59,92],lacu:80,laid:63,lambda:[61,63,77,89],lang:76,languag:[76,77,78,89,90],laoreet:80,larg:[57,59,63,75,77,88,93],larger:[56,75,88],largest:68,last:[2,55,72,92],lastli:89,later:[29,63,95],latest:[65,66,75],launch:89,layer:[46,49,52,53,54,55,61,62,63,73,88,89,91,92,93,97],layer_norm:[54,68],layout:[2,48,68,72,73],ld_library_path:66,ld_preload:94,ldd:66,le:68,lead:77,leader:77,leaky_relu:[54,68],leaky_relu_:68,learn:[63,67,89,93,97],leas:78,least:[77,78],leav:[55,62],lectu:[78,80],left:[75,77],legacy_calibr:71,legend:77,len:[62,68],lenet:[63,90],lenet_script:[63,90],lenetclassifi:90,lenetfeatextractor:90,length:[3,4,44,68,78,91,92],leo:80,let:[46,52,55,61,72,73,75,77,88,89,92],letter:[78,88],level:[23,25,26,39,42,44,50,54,55,56,59,70,73,81,89,90,92],levelnamespac:50,leverag:[83,87,92,93],lgamma:56,lib:[54,55,63,65,66],libero:[78,80],librari:[35,42,43,44,45,52,54,57,58,59,61,63],libtorch:[4,37,61,63,65,66,93],libtorch_pre_cxx11_abi:66,libtorchtrt:[52,63,66],libtorchtrt_plugin:94,libtorchtrt_runtim:94,licens:[42,43,44,45,63,65],light:77,ligula:80,like:[52,53,54,55,58,61,63,64,66,76,77,89,90,92,93,94],limit:[55,70,76,93],line:[52,63,78],linear:[2,56,68,72,90],link:[52,53,62,63,67,75,76,81,94],lint:62,linux:[59,63,66],list:[18,19,20,21,31,49,51,53,56,58,61,62,63,64,66,68,69,71,72,73,81,89,92],listconstruct:[53,56,58,63],listunpack:[58,63],liter:78,literal:78,literal_block:77,live:[61,77],ll:92,lo:68,load:[52,56,58,63,64,71,73,88,89,92,93,94,95,96],load_librari:94,loading_data_recip:93,loborti:[78,80],local:[52,55,63,75],localhost:89,locat:[54,62,66,93],lock:76,log:[15,16,19,20,38,44,50,51,55,61,67,68,69,72,85,86,92],log_debug:61,logger:[62,70],logger_level:69,loggingenum:50,logic:92,login:89,logist:92,loglevel:70,logo_onli:75,lone:78,longer:[75,94],look:[53,55,89,90,91,93,96],loop:[56,92],loop_unrol:55,lorem:[78,80],lose:75,loss:[88,93],lot:61,low:[48,92],lower:[16,54,67,69,70,72,78,85,86,88,91,92],lower_exampl:92,lower_graph:55,lower_precis:[69,92],lower_tupl:55,loweralltupl:55,lowerprecis:[69,92],lowersimpletupl:55,lstm_cell:68,lt:68,ltorchtrt:94,luctu:80,lvl:[25,26,42],m:78,machin:[58,89,93],macro:[5,6,7,8,9,10,11,12,15,18,20,21,42,44,45,50,51],mad:77,made:[55,57,59,77],maecena:80,magna:80,mai:[53,56,58,59,63,64,73,77,78,84,85,86,89,90,92,93],main:[55,56,57,58,59,61,63,75,77,79,92],mainli:92,maintain:[56,58,61],major:[59,92],make:[53,54,63,64,66,77,79,83,87,88,89,92,93,97],make_data_load:[4,93],make_int8_cache_calibr:[40,44,50,93],make_int8_calibr:[29,40,44,50,93],malesuada:80,man:[77,78],manag:[49,52,53,57,59,61,63,65,70,72,73],mangag:55,mani:[56,75,77,78,92],manipul:62,mantissa:[49,73],manual:[76,77,92],map:[1,46,53,54,55,57,59,61,63,84,88,89,92,93,96],mapper:92,mark:[55,56,75],marknodesforfallback:55,markup:[78,81],markup_process:77,mask:68,masked_fil:68,massa:80,master:[60,66,93,94],mat1:54,mat2:[54,68],match:[55,66],math:81,matmul:[54,55,63,68],matrix:60,matter:92,matti:78,matur:59,mauri:[78,80],max:[48,52,61,68,72,75,91],max_batch_s:[69,89,92],max_c:52,max_h:52,max_input_shap:69,max_n:52,max_pool1d:68,max_pool2d:[63,68,90],max_pool3d:68,max_shap:[45,48,64,72,73,88,91,92],max_val:[61,68],max_w:52,max_workspace_s:[69,92],maximu:80,maximum:[48,49,52,69,73,85,86,89,92],mayb:77,mb:52,md:60,me:[77,78],mean:[54,56,61,67,68,69,84,89,91,92],mechan:[61,88,92],medium:77,meet:72,member:[46,47,48,49,72],memeori:2,memori:[20,21,44,45,55,61,63,64,72,73],memory_format:[68,72],memoryformat:[2,45],men:77,mental:77,mention:[54,91],menu:[52,75,77],menuselect:77,merg:56,messag:[16,25,26,52,70],meta:[81,92],metadata:[49,52,58,61,73,75,95],meth:77,method:[31,32,33,34,48,52,55,61,63,66,72,73,77,88,90,96],method_nam:[31,34,45,52,63,72,73],metu:80,mi:80,microsoft:65,middl:77,might:[55,66,75,91],min:[48,52,61,68,72,91],min_acc_module_s:69,min_block_s:[45,49,56,73,84,85,86],min_c:52,min_h:52,min_input_shap:69,min_n:52,min_shap:[45,48,64,72,73,88,91,92],min_val:[61,68],min_w:52,mind:77,mine:77,minim:[69,93],minimum:[48,49,52,56,70,73],minmax:[29,30,93],minmax_calibr:71,minor:65,minut:[84,85,86],misbuild:75,miss:[63,77],mix:91,mkdir:66,mm:89,mmb:77,mobilenet_v2:96,mobilenetv2:88,mock:91,mod:[52,56,63,81,92,93],mode:[64,91,92,93],mode_:93,model:[52,56,57,58,59,63,64,67,69,70,83,87,90,91,93,96],model_half:84,model_nam:89,model_output:91,model_repositori:89,model_torchtrt:70,model_trt:70,modif:[54,62],modifi:[56,62,78,91,92],modified_graph:62,modul:[31,32,33,34,45,49,52,56,57,58,59,61,64,65,66,67,69,70,71,72,73,76,77,78,84,88,91,92,93,95,96,97],modular:63,module_fallback:55,module_nam:52,molesti:80,momentum:68,morbi:80,more:[53,54,63,64,66,67,72,75,78,85,86,89,90,92,93,94,96],most:[59,66,69,89,92,94],mother:77,motion:77,mous:77,move:[30,44,55,58,63,73,93],ms:65,msg:[26,42,70],msvc:65,msvc_x64_x64:65,mu:77,much:[61,75,77,93],mul:[54,56,68],mul_:68,multi:52,multipl:[58,77,78,89,93],multipli:[49,73],must:[33,48,49,52,55,56,61,62,63,66,69,72,73,77,78,91,92,94],mutil:78,my:77,my_custom_pass:62,my_pytorch_model:92,myclass:77,mymodel:[56,64,91,95],mymodul:91,myself:78,n:[52,61,62,63,93],nabla:77,nam:[78,80],name:[3,4,31,33,34,44,54,56,58,61,63,65,66,69,70,71,72,73,77,78,89,90,92,96],namedtupl:92,namespac:[42,43,44,45,51,55,67,93],narrow:68,nativ:[59,60,63],native_funct:60,natur:77,nav:[75,81],navig:75,navigation_depth:75,nbbind:[3,4,44],nchw:[2,72,73],ne:[55,68],nec:80,necessari:[42,62,94],need:[0,1,2,25,29,43,46,53,54,55,61,63,64,66,69,77,88,89,91,92,93,94,95],neg:68,negative_slop:68,nequ:[78,80],nest:[45,49,50,77,78],net:[61,63,77,78],netu:80,network:[29,30,54,61,63,88,89,92,93,97],neural:97,new_batch_size_input:85,new_batch_size_output:85,new_input:[85,86],new_lay:61,new_local_repositori:66,new_output:[85,86],new_siz:93,newer:66,next:[3,4,53,58,69,75,77,78,84,89,91,93],ngc:[66,89],nhwc:[2,52,72],nibh:[78,80],nice:66,nickel:77,night:78,nightli:92,ninja:[65,66],nisi:80,nisl:80,nlp:[29,30,93],nn:[55,60,63,64,69,72,73,84,90,91,92],nn_ops_convert:54,node:[54,55,56,57,59,61,62,63,69,88,91,92,95],node_info:[61,63],noexcept:[44,93],noexceptoverrid:[3,4],non:[78,80],non_block:68,none:[54,61,68,69,70,71,72,73,75,77,84,92],nonetheless:77,nonexist:77,norm:68,normal:[0,1,2,46,54,63,77,89,90,92,93,97],normalized_shap:68,noskipw:44,notat:72,notatemoduleforfallback:55,note:[1,46,48,54,61,62,63,65,66,72,75,77,91,92,95,97],notebook:[59,67,84,85,86,87],now:[55,56,59,61,63,66,77,92,96],np:89,nu:77,nulla:80,nullptr:[44,45,49],num:52,num_avg_timing_it:[45,49,73,96],num_it:52,num_op:52,num_us:91,num_work:93,number:[3,4,49,52,55,56,61,63,64,69,72,73,75,83,85,86,87,88,92],numel:68,numer:[52,78,92],numpi:89,nunc:80,nvcr:89,nvidia:[32,34,42,43,44,45,52,60,63,66,72,73,84,85,86,89,92,97],nvinfer1:[3,4,29,30,44,45,49,61,93],nvinfer:[20,44],o:[66,77,89],obj:68,object:[0,1,2,3,4,46,48,49,52,54,58,61,62,70,71,72,73,91,93,95,96],obtain:88,obvious:90,occasion:[84,85,86],odio:[78,80],off:[56,58],offer:62,offici:66,ofstream:[44,63],often:77,oh:78,ok:[63,77],okai:49,older:59,onc:[42,43,44,45,53,55,56,58,89,92,93,94],one:[47,54,55,61,63,64,70,72,77,84,85,86,89,90,92],ones:[42,54,56,57,59,63,66,77],onli:[1,3,4,16,29,44,46,48,52,55,56,59,61,66,69,70,72,77,91,92,93,94,97],onnx:55,onto:[52,58],op:[52,53,54,55,56,57,59,61,62,63,72,84,91,94],op_and_target:92,op_evalu:54,op_nam:52,opcod:54,open:[65,88,89],oper:[0,1,2,3,4,31,44,45,46,49,52,53,55,56,57,58,59,61,62,64,67,72,73,85,86,91,92,93,97],opoverload:54,opoverloadpacket:54,oppos:73,opset:[57,59],opt:[48,66,72,91],opt_c:52,opt_h:52,opt_n:52,opt_shap:[45,48,64,72,73,88,91],opt_w:52,optim:[48,52,55,63,64,67,69,85,86,88,90,91,92],optimin:48,optimiz:90,optimization_level:84,optimization_profile_field:72,optimize_target_shap:92,optimized_input_shap:69,optimized_model:[84,85,86],optimized_model_custom:84,optimz:89,option:[44,48,52,54,56,57,59,62,65,66,71,72,73,77,81,84,91,92,93,94,97],orchestra:77,orci:80,order:[33,49,56,61,62,63,64,66,69,73,92],org:[60,63,66,75,77,90,93],organ:78,origin:[33,54,69,92],ornar:[78,80],os:45,ostream:45,other:[0,1,2,45,46,52,53,54,55,58,62,63,64,66,67,68,76,77,92,94],otherwis:[66,69,92,94],our:[56,59,63,89,90,91],out:[31,44,53,55,56,57,59,61,63,65,66,70,73,77,89,91],out_shap:63,out_tensor:[61,63],output0:55,output:[24,27,33,49,52,53,54,55,56,58,61,62,63,66,70,73,75,77,78,88,89,91,92,95],output__0:89,output_binding_nam:[33,45,73],output_file_path:[52,97],output_nam:[69,92],output_pad:68,output_s:68,outself:63,outsid:77,over:[54,57,59,77,89,92],overal:[54,88,91],overrid:[29,30,44,72,92,93],overview:[60,67,84],own:[56,61,63,66,77,89],p:[52,63,68,89,97],packag:[52,55,63],pad:[68,91],padding_idx:68,page:[67,79,81,89],pair:[55,61,66,77,88,93],pane:77,paragraph:[78,81],param:[71,76],paramet:[0,1,2,3,4,25,26,27,29,30,31,32,33,34,36,46,48,49,53,54,55,61,63,69,70,72,73,81,90,92],paramt:54,parent:[14,15,18,19,20,21],pars:[63,77],parser:77,part:[52,56,59,75,76,77,91,92],partial:[52,77],particular:91,partit:55,partitioninfo:56,pass:[33,53,54,56,57,58,59,61,63,66,67,70,71,73,90,92,93],pass_manag:62,passlist:62,passmanag:62,past:77,patch:91,path:[4,13,14,15,29,30,52,54,63,65,66,71,72,89,90,92,93],path_to_torchtrt_root:66,pathwai:90,pattern:[61,63,72],payment:76,pbtxt:89,peephole_optimz:55,pellentesqu:80,peopl:77,pep:77,perforamnc:92,perform:[29,30,62,88,89,91,93],permit:77,permut:[68,92],persist:77,pharetra:80,phase:[16,61,63,91],phasellu:80,phi:77,philosoph:77,phrase:77,pi:77,pick:90,pickler:58,piec:88,pil:89,pin:76,pin_memori:68,pip3:66,pip:[66,89],pipelin:[52,97],piplein:63,pixel_shuffl:68,pl:76,place:[48,54,55,62,66,77,78,79,92,93],placehold:[62,91],placerat:80,plan:[52,59,91],platea:80,platform:[45,52,59,65,66,89,97],pleas:[54,63,66,77,89,91,92],plugin:92,point:[63,72,75,76,77,89],pointer:[3,4,93],polish:76,pool:97,pop:58,popul:69,popular:[66,76,88],port:54,portabl:[58,73],portion:[56,77],porttitor:[78,80],posit:[52,72,75,92],possibl:[66,77,88,89],post:[29,30,49,52,63,67,91],posuer:[78,80],potenti:[49,80],pow:68,power:[63,77,88,92],pr:63,praesent:80,pragma:[42,43,44,45,93],pre:[33,54,55,71,73,93,94],pre_cxx11_abi:66,preced:[54,77],precis:[49,52,63,64,67,72,85,86,92,93,97],prefer:63,prefix:[27,28,42,70,77],preinstal:66,prelu:68,prepar:[89,92],preprint:93,preproc:71,preprocess:[89,93],prerequisit:65,present:[54,65],preserv:[77,90,93],prespect:90,press:[65,77],pretium:80,pretrain:[85,88,89,96],pretti:63,prev_next_buttons_loc:75,prevent:[49,52,56],previou:[65,75,84],previous:[29,33,63],prim:[53,55,56,58,63,68,90],prim_devic:68,primal:77,primari:95,primarili:[59,63],print:[16,31,44,62,63,70,72,73,77,85,86,89,96],prior:91,priorit:66,prioriti:54,privat:[3,4,44,45,93],problem:77,problemat:77,proce:89,proceed:89,process:[52,56,76,77,84,88,89,90,93,96],prod:68,produc:[48,53,54,58,61,63,72,77,88],product:49,profil:[48,69],profiling_verbos:92,program:[18,19,20,21,29,51,52,57,58,59,67,90,95],programm:77,progress:78,proin:80,project:[66,76,81],projectdir:65,promis:92,prop:69,properli:66,properti:75,propog:55,prose:77,provid:[3,4,49,52,56,58,61,62,63,64,66,69,72,73,77,83,84,87,89,91,92,93,94,96],providi:[57,59],provok:77,pt:[52,63,89,92],ptq:[3,4,15,18,19,38,50,51,52,67,72,73],ptq_calibr:[3,4,45,49,93],ptqtemplat:50,publish:89,pull:[66,89],purchas:76,pure:31,purpos:[66,88,89,92],puru:80,push:58,push_back:[44,56],put:[77,88],put_binding_nam:73,pwd:[65,89],py3:89,py:[54,55,59,62,63,66,75,77,82,84,85,86,90,91,92,93],pyindex:[66,89],pypi:66,python3:[55,63,66],python:[52,54,56,59,62,63,69,72,73,77,78,84,85,86,87,88,89,92,94,96,97],python_api:60,pytorch:[33,48,49,52,55,56,57,58,59,61,63,64,66,71,72,73,83,87,89,90,91,93,94,95],pytorch_libtorch:89,pytorch_sphinx_them:[75,82],qat:88,qualnam:[70,71],quant_max:68,quant_min:68,quantiz:[29,30,52,63,67],quantizatiom:49,quartznet:88,question:63,qui:[78,80],quickli:[52,63,93],quisqu:80,quit:[61,63,88],quot:78,r:77,rais:[55,92],raiseexcept:55,ram:[49,52,73],rand:[63,84,92],randint:[86,91],randn:[56,63,72,73,85,91,95,96],rang:[48,49,52,54,72,88,91,92],rank:75,rather:55,raw:75,re:[77,92],read:[3,4,29,30,44,75,77,93],read_calibration_cach:71,readcalibrationcach:[3,4,44],reader:77,realiz:58,realli:61,reason:[0,90,92],reattribut:78,recalibr:29,receiv:92,recip:93,reciproc:68,recognit:[88,93],recomend:[29,30],recommend:[29,30,63,66,77,89,92],recompil:[62,85,86,91],record:[53,90],recurs:53,redistribut:78,reduc:[54,55,56,57,59,88,92,93],redund:92,ref:77,refer:[48,57,59,63,76,81,89,91,92,93],referenc:66,refit:[45,49,73,96],reflect:45,reflection_pad1d:68,reflection_pad2d:68,regard:[66,77],regardless:[78,85,86],region:92,regist:[33,54,58,61,73,92],register_acc_op:92,register_acc_op_map:92,register_custom_acc_mapper_fn:92,register_decomposit:54,registeri:54,registernodeconversionpattern:[61,63],registr:[54,62,92],registri:[53,54,63],reinterpret_cast:44,rel:[52,54,56],relat:[46,77,84,85,86],relationship:50,releas:[65,77,83,87,91,95],reload_model_output:92,reload_trt_mod:92,relu:[56,63,68,84,90],relu_:68,relwithdebinfo:65,remain:[55,93],rememb:92,remov:[62,75],remove_contigu:55,remove_dropout:55,remove_to:55,render:75,rent:78,reorder:56,repack:58,repair:62,repair_input_as_output:62,repeat:[52,68],repeat_interleav:68,replac:[54,56,62,66],replace_input_with:62,replication_pad1d:68,replication_pad2d:68,replication_pad3d:68,repo:65,report:[23,44],reportable_log_level:70,repositori:[59,75,82,89],repres:[48,49,61,70,77,92],represent:[55,61,88,90,92],request:[63,72,89],requir:[29,49,52,53,55,63,70,72,73,75,89,91,92,93,94],require_full_compil:[45,49,73],requires_grad:68,research:92,reserv:[42,43,44,45],reset:[44,84,85,86],reshap:[68,89],resiz:89,resnet18:85,resnet50:89,resnet:[58,83,87,88,89],resnet_trt:58,resolut:88,resolv:[53,55,57,59,84,85,86],resourc:[53,93],respect:54,respons:[29,58,77],rest:[77,78,92],restrict:[49,73],restructuredtext:[77,78],result:[53,54,55,56,64,70,73,75,89,90,91,95],ret:55,reus:[55,92,93],revert:75,revis:[77,78],revisit:77,rewrit:[56,62],rfc:77,rho_:77,rhoncu:80,right:[42,43,44,45,55,59,61,65,77],risu:80,rm:89,rn50_preprocess:89,robust:91,role:77,roll:68,roman:78,room:77,root:[42,43,44,45,66,75,93],roughli:56,round:[49,73],rounding_mod:68,row:78,rst:[75,77],rsub:68,rtol:52,rule:[66,73,92],ruler:77,run:[1,34,46,49,52,53,55,56,57,58,59,61,63,64,66,67,69,72,73,77,84,85,86,88,89,90,91,92,93,94,95,96,97],running_mean:68,running_var:68,runtim:[63,67,84,85,86],runtimeerror:92,rutrum:[78,80],s:[48,49,54,56,58,61,63,65,66,67,69,72,75,77,78,88,89,90,92,93],safe:[61,73],safe_dla:72,safe_gpu:72,safeti:[49,52,72],sage:77,sagitti:[78,80],sai:[78,88],said:77,same:[54,56,58,62,63,66,75,77,85,86,89,90,91,92,95,96],sampl:[64,77,84,85,86,89,92,93],sample_input:[62,84,92],sample_inputs_half:84,sapien:80,satisfi:[56,62,92],save:[29,44,52,57,58,59,63,64,67,72,73,88,89,92,94],save_timing_cach:[69,92],saving_model:67,saw:63,scalar:[54,61,68],scalaropt_dim:68,scalartyp:[0,45,68],scale:[68,88,93],scale_factor:68,scale_grad_by_freq:[54,68],scales_d:68,scales_h:68,scales_w:68,scatter:68,scelerisqu:80,scenario:62,schedul:[72,89],schema:[54,61,63],scheme:92,scientist:77,scope:[55,84,85,86],scratch:29,scratch_spac:89,screen:75,script:[31,55,56,63,64,72,73,84,85,86,90,96],script_model:[90,96],scriptclass:73,scripted_model:97,scriptmodul:[63,64,72,73,95],scroll:[75,79],sdk:60,se:88,seamlessli:67,search:[67,75],second:[55,64,77,84,85,86,92],secondli:89,section:[54,63,75,77,78,79,81,89,91,92,93],sed:[78,80],see:[31,54,55,56,58,62,63,64,66,72,73,77,84,90,92],seen:[77,78],segment:[56,85,86,88],segmentmodelwithdependencyawar:56,select:[17,29,30,34,49,52,54,58,64,65,66,68,72,73,76,79,92,93],self:[55,58,61,63,64,68,71,84,88,90,91,97],self_1:[58,63],self_int:68,sell:78,seller:76,seller_id:76,sem:80,semant:77,semper:80,send:89,senectu:80,sens:[63,77],sentenc:[77,88],sentinel:[0,2],separ:[56,57,59],seper:54,sequenc:[54,62,69,72,73,77,88,91,92],seri:56,serial:[33,34,52,57,59,63,72,73,95],seriali:73,serializ:[58,90],serialized_cach:[69,92],serialized_engin:73,seril:58,serv:[52,58,67,92],servic:77,session:77,session_nam:77,set:[3,4,16,21,25,27,29,32,34,36,45,46,48,49,52,53,55,56,57,58,59,63,64,65,66,67,69,70,72,73,75,79,82,88,90,91,92,93,97],set_data_from_numpi:89,set_devic:[38,45,50,72],set_is_colored_output_on:[39,42,50,70],set_logging_prefix:[39,42,50,70],set_reportable_log_level:[39,42,50,70],setalpha:61,setbeta:61,setnam:[61,63],setreshapedimens:63,setup:[43,89,93],sever:[16,26,70],sh:66,sha256:66,shape:[45,47,48,49,52,56,61,64,68,69,72,73,89,92,97],shape_mod:72,shape_rang:[69,92],share:[49,52,65,66,73],shell_command:77,shift:[65,66,68,77],ship:[63,94],shorthand:77,should:[0,3,4,29,45,49,52,53,54,55,56,57,59,61,67,70,72,73,75,77,80,89,92,93],show:[75,77,88],shown:[63,75,77],shuffl:[63,93],side:[55,63,75],sidebar:[75,81],sigmoid:[68,92],sigmoid_:68,sign:89,signatur:73,signifi:[48,55],signific:77,significantli:[55,75],similar:[54,61,63,92,94,96],similarli:54,simonyan:93,simpil:93,simpl:[77,78,88,89,90,92],simplest:89,simpli:[55,84,88],simplifi:[53,54],simul:88,sin:[68,77],sinc:[54,55,63,77,90,92,93],sing:77,singl:[48,52,55,56,63,72,77,90,92,93],singular:61,sinh:68,sink:77,sit:[78,80],site:[55,63,66,77],six:77,sixth:78,size:[3,4,44,48,49,52,55,56,63,68,69,72,73,75,85,86,88,91,92,93],size_t:[3,4,44,93],skip:52,slash:75,slice:[54,68],slightli:[92,95],sm:58,small:[55,89],smaller:[88,91],so:[0,44,52,53,54,55,58,59,61,62,63,66,67,69,76,77,78,84,85,86,91,92,93],sodal:80,softmax:[54,55,68,92],softwar:[49,52,73,77],sole:[64,93],sollicitudin:80,solv:89,some:[53,54,55,56,57,58,59,61,62,63,76,77,91,92,93],some_funct:77,someth:[43,55,77,89],sometim:91,someurl:77,soon:54,sort:[61,68,96],sourc:[42,43,44,45,59,65,69,70,71,72,73,84,85,86,87,92],source_ir:54,sourceforg:[77,78],sourceir:54,space:[77,78,93],spaces_and_linebreak:77,span:78,spars:[52,54,68],sparse_weight:[45,49,73,92],sparsiti:[49,52,73,92],spec:[45,48,49,52,70,72,73,96],special:[54,56],specif:[32,49,55,57,59,62,72,73,77,87,88],specifi:[3,4,33,52,54,61,64,66,67,70,72,73,75,77,89,91,92,96],specifii:72,speech:88,speedup:88,sphinx:[75,76,77,78,82,84,85,86,87],sphinx_rtd_them:[77,78],spin:89,spirit:77,split:[56,68,92],split_siz:68,split_with_s:68,sqrt:[54,68],squar:68,squeez:[68,88],sram:52,src:[58,60,68],ss:44,ssd300_trt:58,ssd:58,ssd_trace:52,ssd_trt:52,sstream:[20,44],stabl:[60,73,75],stack:[58,68,93],stage:[53,92],stai:95,stand:[58,77],standalon:77,standard:[52,58,67,77,88,94,96],stapl:78,start:[53,56,63,66,68,70,71,78,88,92,95,96],start_dim:[63,68],start_step:68,state:[53,61,62,63],state_dict:95,statement:[55,77],static_cast:44,statu:[44,78],std:[3,4,22,26,28,29,30,31,33,34,35,42,44,45,47,48,49,56,63,89,93,97],stdout:[37,70,72],steamlin:93,step:[56,67,68,88,91,92,93],stich:95,stick:75,sticki:[75,81],sticky_navig:[75,79],still:[44,56,84,92,93],stitch:[56,63],stop:63,storag:93,store:[2,4,49,52,53,58,61,63,73,90,91,92],str:[19,43,44,50,54,68,70,72,73,92],straight:61,strang:77,strategi:[56,72],street:78,strict:94,strict_type_constraint:92,stride:68,string:[3,4,18,20,21,22,26,28,29,30,31,33,34,35,42,44,45,49,54,56,58,61,63,65,72,75,93],stringstream:44,strip_prefix:66,strong:77,strongli:77,struct:[1,21,38,41,45,93],structur:[29,46,49,54,56,59,61,75,77,81,89,90],structuredtext:77,stub:78,stuff:77,style:[42,43,44,45,75,77,78],style_external_link:75,sub:[68,77,84,90],sub_:68,subdirectori:51,subexpress:55,subgraph:[49,52,53,55,61,62,63],subject:[59,62],submenu:81,submodul:[69,90,91,95],suboper:54,suboptim:56,subscript:77,subsect:77,subset:[88,93],substitut:77,subtitl:77,subtre:82,subword:88,success:65,sudo:66,suffic:55,suggest:89,suit:67,suitabl:92,sum:[49,68,73,92],superscript:77,supervis:88,suppli:77,support:[0,1,2,27,31,46,48,49,52,54,56,60,63,64,65,66,67,69,72,73,75,76,85,86,89,90,91,92,97],sure:[63,64,66,89,97],suscipit:[78,80],suspendiss:80,swap:91,sym_siz:91,symbol:[33,66,73,77,92,94],symbolic_trac:[54,91],symlink:82,system:[53,61,62,66,67,73],t1:68,t2:68,t:[0,1,2,45,46,55,61,63,66,68,72,75,77,78,89,90,91,92,93],t_:77,tabl:[66,81],tag:[77,89],take:[31,32,33,34,53,54,57,58,59,61,62,63,69,72,73,75,77,84,88,91,92,93,96],taken:77,talk:67,tan:68,tanh:68,tanh_:68,tar:[66,77,93],tarbal:[63,93],target:[1,33,45,46,48,49,52,54,56,58,59,64,65,67,72,73,91,92,93,96,97],targets_:93,task:[29,30,88,92,93],techinqu:63,techniqu:93,tell:[55,56,57,58,59,61,77],tellu:80,tem:52,templat:[20,40,44,45,50,63,75],temporari:92,tempu:80,tensor:[2,33,44,45,48,49,52,53,54,55,56,58,61,62,63,64,68,69,72,73,84,88,90,91,92,93],tensor_domain:[45,48,72],tensor_mod:68,tensor_scalar:68,tensor_tensor:68,tensorcontain:61,tensorformat:[21,38,45,48,50,72],tensorformatenum:50,tensorlist:[56,61],tensorrt:[0,1,3,4,29,30,31,32,33,34,37,44,45,46,48,49,52,53,54,55,56,57,59,61,62,69,71,72,73,83,84,90,93],tensorrt_bind:72,tensorrt_convert:[54,92],tensorrt_root:65,tensorrtcompilespec:[73,96],tensort:92,teo:52,term:[65,72,77,78,88,93],termin:[27,52,63],test:[52,56,59,65,66,77,78,88,89,92,93],test_acc_trac:92,test_decomposit:54,test_ptq_dataloader_calibr:93,test_ptq_trt_calibr:93,test_py_modul:[77,81],test_segment:56,testing_dataload:93,testing_dataset:93,text:[70,78,80,88],tf32:[49,52],than:[55,67,76,77,88,94],thats:[53,93],the_model_repositori:89,thei:[46,52,53,54,55,58,61,64,66,72,75,77,92],them:[54,55,56,58,63,66,75,88,91,92],theori:[53,77],therebi:[58,88],therefor:[29,58,63,77,88,92],theres:94,therfor:94,theta:77,thi:[0,1,2,29,30,42,43,44,45,46,47,48,49,52,53,54,55,56,57,58,59,61,62,63,65,66,69,72,73,75,76,77,79,80,83,84,85,86,87,88,89,90,91,92,93,94,95,96],thicker:77,thin:77,thing1:77,thing2:77,thing3:77,thing:[66,77,92],think:[61,77],third:[78,92],third_parti:[59,66],this_arg_is_opt:92,those:[53,54,62,77],though:[52,59,61,63,90],thought:77,three:[48,57,59,69,72,77,78,88,89,92],threshold:52,through:[48,53,54,55,56,58,63,64,67,70,71,77,88,92],throught:92,thrown:[49,73],thu:77,tile_to_repeat:55,time:[49,52,53,55,56,57,58,59,61,63,69,73,75,77,84,85,86,92,93],timing_cach:92,timing_cache_prefix:[69,92],tincidunt:80,tini:93,titles_onli:75,tmp:63,toctre:75,tocustomclass:61,todim:63,todo:[75,92],togeth:[53,61,63,95],token:[88,91],toler:52,too:[66,75,77,78],tool:[61,63,65,88,92],toolchain:[59,66],top:[59,75,79],topk:68,torch:[0,1,2,4,20,21,29,30,31,32,33,34,37,44,45,46,47,48,49,52,53,54,55,56,57,58,59,61,62,66,69,72,73,90,93,97],torch_compil:[84,85,86],torch_compile_advanced_usag:84,torch_compile_resnet_exampl:85,torch_compile_transformers_exampl:86,torch_dir:65,torch_disabled_decomposit:54,torch_enabled_decomposit:54,torch_executed_modul:[45,49,56,73],torch_executed_op:[45,49,56,73,84,85,86],torch_input_1:91,torch_input_2:91,torch_scirpt_modul:90,torch_script_modul:63,torch_tensorrt:[0,1,2,3,4,14,16,17,42,43,44,46,47,48,49,50,51,52,54,56,62,63,64,67,83,84,87,88,89,91,92,93,94,95,96,97],torch_tensorrt_build:65,torch_tensorrt_export:43,torch_tensorrt_major_vers:[19,43,50],torch_tensorrt_minor_vers:[19,43,50],torch_tensorrt_patch_vers:[19,43,50],torch_tensorrt_vers:[19,43,50],torch_tensorrtfil:50,torch_tensorrtnamespac:50,torchbind:58,torchdynamo:91,torchhub:89,torchscript:[19,21,38,43,45,49,50,52,56,57,58,59,64,69,72,73,88,91,95,96,97],torchscriptstruct:50,torchtrt:[43,56],torchtrt_api:[0,2,19,22,23,24,25,26,27,28,31,32,33,34,35,36,37,42,43,44,45,48,49,50],torchtrt_check:61,torchtrt_hidden:[19,43,50],torchtrt_runtime_exampl:94,torchtrt_unus:61,torchtrtc:[66,67,97],torchvis:[58,85,89,93,96],toronto:93,tortor:80,total:[84,85,86],totensor:[89,93],tovec:63,toward:93,trace:[54,56,63,73,90,91,92,95],trace_input:91,traced_model:90,track:[61,93],tradit:[48,73,93],traget:32,trail:75,train:[29,30,49,52,63,64,67,68],trainabl:55,transcrib:88,transfer:76,transform:[63,83,87,89,91,93,95],transformed_img:89,transformers_trac:91,translat:63,transmit:77,transpos:[68,92],trash:77,travers:[56,57,59],treat:52,tree:[42,43,44,45,75,93,94],trigger:[56,63,92],trim:93,tristiqu:80,triton:67,triton_to_np_dtyp:89,tritoncli:89,tritonserv:89,trt:[0,1,3,4,46,48,53,55,58,61,62,63,68,85,86,92],trt_exp_program:95,trt_gm:[91,95],trt_interpreter_result:92,trt_lenet_script:63,trt_mod:[56,63,91,93,97],trt_model:[56,89,95,96],trt_script_model:95,trt_t:95,trt_ts_modul:[56,64],trtinterpret:[69,92],trtinterpreterresult:[69,92],trtmodul:[69,92],trtnetwork:54,trttensor:54,truncat:[49,52,73],truncate_long_and_doubl:[45,49,73],ts:[43,52,56,63,64,67,72,90,91,95,96],ts_model:[56,63],tt:77,tue:78,tup:68,tupl:[54,58,64,69,72,73,92],tupleconstruct:[55,58],tupleindex:68,tupleunpack:55,turn:[69,92],turpi:80,tutori:[90,93],two:[52,54,55,61,62,64,66,77,78,82,89,90,91,92,93,95],type:[0,1,2,30,49,50,52,53,54,56,58,61,62,63,64,65,69,70,71,72,73,77,88,92,93],type_fp32:89,typenam:[3,4,29,30,44],typic:[53,61,89],ugli:77,ui:76,uint64_t:[45,49],ultric:80,un:93,unabl:[61,63],unari:54,unbind:68,unbroken:77,uncas:[86,88,91],uncom:66,undefin:91,under:[42,43,44,45,54,59,77,92],underli:[0,1,2,46,61],understand:91,unexpected_op:54,uniformli:88,union:[54,61,63,72,73],uniqu:[4,64],unique_ptr:[4,30],unit:92,unittest:91,univers:77,unknown:72,unless:92,unlik:[66,67,96],unlimit:75,unpack_addmm:55,unpack_log_softmax:55,unqiue_ptr:4,unreferenc:77,unrestrict:77,unsqueez:68,unstabl:59,unsupport:[31,49,65],unsur:61,untest:59,until:[53,56,59,61,66,91],unwrap:61,unwraptodoubl:61,unwraptoint:63,unzip:66,up:[53,55,56,57,58,59,62,77,84,85,86,88,90,92],updat:[65,69,92],upload:89,upon:[75,84,85,86],upper:78,upsample_bilinear2d:68,upsample_linear1d:68,upsample_nearest1d:68,upsample_nearest2d:68,upsample_nearest3d:68,upsample_trilinear3d:68,upscale_factor:68,upstream:63,uri:77,url:[66,75,89],urna:80,us:[0,1,2,3,4,29,30,32,34,36,43,44,45,46,48,49,52,53,54,56,58,59,61,62,65,67,69,70,71,72,73,75,76,77,78,83,87,89,90,91,92,93,94,95,97],usag:[63,71,77,83,87,91,92],use_cach:[3,4,30,44,71,93],use_cache_:44,use_cmake_generated_export_head:43,use_experimental_fx_rt:69,use_input_stat:68,use_python_runtim:84,use_subset:93,usecas:[64,66,87],user:[42,48,54,56,57,58,59,62,63,64,66,77,78,87,89,91,93],using_int:[63,68],usr:66,usual:[75,92],ut:80,utf:[77,78],util:[61,62,63,73,84,85,86,88,89,91,93],v0:[74,89],v143:65,v2:[29,30,77],v:[52,78,89],valid:[1,46,54,56,61,62,72],valu:[0,1,2,16,17,45,46,48,53,56,58,61,63,65,68,70,71,72,75,84,85,86,88,91],value_tensor_map:[53,61],vanilla:92,vari:[69,91,95],variabl:[48,65,72,92],variant:94,varient:55,varieti:89,variou:[92,97],variu:80,vcs_pageview_mod:75,vec:68,vector:[20,21,33,44,45,47,48,49,56,58,63,93,97],vehicula:80,vel:80,velit:80,venenati:80,verbios:52,verbos:[52,69,78,85,86,92],verbose_log:[69,92],veri:[78,79,89,92,93,96],verifi:56,verison:66,version:[35,37,59,62,65,66,75,78,88,89,92,95],vertic:[75,77],vestibulum:[78,80],vgg16:93,vgg:93,vi:77,via:[54,64,67,72,73,75,81,84,85,86,88,91,92,93,94],view:[68,75,91],virtual:93,vision:[89,92],visitor:75,vita:[78,80],vivamu:80,viverra:80,vm:78,volutpat:80,vs:[0,1,2,46,55,73,96],vscode:65,vulput:80,w:52,w_hh:68,w_ih:68,wa:[54,55,58,62,63,77,92],wai:[52,63,66,83,87,88,90,92,93,95],walkthrough:88,want:[42,54,56,63,69,84,89,90,92,93,96],warn:[16,44,52,61,70],wash:77,we:[42,44,53,54,55,56,57,58,59,61,62,63,69,75,77,83,84,85,86,87,88,89,90,91,92,93,95],weak:77,web:77,websit:66,weight:[48,49,52,53,63,68,73,77,88,92],welcom:[63,92],well:[63,66,70,77,93,95],were:63,wget:89,what:[4,55,63,64,77,90,92],whatev:[58,92],wheel:66,when:[27,44,45,46,52,53,54,55,56,57,58,59,61,63,66,70,72,73,75,77,79,88,90,91,92,93],where:[53,54,55,61,62,63,73,78,91,92,93],wherea:54,wherev:92,whether:[4,52,54,69,72,76,85,86,92,93],which:[1,2,29,32,34,46,49,53,54,55,56,57,58,59,61,62,63,64,66,69,71,72,73,75,77,78,84,88,89,90,91,92,93,94,95,96],white:77,whitespac:77,whl:66,who:77,whole:92,whose:[55,92],why:77,wide:81,width:[77,88],win:65,window:77,window_nam:77,wish:78,within:[49,52,57,59,73,75,77,95],without:[56,61,63,75,77,93],wl:94,wooden:77,word:[77,88],work:[44,55,59,61,77,78,84,91,92,93],worker:93,workflow:[69,85,86,88,91,92,96],workspac:[49,52,66,69,73,84,85,86,92],workspace_s:[45,49,52,73,85,86],workspacefold:65,world:77,would:[52,54,61,63,64,66,89,91,92,94,96],wp:89,wrap:[57,58,59,63,77,80,84,85,86,92,96],wrapper:[61,92],write:[3,4,29,30,44,53,54,63,67,77,89,92,93],write_calibration_cach:71,writecalibrationcach:[3,4,44],wrote:77,www:[63,66,75,77,89,93],x64:65,x86:94,x86_64:[59,66],x9:55,x:[5,10,33,43,55,56,63,65,66,73,78,84,90,91,95],x_0:77,x_1:77,x_2:77,x_3:77,x_4:77,x_:77,x_lgamma:56,x_out:84,x_y_out:84,xavier:[45,97],xstr:[19,43,50],xx:89,xxx:92,y:[33,56,73,78,84],y_lgamma:56,y_out:84,yahoo:78,yaml:60,yet:[88,92],you:[0,1,2,29,30,46,48,49,52,53,54,55,56,58,59,61,63,64,66,67,72,73,75,77,78,79,83,87,88,89,90,91,92,93,94,95,96],your:[61,63,64,66,67,75,77,78,82,90,91,94,96],yourself:63,yy:89,z:78,zero:91,zero_point:68,zip:[58,65,66,87],zisserman:93},titles:["Class DataType","Class Device::DeviceType","Class TensorFormat","Template Class Int8CacheCalibrator","Template Class Int8Calibrator","Define STR","Define TORCH_TENSORRT_PATCH_VERSION","Define TORCH_TENSORRT_MAJOR_VERSION","Define TORCH_TENSORRT_MINOR_VERSION","Define TORCHTRT_API","Define XSTR","Define TORCHTRT_HIDDEN","Define TORCH_TENSORRT_VERSION","Directory cpp","Directory include","Directory torch_tensorrt","Enum Level","Enum EngineCapability","File logging.h","File macros.h","File ptq.h","File torch_tensorrt.h","Function torch_tensorrt::logging::get_logging_prefix","Function torch_tensorrt::logging::get_reportable_log_level","Function torch_tensorrt::logging::get_is_colored_output_on","Function torch_tensorrt::logging::set_reportable_log_level","Function torch_tensorrt::logging::log","Function torch_tensorrt::logging::set_is_colored_output_on","Function torch_tensorrt::logging::set_logging_prefix","Template Function torch_tensorrt::ptq::make_int8_cache_calibrator","Template Function torch_tensorrt::ptq::make_int8_calibrator","Function torch_tensorrt::torchscript::check_method_operator_support","Function torch_tensorrt::torchscript::compile","Function torch_tensorrt::torchscript::embed_engine_in_new_module","Function torch_tensorrt::torchscript::convert_method_to_trt_engine","Function torch_tensorrt::get_build_info","Function torch_tensorrt::set_device","Function torch_tensorrt::dump_build_info","Namespace torch_tensorrt","Namespace torch_tensorrt::logging","Namespace torch_tensorrt::ptq","Namespace torch_tensorrt::torchscript","Program Listing for File logging.h","Program Listing for File macros.h","Program Listing for File ptq.h","Program Listing for File torch_tensorrt.h","Struct Device","Struct GraphInputs","Struct Input","Struct CompileSpec","Torch-TensorRT C++ API","Full API","torchtrtc","Conversion Phase","Dynamo Converters","Lowering Phase","Partitioning Phase","Compiler Phases","Runtime Phase","System Overview","Useful Links for Torch-TensorRT Development","Writing Converters","Writing Dynamo ATen Lowering Passes","Using Torch-TensorRT in C++","Using Torch-TensorRT in Python","Building Torch-TensorRT on Windows","Installation","Torch-TensorRT","Operators Supported","torch_tensorrt.fx","torch_tensorrt.logging","torch_tensorrt.ptq","torch_tensorrt","torch_tensorrt.ts","Changelog","Configuration","5. :mod:`test_py_module`","3. Paragraph Level Markup","4. Lists & Tables","1. Long Sticky Nav","1. Structural Elements","<no title>","Installation","Dynamo / torch.compile","Torch Compile Advanced Usage","Compiling ResNet using the Torch-TensorRT torch.compile Backend","Compiling a Transformer using torch.compile and TensorRT","Torch-TensorRT Tutorials","Example notebooks","Serving a Torch-TensorRT model with Triton","Creating a TorchScript Module","Dynamic shapes with Torch-TensorRT","Torch-TensorRT (FX Frontend) User Guide","Post Training Quantization (PTQ)","Deploying Torch-TensorRT Programs","Saving models compiled with Torch-TensorRT","Using Torch-TensorRT Directly From PyTorch","DLA"],titleterms:{"1":[79,89],"10":79,"11":79,"12":79,"13":79,"14":79,"15":79,"16":79,"17":79,"18":79,"19":79,"2":[79,80,89],"20":79,"3":[79,89],"4":79,"5":79,"6":79,"7":79,"8":79,"9":79,"class":[0,1,2,3,4,20,21,38,40,41,50,69,71,72],"default":84,"enum":[16,17,38,39,50,71,72],"function":[22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,50,60,69,72,73],"import":[84,85,86],"long":[79,81],"static":91,A:77,And:77,But:78,By:[18,19],Or:55,The:[63,77],To:55,With:65,aarch64:66,abi:[58,66],acc:92,acceler:88,add:92,addmm:55,admonit:77,advanc:84,advic:61,ahead:67,an:81,api:[50,51,60,66,67],applic:93,arg:[61,76],argument:[85,86],aten:62,automat:56,avail:60,awar:[56,88],backend:85,background:[58,61],base:[3,4,48,75],basic:62,bert:[88,91],binari:66,block:77,branch:55,build:[65,66,75,89],bullet:78,c:[50,60,63,66,67,88,93],can:78,caption:[78,81],center:77,ch:77,changelog:74,check_method_operator_support:31,choos:66,citat:[77,93],citrinet:88,cleanup:[84,85,86],cli:[66,67],client:89,cmake:66,code:[55,65,77],compil:[32,57,59,63,65,66,67,83,84,85,86,87,88,91,95],compilespec:49,compound:77,configur:[65,75],constraint:91,construct:58,content:[18,19,20,21,38,39,40,41,75,76,77,78,79,80],context:[61,75],contigu:55,contract:61,contributor:67,convers:[53,57,59,61],convert:[53,54,61,63,68,92],convert_method_to_trt_engin:34,cpp:[13,18,19,20,21,56],creat:[90,93],creativ:77,cuda:[84,85,86],cudnn:66,current:68,custom:[63,84,91],cxx11:66,data:76,datatyp:0,dead:55,debug:66,deep:88,deeper:78,defin:[5,6,7,8,9,10,11,12,19,50],definit:[18,19,20,21,78,84,85,86],demo:81,depend:[56,66],deploi:[88,94],deseri:58,detect:88,develop:60,devic:[1,46],devicetyp:1,dimens:60,direct:77,directli:96,directori:[13,14,15,51],disk:90,distribut:66,dla:97,doctest:77,documen:67,document:[0,1,2,3,4,5,6,7,8,9,10,11,12,16,17,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,46,47,48,49,60,67,80,81],down:78,download:[77,82],driver:[84,85,86],dropout:55,dump_build_info:37,dynam:[88,91],dynamo:[54,62,83,87],easier:60,efficentnet:88,element:80,elimin:55,eliminatecommonsubexpress:55,embed_engine_in_new_modul:33,emphas:77,engin:[58,92],enginecap:17,enumer:78,envior:66,error:[84,85,86],evalu:[53,68],exampl:[62,77,79,88,91],execept:55,executor:58,expect:60,face:88,fallback:[55,56],field:78,figur:77,file:[15,18,19,20,21,42,43,44,45,50,51],flatten:55,footnot:77,format:58,freez:55,from:[66,96],frontend:[88,92],full:[50,51],fuse:55,fx2trt:92,fx:[69,88,92],gaurd:55,gener:76,get:67,get_build_info:35,get_is_colored_output_on:24,get_logging_prefix:22,get_reportable_log_level:23,giant:78,git:82,glossari:77,gpu:67,graph:[55,58],graphinput:47,grid:78,guarante:61,guid:[67,92],h:[18,19,20,21,42,43,44,45,56],have:78,hierarchi:50,hlist:78,hole:78,hood:[63,91],how:[75,92,93],html:75,hug:88,ien:77,imag:[77,78],implement:54,includ:[14,18,19,20,21],incred:81,index:76,indic:67,infer:[85,86,89],inherit:[3,4,48],inlin:77,input:[48,85,86],instal:[65,66,82],int8:88,int8cachecalibr:3,int8calibr:4,ir:[60,91],jetson:66,jit:67,languag:88,layer:60,learn:88,lenet:88,level:[16,75,77,78],librari:[66,94],libtorchtrt:94,like:78,limit:91,line:77,linear:55,link:[60,77],list:[42,43,44,45,78],liter:77,local:66,log:[18,22,23,24,25,26,27,28,39,42,70],logsoftmax:55,loop:55,lower:[55,57,59,62],macro:[19,43],make_int8_cache_calibr:29,make_int8_calibr:30,markup:77,mask:88,math:77,menu:[79,81],meta:77,miss:92,mlm:88,mod:76,model:[84,85,86,88,89,92,95],modul:[55,63,90],namespac:[18,19,20,21,38,39,40,41,50],nativ:66,native_op:60,nav:79,nest:[1,46],node:53,note:[84,85,86],notebook:88,number:[77,78],nvidia:67,object:88,one:78,op:[58,92],oper:[54,63,68],optim:89,optimz:55,option:[75,76,78,85,86],other:61,overview:59,own:93,packag:[66,94],page:75,paragraph:[77,80],paramet:76,partit:[56,57,59],partitoninfo:56,pass:[55,62],pattern:55,peephol:55,phase:[53,55,56,57,58,59],plugin:94,post:93,pre:66,precompil:66,prerequisit:66,program:[42,43,44,45,94],project:75,ptq:[20,29,30,40,44,71,93],python:[60,64,66,67,90,93],pytorch:[60,67,88,92,96],quantiz:[88,93],queri:89,quickstart:63,quot:77,rabbit:78,read:60,redund:55,refer:77,regist:[62,63],relationship:[1,3,4,46,48],releas:66,remov:55,repeat:55,replac:[55,77],requir:62,resnet50:88,resnet:85,respons:61,result:58,right:66,rubric:77,runtim:[57,58,59,94],save:[90,95],second:78,section:80,segmentedblock:56,serial:58,serv:[88,89],server:89,set:[54,84,89],set_devic:36,set_is_colored_output_on:27,set_logging_prefix:28,set_reportable_log_level:25,setup:66,shape:[88,91],shape_analysi:56,sidebar:77,so:94,sometim:60,sourc:66,ssd:88,start:67,step:[54,89],sticki:79,str:5,struct:[46,47,48,49,50],structur:80,studio:65,subdirectori:[13,14],submenu:79,submodul:72,subsect:80,subsubmenu:79,subsubsect:80,support:68,system:59,tabl:[75,76,77,78,79,80],tarbal:66,target:77,templat:[3,4,29,30],tensorformat:2,tensorrt:[50,58,60,63,64,65,66,67,85,86,87,88,89,91,92,94,95,96],test:54,test_py_modul:76,text:77,theme:[75,81],thi:[78,81],through:68,tile:55,time:67,titl:77,toc:75,topic:77,torch:[50,60,63,64,65,67,83,84,85,86,87,88,89,91,92,94,95,96],torch_compil:91,torch_tensorrt:[15,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,45,69,70,71,72,73,85,86],torch_tensorrt_major_vers:7,torch_tensorrt_minor_vers:8,torch_tensorrt_patch_vers:6,torch_tensorrt_vers:12,torchscript:[31,32,33,34,41,63,67,90],torchtrt_api:9,torchtrt_hidden:11,torchtrtc:[52,63],tracer:92,train:[88,93],transform:[86,88],triton:89,ts:73,tupl:55,tutori:[67,87],type:[3,4,46,48],under:[63,91],unpack:55,unrol:55,unsupport:63,up:89,us:[55,60,63,64,66,84,85,86,88,96],usag:84,user:[67,92],version:58,via:82,visual:65,wai:77,weight:61,what:61,wide:75,window:65,work:[63,90],workaround:91,write:[61,62],xstr:10,your:[89,93]}}) \ No newline at end of file diff --git a/docs/src/pytorch-sphinx-theme/docs/changelog.html b/docs/src/pytorch-sphinx-theme/docs/changelog.html index 72cd661ff7..ebcef648ec 100644 --- a/docs/src/pytorch-sphinx-theme/docs/changelog.html +++ b/docs/src/pytorch-sphinx-theme/docs/changelog.html @@ -10,7 +10,7 @@ - Changelog — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Changelog — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/src/pytorch-sphinx-theme/docs/configuring.html b/docs/src/pytorch-sphinx-theme/docs/configuring.html index 6210cba57c..1b9d4a7233 100644 --- a/docs/src/pytorch-sphinx-theme/docs/configuring.html +++ b/docs/src/pytorch-sphinx-theme/docs/configuring.html @@ -10,7 +10,7 @@ - Configuration — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Configuration — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/api.html b/docs/src/pytorch-sphinx-theme/docs/demo/api.html index f447a4c645..56ecd2efec 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/api.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/api.html @@ -10,7 +10,7 @@ - 5. :mod:`test_py_module` — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + 5. :mod:`test_py_module` — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/demo.html b/docs/src/pytorch-sphinx-theme/docs/demo/demo.html index 3afd01b139..30a920c569 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/demo.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/demo.html @@ -12,7 +12,7 @@ - 3. Paragraph Level Markup — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + 3. Paragraph Level Markup — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    @@ -582,7 +585,7 @@

    3.4.4.

    3.4.5. Code Blocks

    # parsed-literal test
    -curl -O http://someurl/release-v2.0.0.dev0+1fec519.tar-gz
    +curl -O http://someurl/release-v2.2.0.dev0+50ab2c1.tar-gz

    Code Blocks can have captions.
    {
    diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/lists_tables.html b/docs/src/pytorch-sphinx-theme/docs/demo/lists_tables.html
    index 07fdd600a2..66efa473dc 100644
    --- a/docs/src/pytorch-sphinx-theme/docs/demo/lists_tables.html
    +++ b/docs/src/pytorch-sphinx-theme/docs/demo/lists_tables.html
    @@ -10,7 +10,7 @@
     
       
       
    -  4. Lists & Tables — Torch-TensorRT v2.0.0.dev0+1fec519 documentation
    +  4. Lists & Tables — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation
       
     
       
    @@ -223,7 +223,7 @@
                   
                   
                     
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/long.html b/docs/src/pytorch-sphinx-theme/docs/demo/long.html index 70fb1ed221..c08de37334 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/long.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/long.html @@ -10,7 +10,7 @@ - 1. Long Sticky Nav — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + 1. Long Sticky Nav — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/src/pytorch-sphinx-theme/docs/demo/structure.html b/docs/src/pytorch-sphinx-theme/docs/demo/structure.html index a6b54f077b..35a9727672 100644 --- a/docs/src/pytorch-sphinx-theme/docs/demo/structure.html +++ b/docs/src/pytorch-sphinx-theme/docs/demo/structure.html @@ -10,7 +10,7 @@ - 1. Structural Elements — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + 1. Structural Elements — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/src/pytorch-sphinx-theme/docs/index.html b/docs/src/pytorch-sphinx-theme/docs/index.html index c8adc5149b..ed2dcfe345 100644 --- a/docs/src/pytorch-sphinx-theme/docs/index.html +++ b/docs/src/pytorch-sphinx-theme/docs/index.html @@ -10,7 +10,7 @@ - <no title> — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + <no title> — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/src/pytorch-sphinx-theme/docs/installing.html b/docs/src/pytorch-sphinx-theme/docs/installing.html index 45f233d0e6..4977045709 100644 --- a/docs/src/pytorch-sphinx-theme/docs/installing.html +++ b/docs/src/pytorch-sphinx-theme/docs/installing.html @@ -10,7 +10,7 @@ - Installation — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Installation — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/tutorials/_rendered_examples/dynamo/index.html b/docs/tutorials/_rendered_examples/dynamo/index.html index dd334fb068..a07f008093 100644 --- a/docs/tutorials/_rendered_examples/dynamo/index.html +++ b/docs/tutorials/_rendered_examples/dynamo/index.html @@ -10,7 +10,7 @@ - Dynamo / torch.compile — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Dynamo / torch.compile — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.html b/docs/tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.html index 607bd80866..ac52d9c0b3 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_compile_advanced_usage.html @@ -10,7 +10,7 @@ - Torch Compile Advanced Usage — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Torch Compile Advanced Usage — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.html b/docs/tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.html index b1447676b6..9cc0c0b790 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_compile_resnet_example.html @@ -10,7 +10,7 @@ - Compiling ResNet using the Torch-TensorRT torch.compile Backend — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Compiling ResNet using the Torch-TensorRT torch.compile Backend — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.html b/docs/tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.html index 31e524ca68..aa6dbab619 100644 --- a/docs/tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.html +++ b/docs/tutorials/_rendered_examples/dynamo/torch_compile_transformers_example.html @@ -10,7 +10,7 @@ - Compiling a Transformer using torch.compile and TensorRT — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Compiling a Transformer using torch.compile and TensorRT — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/tutorials/_rendered_examples/index.html b/docs/tutorials/_rendered_examples/index.html index a02edb770b..eeaf2857a4 100644 --- a/docs/tutorials/_rendered_examples/index.html +++ b/docs/tutorials/_rendered_examples/index.html @@ -10,7 +10,7 @@ - Torch-TensorRT Tutorials — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Torch-TensorRT Tutorials — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -223,7 +223,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -267,6 +267,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -302,6 +304,7 @@

    Indices

    diff --git a/docs/tutorials/notebooks.html b/docs/tutorials/notebooks.html index 56a86eac08..dd264be243 100644 --- a/docs/tutorials/notebooks.html +++ b/docs/tutorials/notebooks.html @@ -10,7 +10,7 @@ - Example notebooks — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Example notebooks — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/tutorials/serving_torch_tensorrt_with_triton.html b/docs/tutorials/serving_torch_tensorrt_with_triton.html index 7dd549f255..3cd0360106 100644 --- a/docs/tutorials/serving_torch_tensorrt_with_triton.html +++ b/docs/tutorials/serving_torch_tensorrt_with_triton.html @@ -10,7 +10,7 @@ - Serving a Torch-TensorRT model with Triton — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Serving a Torch-TensorRT model with Triton — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/user_guide/creating_torchscript_module_in_python.html b/docs/user_guide/creating_torchscript_module_in_python.html index 996b36eacf..8e4a4421d8 100644 --- a/docs/user_guide/creating_torchscript_module_in_python.html +++ b/docs/user_guide/creating_torchscript_module_in_python.html @@ -10,7 +10,7 @@ - Creating a TorchScript Module — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Creating a TorchScript Module — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/user_guide/dynamic_shapes.html b/docs/user_guide/dynamic_shapes.html new file mode 100644 index 0000000000..719ccb7fac --- /dev/null +++ b/docs/user_guide/dynamic_shapes.html @@ -0,0 +1,917 @@ + + + + + + + + + + + + + Dynamic shapes with Torch-TensorRT — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    +
    +
    + + + + + + + + + + + + + + + + +
    + +
      + +
    • + + + Docs + + > +
    • + + +
    • Dynamic shapes with Torch-TensorRT
    • + + +
    • + + + + + +
    • + +
    + + +
    +
    + +
    + Shortcuts +
    +
    + +
    +
    + + + + + + +
    + +
    +
    + +
    +

    Dynamic shapes with Torch-TensorRT

    +

    By default, you can run a pytorch model with varied input shapes and the output shapes are determined eagerly. +However, Torch-TensorRT is an AOT compiler which requires some prior information about the input shapes to compile and optimize the model. +In the case of dynamic input shapes, we must provide the (min_shape, opt_shape, max_shape) arguments so that the model can be optimized for +these range of input shapes. An example usage of static and dynamic shapes is as follows.

    +

    NOTE: The following code uses dynamo IR. Incase of Torchscript IR, please swap out ir=dynamo with ir=ts and the behavior is exactly the same.

    +
    import torch
    +import torch_tensorrt
    +
    +model = MyModel().eval().cuda()
    +# Compile with static shapes
    +inputs = torch_tensorrt.Input(shape=[1, 3, 224, 224], dtype=torch.float32)
    +# or compile with dynamic shapes
    +inputs = torch_tensorrt.Input(min_shape=[1, 3, 224, 224],
    +                              opt_shape=[4, 3, 224, 224],
    +                              max_shape=[8, 3, 224, 224],
    +                              dtype=torch.float32)
    +trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs)
    +
    +
    +
    +

    Under the hood

    +

    There are two phases of compilation when we use torch_tensorrt.compile API with ir=dynamo (default).

    +
      +
    • aten_tracer.trace (which uses torch.export to trace the graph with the given inputs)

    • +
    +

    In the tracing phase, we use torch.export along with the constraints. In the case of +dynamic shaped inputs, the range can be provided to the tracing via constraints. Please +refer to this docstring +for detailed information on how to set constraints. In short, we create new inputs for +torch.export tracing and provide constraints on the min and max values(provided by the user), a particular dimension can take. +Please take a look at aten_tracer.py file to understand how this works under the hood.

    +
      +
    • dynamo.compile (which compiles a torch.fx.GraphModule object using TensorRT)

    • +
    +

    In the conversion to TensorRT, we use the user provided dynamic shape inputs. +We perform shape analysis using dummy inputs (across min, opt and max shapes) and store the +intermediate output shapes which can be used in case the graph has a mix of Pytorch +and TensorRT submodules.

    +
    +
    +

    Custom Constraints

    +

    Given an input x = torch_tensorrt.Input(min_shape, opt_shape, max_shape, dtype), +Torch-TensorRT automatically sets the constraints during torch.export tracing as follows

    +
    for dim in constraint_dims:
    +    if min_shape[dim] > 1:
    +        constraints.append(min_shape[dim] <= dynamic_dim(trace_input, dim))
    +    if max_shape[dim] > 1:
    +        constraints.append(dynamic_dim(trace_input, dim) <= max_shape[dim])
    +
    +
    +

    Sometimes, we might need to set additional constraints and Torchdynamo errors out if we don’t specify them. +For example, in the case of BERT model compilation, there are two inputs and a constraint has to be set involving the sequence length size of these two inputs.

    +
    constraints.append(dynamic_dim(trace_inputs[0], 0) == dynamic_dim(trace_inputs[1], 0))
    +
    +
    +

    If you have to provide any custom constraints to your model, the overall workflow for model compilation using ir=dynamo would involve a few steps.

    +
    import torch
    +import torch_tensorrt
    +from torch_tensorrt.dynamo.lowering import apply_lowering_passes, get_decompositions
    +# Assume the model has two inputs
    +model = MyModel()
    +torch_input_1 = torch.randn((1, 14), dtype=torch.int32).cuda()
    +torch_input_2 = torch.randn((1, 14), dtype=torch.int32).cuda()
    +
    +dynamic_inputs = [torch_tensorrt.Input(min_shape=[1, 14],
    +                    opt_shape=[4, 14],
    +                    max_shape=[8, 14],
    +                    dtype=torch.int32),
    +                  torch_tensorrt.Input(min_shape=[1, 14],
    +                    opt_shape=[4, 14],
    +                    max_shape=[8, 14],
    +                    dtype=torch.int32)]
    +
    +# Export the model with additional constraints
    +constraints = []
    +# The following constraints are automatically added by Torch-TensorRT in the
    +# general case when you call torch_tensorrt.compile directly on MyModel()
    +constraints.append(dynamic_dim(torch_input_1, 0) < 8)
    +constraints.append(dynamic_dim(torch_input_2, 0) < 8)
    +# This is an additional constraint as instructed by Torchdynamo
    +constraints.append(dynamic_dim(torch_input_1, 0) == dynamic_dim(torch_input_2, 0))
    +with unittest.mock.patch(
    +    "torch._export.DECOMP_TABLE", get_decompositions(experimental_decompositions)
    +):
    +    graph_module = export(
    +        model, (torch_input_1, torch_input_2), constraints=constraints
    +    ).module()
    +
    +# Use the dynamo.compile API
    +trt_mod = torch_tensorrt.dynamo.compile(graph_module, inputs=dynamic_inputs, **compile_spec)
    +
    +
    +
    +
    +

    Limitations

    +

    If there are operations in the graph that use the dynamic dimension of the input, Pytorch +introduces torch.ops.aten.sym_size.int ops in the graph. Currently, we cannot handle these operators and +the compilation results in undefined behavior. We plan to add support for these operators and implement +robust support for shape tensors in the next release. Here is an example of the limitation described above

    +
    import torch
    +import torch_tensorrt
    +
    +class MyModule(torch.nn.Module):
    +    def __init__(self):
    +        super().__init__()
    +        self.avgpool = torch.nn.AdaptiveAvgPool2d((1, 1))
    +
    +    def forward(self, x):
    +        x = self.avgpool(x)
    +        out = torch.flatten(x, 1)
    +        return out
    +
    +model = MyModel().eval().cuda()
    +# Compile with dynamic shapes
    +inputs = torch_tensorrt.Input(min_shape=(1, 512, 1, 1),
    +                     opt_shape=(4, 512, 1, 1),
    +                     max_shape=(8, 512, 1, 1),
    +                     dtype=torch.float32)
    +trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs)
    +
    +
    +

    The traced graph of MyModule() looks as follows

    +
    Post export graph: graph():
    +%arg0_1 : [num_users=2] = placeholder[target=arg0_1]
    +%mean : [num_users=1] = call_function[target=torch.ops.aten.mean.dim](args = (%arg0_1, [-1, -2], True), kwargs = {})
    +%sym_size : [num_users=1] = call_function[target=torch.ops.aten.sym_size.int](args = (%arg0_1, 0), kwargs = {})
    +%view : [num_users=1] = call_function[target=torch.ops.aten.view.default](args = (%mean, [%sym_size, 512]), kwargs = {})
    +return (view,)
    +
    +
    +

    Here the %sym_size node captures the dynamic batch and uses it in the aten.view layer. This requires shape tensors support +which would be a part of our next release.

    +
    +
    +

    Workaround (BERT static compilation example)

    +

    In the case where you encounter the issues mentioned in the Limitations section, +you can compile the model (static mode) with max input size that can be provided. In the cases of smaller inputs, +we can pad them accordingly. This is only a workaround until we address the limitations.

    +
    import torch
    +import torch_tensorrt
    +from transformers.utils.fx import symbolic_trace as transformers_trace
    +
    +model = BertModel.from_pretrained("bert-base-uncased").cuda().eval()
    +
    +# Input sequence length is 20.
    +input1 = torch.randint(0, 5, (1, 20), dtype=torch.int32).to("cuda")
    +input2 = torch.randint(0, 5, (1, 20), dtype=torch.int32).to("cuda")
    +
    +model = transformers_trace(model, input_names=["input_ids", "attention_mask"]).eval().cuda()
    +trt_mod = torch_tensorrt.compile(model, inputs=[input1, input2], **compile_spec)
    +model_outputs = model(input, input2)
    +
    +# If you have a sequence of length 14, pad 6 zero tokens and run inference
    +# or recompile for sequence length of 14.
    +input1 = torch.randint(0, 5, (1, 14), dtype=torch.int32).to("cuda")
    +input2 = torch.randint(0, 5, (1, 14), dtype=torch.int32).to("cuda")
    +trt_mod = torch_tensorrt.compile(model, inputs=[input1, input2], **compile_spec)
    +model_outputs = model(input, input2)
    +
    +
    +
    +
    +

    Dynamic shapes with ir=torch_compile

    +

    torch_tensorrt.compile(model, inputs, ir="torch_compile") returns a torch.compile boxed function with the backend +configured to Tensorrt. In the case of ir=torch_compile, users have to recompile for different input shapes. +In the future, we plan to explore the option of compiling with dynamic shapes in the first execution of the model.

    +
    import torch
    +import torch_tensorrt
    +
    +model = MyModel().eval().cuda()
    +inputs = torch.randn((1, 3, 224, 224), dtype=float32)
    +trt_gm = torch_tensorrt.compile(model, ir="torch_compile", inputs)
    +# Compilation happens when you call the model
    +trt_gm(inputs)
    +
    +# Recompilation happens with modified batch size
    +inputs_bs2 = torch.randn((2, 3, 224, 224), dtype=torch.float32)
    +trt_gm = torch_tensorrt.compile(model, ir="torch_compile", inputs_bs2)
    +
    +
    +
    +
    + + +
    + +
    + + +
    +
    + + +
    +
    + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    +
    +
    +
    +

    Docs

    +

    Access comprehensive developer documentation for PyTorch

    + View Docs +
    + +
    +

    Tutorials

    +

    Get in-depth tutorials for beginners and advanced developers

    + View Tutorials +
    + +
    +

    Resources

    +

    Find development resources and get your questions answered

    + View Resources +
    +
    +
    +
    + + + + + + + + + +
    +
    +
    +
    + + +
    +
    +
    + + +
    + + + + + + + + \ No newline at end of file diff --git a/docs/user_guide/getting_started_with_fx_path.html b/docs/user_guide/getting_started_with_fx_path.html index 3f24613510..a1b0c471bd 100644 --- a/docs/user_guide/getting_started_with_fx_path.html +++ b/docs/user_guide/getting_started_with_fx_path.html @@ -10,7 +10,7 @@ - Torch-TensorRT (FX Frontend) User Guide — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Torch-TensorRT (FX Frontend) User Guide — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/user_guide/ptq.html b/docs/user_guide/ptq.html index 27154ea40f..fd93b60fbc 100644 --- a/docs/user_guide/ptq.html +++ b/docs/user_guide/ptq.html @@ -10,7 +10,7 @@ - Post Training Quantization (PTQ) — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Post Training Quantization (PTQ) — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -225,7 +225,7 @@
    - v2.0.0.dev0+1fec519 + v2.2.0.dev0+50ab2c1
    @@ -269,6 +269,8 @@
  • Torch-TensorRT (FX Frontend) User Guide
  • Post Training Quantization (PTQ)
  • Deploying Torch-TensorRT Programs
  • +
  • Saving models compiled with Torch-TensorRT
  • +
  • Dynamic shapes with Torch-TensorRT
  • Using Torch-TensorRT Directly From PyTorch
  • DLA
  • @@ -304,6 +306,7 @@

    Indices

    diff --git a/docs/user_guide/runtime.html b/docs/user_guide/runtime.html index a4cdd70e22..751f067279 100644 --- a/docs/user_guide/runtime.html +++ b/docs/user_guide/runtime.html @@ -10,7 +10,7 @@ - Deploying Torch-TensorRT Programs — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Deploying Torch-TensorRT Programs — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -39,7 +39,7 @@ - + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    +
    +
    + + + + + + + + + + + + + + + + +
    + +
      + +
    • + + + Docs + + > +
    • + + +
    • Saving models compiled with Torch-TensorRT
    • + + +
    • + + + + + +
    • + +
    + + +
    +
    + +
    + Shortcuts +
    +
    + +
    +
    + + + + + + +
    + +
    +
    + +
    +

    Saving models compiled with Torch-TensorRT

    +

    Saving models compiled with Torch-TensorRT varies slightly with the ir that has been used for compilation.

    +
      +
    1. Dynamo IR

    2. +
    +

    Starting with 2.1 release of Torch-TensorRT, we are switching the default compilation to be dynamo based. +The output of ir=dynamo compilation is a torch.fx.GraphModule object. There are two ways to save these objects

    +

    a) Converting to Torchscript +torch.fx.GraphModule objects cannot be serialized directly. Hence we use torch.jit.trace to convert this into a ScriptModule object which can be saved to disk. +The following code illustrates this approach.

    +
    import torch
    +import torch_tensorrt
    +
    +model = MyModel().eval().cuda()
    +inputs = torch.randn((1, 3, 224, 224)).cuda()
    +trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) # Output is a torch.fx.GraphModule
    +trt_script_model = torch.jit.trace(trt_gm, inputs)
    +torch.jit.save(trt_script_model, "trt_model.ts")
    +
    +# Later, you can load it and run inference
    +model = torch.jit.load("trt_model.ts").cuda()
    +model(inputs)
    +
    +
    +

    b) ExportedProgram +torch.export.ExportedProgram is a new format introduced in Pytorch 2.1. After we compile a Pytorch module using Torch-TensorRT, the resultant +torch.fx.GraphModule along with additional metadata can be used to create ExportedProgram which can be saved and loaded from disk.

    +
    import torch
    +import torch_tensorrt
    +from torch_tensorrt.dynamo.export import transform, create_exported_program
    +
    +model = MyModel().eval().cuda()
    +inputs = torch.randn((1, 3, 224, 224)).cuda()
    +trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs) # Output is a torch.fx.GraphModule
    +# Transform and create an exported program
    +trt_gm = transform(trt_gm, inputs)
    +trt_exp_program = create_exported_program(trt_gm, call_spec, trt_gm.state_dict())
    +torch._export.save(trt_exp_program, "trt_model.ep")
    +
    +# Later, you can load it and run inference
    +model = torch._export.load("trt_model.ep")
    +model(inputs)
    +
    +
    +

    torch_tensorrt.dynamo.export.transform inlines the submodules within a GraphModule to their corresponding nodes and stiches all the nodes together. +This is needed as torch._export serialization cannot handle serializing and deserializing of submodules (call_module nodes).

    +

    NOTE: This way of saving the models using ExportedProgram is experimental. Here is a known issue : https://github.com/pytorch/TensorRT/issues/2341

    +
      +
    1. Torchscript IR

    2. +
    +
    +

    In Torch-TensorRT 1.X versions, the primary way to compile and run inference with Torch-TensorRT is using Torchscript IR. +This behavior stays the same in 2.X versions as well.

    +
    import torch
    +import torch_tensorrt
    +
    +model = MyModel().eval().cuda()
    +inputs = torch.randn((1, 3, 224, 224)).cuda()
    +trt_ts = torch_tensorrt.compile(model, ir="ts", inputs) # Output is a ScriptModule object
    +torch.jit.save(trt_ts, "trt_model.ts")
    +
    +# Later, you can load it and run inference
    +model = torch.jit.load("trt_model.ts").cuda()
    +model(inputs)
    +
    +
    +
    +
    + + +
    + +
    + + +
    +
    + + +
    +
    + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    +
    +
    +
    +

    Docs

    +

    Access comprehensive developer documentation for PyTorch

    + View Docs +
    + +
    +

    Tutorials

    +

    Get in-depth tutorials for beginners and advanced developers

    + View Tutorials +
    + +
    +

    Resources

    +

    Find development resources and get your questions answered

    + View Resources +
    +
    +
    +
    + + + + + + + + + +
    +
    +
    +
    + + +
    +
    +
    + + +
    + + + + + + + + \ No newline at end of file diff --git a/docs/user_guide/use_from_pytorch.html b/docs/user_guide/use_from_pytorch.html index 836ae7faf6..4a866a5c7a 100644 --- a/docs/user_guide/use_from_pytorch.html +++ b/docs/user_guide/use_from_pytorch.html @@ -10,7 +10,7 @@ - Using Torch-TensorRT Directly From PyTorch — Torch-TensorRT v2.0.0.dev0+1fec519 documentation + Using Torch-TensorRT Directly From PyTorch — Torch-TensorRT v2.2.0.dev0+50ab2c1 documentation @@ -40,7 +40,7 @@ - +