Skip to content

Commit 4e47be6

Browse files
committed
docs: minor changes in Resnet50 notebook
1 parent 5a7f00e commit 4e47be6

File tree

1 file changed

+3
-4
lines changed

1 file changed

+3
-4
lines changed

notebooks/Resnet50-example.ipynb

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -208,8 +208,7 @@
208208
"<a id=\"3\"></a>\n",
209209
"## 3. Running the model without optimizations\n",
210210
"\n",
211-
"\n",
212-
"PyTorch has a model repository called `timm`, which is a source for high quality implementations of computer vision models. We can get our EfficientNet model from there pretrained on ImageNet."
211+
"\n"
213212
]
214213
},
215214
{
@@ -850,7 +849,7 @@
850849
"# The compiled module will have precision as specified by \"op_precision\".\n",
851850
"# Here, it will have FP16 precision.\n",
852851
"trt_model_fp16 = torch_tensorrt.compile(model, inputs = [torch_tensorrt.Input((128, 3, 224, 224), dtype=torch.half)],\n",
853-
" enabled_precisions = {torch.half}, # Run with FP32\n",
852+
" enabled_precisions = {torch.half}, # Run with FP16\n",
854853
" workspace_size = 1 << 22\n",
855854
")\n"
856855
]
@@ -894,7 +893,7 @@
894893
"<a id=\"5\"></a>\n",
895894
"## 5. Conclusion\n",
896895
"\n",
897-
"In this notebook, we have walked through the complete process of compiling TorchScript models with Torch-TensorRT for EfficientNet-B0 model and test the performance impact of the optimization. With Torch-TensorRT, we observe a speedup of **1.84x** with FP32, and **5.2x** with FP16 on an NVIDIA 3090 GPU. These acceleration numbers will vary from GPU to GPU(as well as implementation to implementation based on the ops used) and we encorage you to try out latest generation of Data center compute cards for maximum acceleration.\n",
896+
"In this notebook, we have walked through the complete process of compiling TorchScript models with Torch-TensorRT for ResNet-50 model and test the performance impact of the optimization. With Torch-TensorRT, we observe a speedup of **1.84x** with FP32, and **5.2x** with FP16 on an NVIDIA 3090 GPU. These acceleration numbers will vary from GPU to GPU(as well as implementation to implementation based on the ops used) and we encorage you to try out latest generation of Data center compute cards for maximum acceleration.\n",
898897
"\n",
899898
"### What's next\n",
900899
"Now it's time to try Torch-TensorRT on your own model. If you run into any issues, you can fill them at https://github.com/pytorch/TensorRT. Your involvement will help future development of Torch-TensorRT.\n"

0 commit comments

Comments
 (0)