As of v3.0 release, the HugeCTR container is no longer being released separately. The unified Merlin container is now available on the NVIDIA GPU Cloud (NGC) and can be used for research and production purposes. It provides support for preprocessing data using NVTabular, model training using HugeCTR, Inference using the Triton HugeCTR Backend, and the Embedding TensorFlow (TF) plugin. If you want to use any one of these use cases with the exception of preprocessing data using NVTabular, you can build a container using the dockerfile on your own.
The train.Dockerfile
supports the devel
and release
build modes. A container in release
mode contains the necessary libraries and executable files so that HugeCTR and its Python interface can be used, but with no source code/dataset/script stored in it. A container in devel
mode can be used for development purposes. After running the container, you can download the HugeCTR source code and build following the steps shown here.
You can build the container by running one of these commands from the current directory:
# release mode
docker build -t hugectr:release -f train.Dockerfile --build-arg RELEASE=true .
# devel mode
docker build -t hugectr:devel -f train.Dockerfile .
If you want to use the HugeCTR inference functionality based on the Triton Inference Server HugeCTR Backend, build a container by running the following command from the current directory:
docker build -t hugectr:inference -f infer.Dockerfile --build-arg RELEASE=true .
If you want to use the HugeCTR embedding TF plugin, build a container by running the following command from the current directory:
docker build -t hugectr:plugin-embedding -f plugin-embedding.Dockerfile .