HugeCTR Example Notebooks
This directory contains a set of Jupyter notebook that demonstrate how to use HugeCTR.
The simplest way to run a one of our notebooks is with a Docker container. A container provides a self-contained, isolated, and reproducible environment for repetitive experiments. Docker images are available from the NVIDIA GPU Cloud (NGC).
1. Clone the HugeCTR Repository
Use the following command to clone the HugeCTR repository:
git clone https://github.com/NVIDIA/HugeCTR
2. Pull the NGC Docker and run it
Pull the container using the following command:
docker pull nvcr.io/nvidia/merlin/merlin-hugectr:24.04
Launch the container in interactive mode (mount the HugeCTR root directory into the container for your convenience) by running this command:
docker run --gpus all --rm -it --cap-add SYS_NICE --ipc=host --ulimit memlock=-1 --ulimit stack=67108864 -u root -v $(pwd):/HugeCTR -w /HugeCTR --network=host --runtime=nvidia nvcr.io/nvidia/merlin/merlin-hugectr:24.04
To run the Sparse Operation Kit notebooks, specify the
nvcr.io/nvidia/merlin/merlin-tensorflow:24.04
container.
3. Customized Building (Optional)
HugeCTR is already installed in the NGC container. But you can also setup HugeCTR from source to customize the build more. This is useful for developmental purposes.
Go to HugeCTR repo and update third party modules
$ cd HugeCTR
$ git submodule update --init --recursive
There are options to customize the build using parameters, which are detailed here Here are some examples of how you can build HugeCTR using these build options:
# Example 1
$ mkdir -p build && cd build
$ cmake -DCMAKE_BUILD_TYPE=Release -DSM=70 .. # Target is NVIDIA V100 with all others by default
$ make -j && make install
# Example 2
$ mkdir -p build && cd build
$ cmake -DCMAKE_BUILD_TYPE=Release -DSM="70;80" -DENABLE_MULTINODES=ON .. # Target is NVIDIA V100 / A100 with the multi-node mode on.
$ make -j && make install
By default, HugeCTR is installed at /usr/local
. However, you can use CMAKE_INSTALL_PREFIX
to install HugeCTR to non-default location:
$ cmake -DCMAKE_INSTALL_PREFIX=/opt/HugeCTR -DSM=70 ..
Refer to the
How to Start Your Development documentation for more details on building HugeCTR From Source
4. Start the Jupyter Notebook
Start Jupyter using these commands:
cd /HugeCTR/notebooks jupyter-notebook --allow-root --ip 0.0.0.0 --port 8888 --NotebookApp.token='hugectr'
Connect to your host machine using the 8888 port by accessing its IP address or name from your web browser:
http://[host machine]:8888
Use the token available from the output by running the command above to log in. For example:
http://[host machine]:8888/?token=aae96ae9387cd28151868fee318c3b3581a2d794f3b25c6b
Optional: Import MPI.
By default, HugeCTR initializes and finalizes MPI when you run the
import hugectr
statement within the NGC Merlin container. If you build and install HugeCTR yourself, specify theENABLE_MULTINODES=ON
argument when you build. See Build HugeCTR from Source.If your program uses MPI for a reason other than interacting with HugeCTR, initialize MPI with the
from mpi4py import MPI
statement before you import HugeCTR.Important Note:
HugeCTR is written in CUDA/C++ and wrapped to Python using Pybind11. The C++ output will not display in Notebook cells unless you run the Python script in a command line manner.
Notebook List
The notebooks are located within the container and can be found in the /HugeCTR/notebooks
directory.
Here’s a list of notebooks that you can run:
hugectr_e2e_demo_with_nvtabular.ipynb: Notebook to preprocess data using NVTabular, train the model with HugeCTR, and do the offline inference with the HugeCTR HPS.
continuous_training.ipynb (deprecated): Notebook to introduce how to deploy continued training with HugeCTR.
~multi_gpu_offline_inference.ipynb~: It was deprecated. Check out this HPS TRT notebook as an alternative.
hps_demo.ipynb: Demonstrate how to utilize HPS Python APIs together with ONNX Runtime APIs to create an ensemble inference model.
training_and_inference_with_remote_filesystem.ipynb: Demonstrates how to train a model with data that is stored in a remote file system such as Hadoop HDFS and AWS S3.
The multi-modal-data series of notebooks demonstrate how to use of multi-modal data such as text and images for the task of movie recommendation. The notebooks use the Movielens-25M dataset.
More notebooks on the Hierarchical Parameter Server (HPS) are available with its TensorRT and Tensorflow plugins.
For Sparse Operation Kit notebooks, refer to the sparse_operation_kit/notebooks/ directory of the repository or the documentation.
System Specifications
The specifications of the system on which each notebook can run successfully are summarized in the table. The notebooks are verified on the system below but it does not mean the minimum requirements.
Notebook |
CPU |
GPU |
#GPUs |
Author |
---|---|---|---|---|
Intel® Xeon® CPU E5-2698 v4 @ 2.20GHz |
Tesla V100-SXM2-32GB |
1 |
Vinh Nguyen |
|
Intel® Xeon® CPU E5-2698 v4 @ 2.20GHz |
Tesla V100-SXM2-32GB |
1 |
Xiaolei Shi |
|
Intel® Xeon® CPU E5-2698 v4 @ 2.20GHz |
Tesla V100-SXM2-32GB |
1 |
Kingsley Liu, Matthias Langer and Yingcan Wei |
|
Intel® Xeon® CPU E5-2698 v4 @ 2.20GHz |
Tesla V100-SXM2-32GB |
1 |
Jerry Shi |
|
Intel® Xeon® CPU E5-2698 v4 @ 2.20GHz |
Tesla V100-SXM2-32GB |
1 |
Jerry Shi |