Scaling Large Datasets with Criteo

Criteo provides the largest publicly available dataset for recommender systems with a size of 1TB of uncompressed click logs that contain 4 billion examples. We demonstrate how to scale NVTabular, as well as:

  • Use multiple GPUs and nodes with NVTabular for feature engineering.

  • Train recommender system models with the NVTabular dataloader for PyTorch.

  • Train recommender system models with the NVTabular dataloader for TensorFlow

  • Train recommender system models with HugeCTR using multiple GPUs.

  • Inference with the Triton Inference Server and TensorFlow or HugeCTR.

There are example compose files for use with Docker in the scaling-criteo <https://github.com/NVIDIA-Merlin/Merlin/tree/main/examples/scaling-criteo>_ directory of the Merlin repository on GitHub. The compose files enable you to run a pair of training and inference containers.

Explore the following notebooks: