Scaling to Large Datasets with Criteo
Criteo provides the largest publicly available dataset for recommender systems. The dataset is 1 TB uncompressed click logs of 4 billion examples. The example notebooks show how to scale NVTabular in the following ways:
- Using multiple GPUs and multiple nodes with NVTabular for ETL. 
- Training recommender system model with NVTabular dataloader for PyTorch. 
Refer to the following notebooks:
- Training a model: HugeCTR | TensorFlow 
- Use Triton Inference Server to serve a model: HugeCTR | TensorFlow