# Copyright 2022 NVIDIA Corporation. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ==============================================================================

# Each user is responsible for checking the content of datasets and the
# applicable licenses and determining if suitable for the intended use.
https://developer.download.nvidia.com/notebooks/dlsw-notebooks/merlin_transformers4rec_getting-started-session-based-02-session-based-xlnet-with-pyt/nvidia_logo.png

Session-based Recommendation with XLNET

This notebook is created using the latest stable merlin-pytorch container.

In this notebook we introduce the Transformers4Rec library for sequential and session-based recommendation. This notebook uses the PyTorch API. Transformers4Rec integrates with the popular HuggingFace’s Transformers and makes it possible to experiment with a cutting-edge implementation of the latest NLP Transformer architectures.

We demonstrate how to build a session-based recommendation model with the XLNET Transformer architecture. The XLNet architecture was designed to leverage the best of both auto-regressive language modeling and auto-encoding with its Permutation Language Modeling training method. In this example we will use XLNET with masked language modeling (MLM) training method, which showed very promising results in the experiments conducted in our ACM RecSys’21 paper.

In the previous notebook we went through our ETL pipeline with the NVTabular library, and created sequential features to be used in training a session-based recommendation model. In this notebook we will learn:

  • Accelerating data loading of parquet files with multiple features on PyTorch using NVTabular library

  • Training and evaluating a Transformer-based (XLNET-MLM) session-based recommendation model with multiple features

Build a DL model with Transformers4Rec library

Transformers4Rec supports multiple input features and provides configurable building blocks that can be easily combined for custom architectures:

  • TabularSequenceFeatures class that reads from schema and creates an input block. This input module combines different types of features (continuous, categorical & text) to a sequence.

  • MaskSequence to define masking schema and prepare the masked inputs and labels for the selected LM task.

  • TransformerBlock class that supports HuggingFace Transformers for session-based and sequential-based recommendation models.

  • SequentialBlock creates the body by mimicking torch.nn.sequential class. It is designed to define our model as a sequence of layers.

  • Head where we define the prediction task of the model.

  • NextItemPredictionTask is the class to support next item prediction task.

  • Trainer extends the Trainer class from HF transformers and manages the model training and evaluation.

You can check the full documentation of Transformers4Rec if needed.

Figure 1 illustrates Transformers4Rec meta-architecture and how each module/block interacts with each other.

tf4rec_meta

Import required libraries

import os

os.environ["CUDA_VISIBLE_DEVICES"]="0"

import glob
import torch 

from transformers4rec import torch as tr
from transformers4rec.torch.ranking_metric import NDCGAt, AvgPrecisionAt, RecallAt
from transformers4rec.torch.utils.examples_utils import wipe_memory
/usr/local/lib/python3.8/dist-packages/tqdm/auto.py:21: TqdmWarning: IProgress not found. Please update jupyter and ipywidgets. See https://ipywidgets.readthedocs.io/en/stable/user_install.html
  from .autonotebook import tqdm as notebook_tqdm
/usr/local/lib/python3.8/dist-packages/merlin/dtypes/mappings/tf.py:52: UserWarning: Tensorflow dtype mappings did not load successfully due to an error: No module named 'tensorflow'
  warn(f"Tensorflow dtype mappings did not load successfully due to an error: {exc.msg}")

Transformers4Rec library relies on a schema object to automatically build all necessary layers to represent, normalize and aggregate input features. As you can see below, schema.pb is a protobuf file that contains metadata including statistics about features such as cardinality, min and max values and also tags features based on their characteristics and dtypes (e.g., categorical, continuous, list, integer).

Set the schema object

We create the schema object by reading the processed train parquet file generated by NVTabular pipeline in the previous, 01-ETL-with-NVTabular, notebook.

INPUT_DATA_DIR = os.environ.get("INPUT_DATA_DIR", "/workspace/data")
OUTPUT_DIR = os.environ.get("OUTPUT_DIR", f"{INPUT_DATA_DIR}/sessions_by_day")
from merlin.schema import Schema
from merlin.io import Dataset

train = Dataset(os.path.join(INPUT_DATA_DIR, "processed_nvt/part_0.parquet"))
schema = train.schema
/usr/local/lib/python3.8/dist-packages/merlin/schema/tags.py:149: UserWarning: Compound tags like Tags.ITEM_ID have been deprecated and will be removed in a future version. Please use the atomic versions of these tags, like [<Tags.ITEM: 'item'>, <Tags.ID: 'id'>].
  warnings.warn(
# You can select a subset of features for training
schema = schema.select_by_name(['item_id-list', 
                                'category-list', 
                                'weekday_sin-list',
                                'age_days-list'])

Let’s print out the schema.

schema
name tags dtype is_list is_ragged properties.num_buckets properties.freq_threshold properties.max_size properties.start_index properties.cat_path properties.embedding_sizes.cardinality properties.embedding_sizes.dimension properties.domain.min properties.domain.max properties.domain.name properties.value_count.min properties.value_count.max
0 item_id-list (Tags.ITEM_ID, Tags.ID, Tags.LIST, Tags.CATEGO... DType(name='int64', element_type=<ElementType.... True False NaN 0.0 0.0 0.0 .//categories/unique.item_id.parquet 497.0 52.0 0.0 496.0 item_id 20 20
1 category-list (Tags.CATEGORICAL, Tags.LIST) DType(name='int64', element_type=<ElementType.... True False NaN 0.0 0.0 0.0 .//categories/unique.category.parquet 143.0 26.0 0.0 142.0 category 20 20
2 weekday_sin-list (Tags.CONTINUOUS, Tags.LIST) DType(name='float32', element_type=<ElementTyp... True False NaN NaN NaN NaN NaN NaN NaN NaN NaN NaN 20 20
3 age_days-list (Tags.CONTINUOUS, Tags.LIST) DType(name='float32', element_type=<ElementTyp... True False NaN NaN NaN NaN NaN NaN NaN NaN NaN NaN 20 20

Define the sequential input module

Below we define our input block using the TabularSequenceFeatures class. The from_schema() method processes the schema and creates the necessary layers to represent features and aggregate them. It keeps only features tagged as categorical and continuous and supports data aggregation methods like concat and elementwise-sum. It also supports data augmentation techniques like stochastic swap noise. It outputs an interaction representation after combining all features and also the input mask according to the training task (more on this later).

The max_sequence_length argument defines the maximum sequence length of our sequential input, and if continuous_projection argument is set, all numerical features are concatenated and projected by an MLP block so that continuous features are represented by a vector of size defined by user, which is 64 in this example.

inputs = tr.TabularSequenceFeatures.from_schema(
        schema,
        max_sequence_length=20,
        continuous_projection=64,
        masking="mlm",
        d_output=100,
)

The output of the TabularSequenceFeatures module is the sequence of interactions embedding vectors defined in the following steps:

    1. Create sequence inputs: If the schema contains non sequential features, expand each feature to a sequence by repeating the value as many times as the max_sequence_length value.

    1. Get a representation vector of categorical features: Project each sequential categorical feature using the related embedding table. The resulting tensor is of shape (bs, max_sequence_length, embed_dim).

    1. Project scalar values if continuous_projection is set : Apply an MLP layer with hidden size equal to continuous_projection vector size value. The resulting tensor is of shape (batch_size, max_sequence_length, continuous_projection).

    1. Aggregate the list of features vectors to represent each interaction in the sequence with one vector: For example, concat will concat all vectors based on the last dimension -1 and the resulting tensor will be of shape (batch_size, max_sequence_length, D) where D is the sum over all embedding dimensions and the value of continuous_projection.

    1. If masking schema is set (needed only for the NextItemPredictionTask training), the masked labels are derived from the sequence of raw item-ids and the sequence of interactions embeddings are processed to mask information about the masked positions.

Define the Transformer block

In the next cell, the whole model is build with a few lines of code. Here is a brief explanation of the main classes:

  • XLNetConfig - We have injected in the HF transformers config classes like XLNetConfigthe build() method that provides default configuration to Transformer architectures for session-based recommendation. Here we use it to instantiate and configure an XLNET architecture.

  • TransformerBlock class integrates with HF Transformers, which are made available as a sequence processing module for session-based and sequential-based recommendation models.

  • NextItemPredictionTask supports the next-item prediction task. We also support other predictions tasks, like classification and regression for the whole sequence.

# Define XLNetConfig class and set default parameters for HF XLNet config  
transformer_config = tr.XLNetConfig.build(
    d_model=64, n_head=4, n_layer=2, total_seq_length=20
)
# Define the model block including: inputs, masking, projection and transformer block.
body = tr.SequentialBlock(
    inputs, tr.MLPBlock([64]), tr.TransformerBlock(transformer_config, masking=inputs.masking)
)

# Define the evaluation top-N metrics and the cut-offs
metrics = [NDCGAt(top_ks=[20, 40], labels_onehot=True),  
           RecallAt(top_ks=[20, 40], labels_onehot=True)]

# Define a head related to next item prediction task 
head = tr.Head(
    body,
    tr.NextItemPredictionTask(weight_tying=True, 
                              metrics=metrics),
    inputs=inputs,
)

# Get the end-to-end Model class 
model = tr.Model(head)

Note that we can easily define an RNN-based model inside the SequentialBlock instead of a Transformer-based model. You can explore this tutorial for a GRU-based model example.

Train the model

We use the Merlin Dataloader’s PyTorch Dataloader for optimized loading of multiple features from input parquet files. You can learn more about this data loader here.

Set Training arguments

per_device_train_batch_size = int(os.environ.get(
    "per_device_train_batch_size", 
    '128'
))

per_device_eval_batch_size = int(os.environ.get(
    "per_device_eval_batch_size", 
    '32'
))
from transformers4rec.config.trainer import T4RecTrainingArguments
from transformers4rec.torch import Trainer
# Set hyperparameters for training 
train_args = T4RecTrainingArguments(data_loader_engine='merlin', 
                                    dataloader_drop_last = True,
                                    gradient_accumulation_steps = 1,
                                    per_device_train_batch_size = per_device_train_batch_size, 
                                    per_device_eval_batch_size = per_device_eval_batch_size,
                                    output_dir = "./tmp", 
                                    learning_rate=0.0005,
                                    lr_scheduler_type='cosine', 
                                    learning_rate_num_cosine_cycles_by_epoch=1.5,
                                    num_train_epochs=5,
                                    max_sequence_length=20, 
                                    report_to = [],
                                    logging_steps=50,
                                    no_cuda=False)

Note that we add an argument data_loader_engine='merlin' to automatically load the features needed for training using the schema. The default value is merlin for optimized GPU-based data-loading. Optionally a PyarrowDataLoader (pyarrow) can also be used as a basic option, but it is slower and works only for small datasets, as the full data is loaded to CPU memory.

Daily Fine-Tuning: Training over a time window

Here we do daily fine-tuning meaning that we use the first day to train and second day to evaluate, then we use the second day data to train the model by resuming from the first step, and evaluate on the third day, so on and so forth.

We have extended the HuggingFace transformers Trainer class (PyTorch only) to support evaluation of RecSys metrics. In this example, the evaluation of the session-based recommendation model is performed using traditional Top-N ranking metrics such as Normalized Discounted Cumulative Gain (NDCG@20) and Hit Rate (HR@20). NDCG accounts for rank of the relevant item in the recommendation list and is a more fine-grained metric than HR, which only verifies whether the relevant item is among the top-n items. HR@n is equivalent to Recall@n when there is only one relevant item in the recommendation list.

# Instantiate the T4Rec Trainer, which manages training and evaluation for the PyTorch API
trainer = Trainer(
    model=model,
    args=train_args,
    schema=schema,
    compute_metrics=True,
)

Define the output folder of the processed parquet files:

start_window_index = int(os.environ.get(
    "start_window_index", 
    '1'
))

final_window_index = int(os.environ.get(
    "final_window_index", 
    '8'
))
start_time_window_index = start_window_index
final_time_window_index = final_window_index
#Iterating over days of one week
for time_index in range(start_time_window_index, final_time_window_index):
    # Set data 
    time_index_train = time_index
    time_index_eval = time_index + 1
    train_paths = glob.glob(os.path.join(OUTPUT_DIR, f"{time_index_train}/train.parquet"))
    eval_paths = glob.glob(os.path.join(OUTPUT_DIR, f"{time_index_eval}/valid.parquet"))
    print(train_paths)
    
    # Train on day related to time_index 
    print('*'*20)
    print("Launch training for day %s are:" %time_index)
    print('*'*20 + '\n')
    trainer.train_dataset_or_path = train_paths
    trainer.reset_lr_scheduler()
    trainer.train()
    trainer.state.global_step +=1
    print('finished')
    
    # Evaluate on the following day
    trainer.eval_dataset_or_path = eval_paths
    train_metrics = trainer.evaluate(metric_key_prefix='eval')
    print('*'*20)
    print("Eval results for day %s are:\t" %time_index_eval)
    print('\n' + '*'*20 + '\n')
    for key in sorted(train_metrics.keys()):
        print(" %s = %s" % (key, str(train_metrics[key]))) 
    wipe_memory()
['/workspace/data/sessions_by_day/1/train.parquet']
********************
Launch training for day 1 are:
********************
***** Running training *****
  Num examples = 1664
  Num Epochs = 5
  Instantaneous batch size per device = 128
  Total train batch size (w. parallel, distributed & accumulation) = 128
  Gradient Accumulation steps = 1
  Total optimization steps = 65
[65/65 00:02, Epoch 5/5]
Step Training Loss
50 5.772000

Training completed. Do not forget to share your model on huggingface.co/models =)
finished
[6/6 00:20]
********************
Eval results for day 2 are:	

********************

 eval_/loss = 5.125837326049805
 eval_/next-item/ndcg_at_20 = 0.15405046939849854
 eval_/next-item/ndcg_at_40 = 0.19994235038757324
 eval_/next-item/recall_at_20 = 0.421875
 eval_/next-item/recall_at_40 = 0.6458333730697632
 eval_runtime = 0.1725
 eval_samples_per_second = 1113.287
 eval_steps_per_second = 34.79
['/workspace/data/sessions_by_day/2/train.parquet']
********************
Launch training for day 2 are:
********************
***** Running training *****
  Num examples = 1664
  Num Epochs = 5
  Instantaneous batch size per device = 128
  Total train batch size (w. parallel, distributed & accumulation) = 128
  Gradient Accumulation steps = 1
  Total optimization steps = 65
[65/65 00:02, Epoch 5/5]
Step Training Loss
50 4.879700

Training completed. Do not forget to share your model on huggingface.co/models =)
finished
***** Running training *****
  Num examples = 1664
  Num Epochs = 5
  Instantaneous batch size per device = 128
  Total train batch size (w. parallel, distributed & accumulation) = 128
  Gradient Accumulation steps = 1
  Total optimization steps = 65
********************
Eval results for day 3 are:	

********************

 eval_/loss = 4.680967807769775
 eval_/next-item/ndcg_at_20 = 0.17892448604106903
 eval_/next-item/ndcg_at_40 = 0.22572219371795654
 eval_/next-item/recall_at_20 = 0.484375
 eval_/next-item/recall_at_40 = 0.7135416865348816
 eval_runtime = 0.1745
 eval_samples_per_second = 1100.148
 eval_steps_per_second = 34.38
['/workspace/data/sessions_by_day/3/train.parquet']
********************
Launch training for day 3 are:
********************
[65/65 00:03, Epoch 5/5]
Step Training Loss
50 4.581400

Training completed. Do not forget to share your model on huggingface.co/models =)
finished
********************
Eval results for day 4 are:	

********************

 eval_/loss = 4.464942455291748
 eval_/next-item/ndcg_at_20 = 0.2020609825849533
 eval_/next-item/ndcg_at_40 = 0.24695619940757751
 eval_/next-item/recall_at_20 = 0.5260416865348816
 eval_/next-item/recall_at_40 = 0.7447916865348816
 eval_runtime = 0.165
 eval_samples_per_second = 1163.669
 eval_steps_per_second = 36.365
['/workspace/data/sessions_by_day/4/train.parquet']
********************
Launch training for day 4 are:
********************
***** Running training *****
  Num examples = 1664
  Num Epochs = 5
  Instantaneous batch size per device = 128
  Total train batch size (w. parallel, distributed & accumulation) = 128
  Gradient Accumulation steps = 1
  Total optimization steps = 65
[65/65 00:02, Epoch 5/5]
Step Training Loss
50 4.519200

Training completed. Do not forget to share your model on huggingface.co/models =)
finished
***** Running training *****
  Num examples = 1664
  Num Epochs = 5
  Instantaneous batch size per device = 128
  Total train batch size (w. parallel, distributed & accumulation) = 128
  Gradient Accumulation steps = 1
  Total optimization steps = 65
********************
Eval results for day 5 are:	

********************

 eval_/loss = 4.440032958984375
 eval_/next-item/ndcg_at_20 = 0.20033857226371765
 eval_/next-item/ndcg_at_40 = 0.24777457118034363
 eval_/next-item/recall_at_20 = 0.5416666865348816
 eval_/next-item/recall_at_40 = 0.7760416865348816
 eval_runtime = 0.1673
 eval_samples_per_second = 1147.626
 eval_steps_per_second = 35.863
['/workspace/data/sessions_by_day/5/train.parquet']
********************
Launch training for day 5 are:
********************
[65/65 00:02, Epoch 5/5]
Step Training Loss
50 4.479100

Training completed. Do not forget to share your model on huggingface.co/models =)
finished
***** Running training *****
  Num examples = 1664
  Num Epochs = 5
  Instantaneous batch size per device = 128
  Total train batch size (w. parallel, distributed & accumulation) = 128
  Gradient Accumulation steps = 1
  Total optimization steps = 65
********************
Eval results for day 6 are:	

********************

 eval_/loss = 4.369968891143799
 eval_/next-item/ndcg_at_20 = 0.18754208087921143
 eval_/next-item/ndcg_at_40 = 0.23825709521770477
 eval_/next-item/recall_at_20 = 0.5416666865348816
 eval_/next-item/recall_at_40 = 0.7864583730697632
 eval_runtime = 0.1616
 eval_samples_per_second = 1188.118
 eval_steps_per_second = 37.129
['/workspace/data/sessions_by_day/6/train.parquet']
********************
Launch training for day 6 are:
********************
[65/65 00:02, Epoch 5/5]
Step Training Loss
50 4.459600

Training completed. Do not forget to share your model on huggingface.co/models =)
finished
********************
Eval results for day 7 are:	

********************

 eval_/loss = 4.5033793449401855
 eval_/next-item/ndcg_at_20 = 0.1925697922706604
 eval_/next-item/ndcg_at_40 = 0.23315435647964478
 eval_/next-item/recall_at_20 = 0.53125
 eval_/next-item/recall_at_40 = 0.7291666865348816
 eval_runtime = 0.1677
 eval_samples_per_second = 1144.864
 eval_steps_per_second = 35.777
***** Running training *****
  Num examples = 1664
  Num Epochs = 5
  Instantaneous batch size per device = 128
  Total train batch size (w. parallel, distributed & accumulation) = 128
  Gradient Accumulation steps = 1
  Total optimization steps = 65
['/workspace/data/sessions_by_day/7/train.parquet']
********************
Launch training for day 7 are:
********************
[65/65 00:02, Epoch 5/5]
Step Training Loss
50 4.455300

Training completed. Do not forget to share your model on huggingface.co/models =)
finished
********************
Eval results for day 8 are:	

********************

 eval_/loss = 4.4695820808410645
 eval_/next-item/ndcg_at_20 = 0.1891615390777588
 eval_/next-item/ndcg_at_40 = 0.24406516551971436
 eval_/next-item/recall_at_20 = 0.5104166865348816
 eval_/next-item/recall_at_40 = 0.7760416865348816
 eval_runtime = 0.1806
 eval_samples_per_second = 1063.043
 eval_steps_per_second = 33.22

Re-compute evaluation metrics of the validation data

eval_data_paths = glob.glob(os.path.join(OUTPUT_DIR, f"{time_index_eval}/valid.parquet"))
# set new data from day 7
eval_metrics = trainer.evaluate(eval_dataset=eval_data_paths, metric_key_prefix='eval')
for key in sorted(eval_metrics.keys()):
    print("  %s = %s" % (key, str(eval_metrics[key])))
  eval_/loss = 4.4695820808410645
  eval_/next-item/ndcg_at_20 = 0.1891615390777588
  eval_/next-item/ndcg_at_40 = 0.24406516551971436
  eval_/next-item/recall_at_20 = 0.5104166865348816
  eval_/next-item/recall_at_40 = 0.7760416865348816
  eval_runtime = 0.1776
  eval_samples_per_second = 1081.009
  eval_steps_per_second = 33.782

Save the model

Let’s save the model to be able to load it back at inference step. Using model.save(), we save the model as a pkl file in the given path.

model_path= os.environ.get("OUTPUT_DIR", f"{INPUT_DATA_DIR}/saved_model")
model.save(model_path)

That’s it! You have just trained your session-based recommendation model using Transformers4Rec. Now you can move on to the next notebook 03-serving-session-based-model-torch-backend. Please shut down this kernel to free the GPU memory before you start the next one.

Tip: We can easily log and visualize model training and evaluation on Weights & Biases (W&B), TensorBoard, or NVIDIA DLLogger. By default, the HuggingFace transformers Trainer (which we extend) uses Weights & Biases (W&B) to log training and evaluation metrics, which provides nice visualization results and comparison between different runs.