Enhancing Transformers without Self-supervised Learning: A Loss Landscape Perspective in Sequential Recommendation

08/20/2023
by   Vivian Lai, et al.
0

Transformer and its variants are a powerful class of architectures for sequential recommendation, owing to their ability of capturing a user's dynamic interests from their past interactions. Despite their success, Transformer-based models often require the optimization of a large number of parameters, making them difficult to train from sparse data in sequential recommendation. To address the problem of data sparsity, previous studies have utilized self-supervised learning to enhance Transformers, such as pre-training embeddings from item attributes or contrastive data augmentations. However, these approaches encounter several training issues, including initialization sensitivity, manual data augmentations, and large batch-size memory bottlenecks. In this work, we investigate Transformers from the perspective of loss geometry, aiming to enhance the models' data efficiency and generalization in sequential recommendation. We observe that Transformers (e.g., SASRec) can converge to extremely sharp local minima if not adequately regularized. Inspired by the recent Sharpness-Aware Minimization (SAM), we propose SAMRec, which significantly improves the accuracy and robustness of sequential recommendation. SAMRec performs comparably to state-of-the-art self-supervised Transformers, such as S^3Rec and CL4SRec, without the need for pre-training or strong data augmentations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/14/2021

Contrastive Self-supervised Sequential Recommendation with Robust Augmentation

Sequential Recommendationdescribes a set of techniques to model dynamic ...
research
02/22/2021

UPRec: User-Aware Pre-training for Recommender Systems

Existing sequential recommendation methods rely on large amounts of trai...
research
08/18/2020

S^3-Rec: Self-Supervised Learning for Sequential Recommendation with Mutual Information Maximization

Recently, significant progress has been made in sequential recommendatio...
research
03/05/2022

Towards Efficient and Scalable Sharpness-Aware Minimization

Recently, Sharpness-Aware Minimization (SAM), which connects the geometr...
research
07/15/2022

A Systematic Review and Replicability Study of BERT4Rec for Sequential Recommendation

BERT4Rec is an effective model for sequential recommendation based on th...
research
06/13/2023

Is Anisotropy Inherent to Transformers?

The representation degeneration problem is a phenomenon that is widely o...
research
07/26/2023

Integrating Offline Reinforcement Learning with Transformers for Sequential Recommendation

We consider the problem of sequential recommendation, where the current ...

Please sign up or login with your details

Forgot password? Click here to reset