Progressive Self-Attention Network with Unsymmetrical Positional Encoding for Sequential Recommendation

Yuehua Zhu, Bo Huang, Shaohua Jiang, Muli Yang, Yanhua Yang, Wenliang Zhong
2022 Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval  
In real-world recommendation systems, the preferences of users are often affected by long-term constant interests and short-term temporal needs. The recently proposed Transformer-based models have proved superior in the sequential recommendation, modeling temporal dynamics globally via the remarkable self-attention mechanism. However, all equivalent item-item interactions in original self-attention are cumbersome, failing to capture the drifting of users' local preferences, which contain
more » ... t short-term patterns. In this paper, we propose a novel interpretable convolutional self-attention, which efficiently captures both short-and long-term patterns with a progressive attention distribution. Specifically, a down-sampling convolution module is proposed to segment the overall long behavior sequence into a series of local subsequences. Accordingly, the segments are interacted with each item in the self-attention layer to produce locality-aware contextual representations, during which the quadratic complexity in original self-attention is reduced to nearly linear complexity. Moreover, to further enhance the robust feature learning in the context of Transformers, an unsymmetrical positional encoding strategy is carefully designed. Extensive experiments are carried out on realworld datasets, e.g., ML-1M, Amazon Books, and Yelp, indicating that the proposed method outperforms the state-of-the-art methods w.r.t. both effectiveness and efficiency. CCS CONCEPTS • Information systems → Recommender systems.
doi:10.1145/3477495.3531800 fatcat:53pypr36ira6fathwdrvrtlxey