Liu, Yuwen and Qi, Lianyong and Mao, Xingyuan and Liu, Weiming and Fan, Xuhui and Ni, Qiang and Zhang, Xuyun and Zhang, Yang and Tian, Yuan and Beheshti, Amin (2025) Hyperbolic-Enhanced Mixture-of-Experts Mamba for Sequential Recommendation. In: The 40th Annual AAAI Conference on Artificial Intelligence : AAAI 2026. AAAI, pp. 1-9. (In Press)
AAAI_2026-Hyperbolic-Enhanced_Mixture.pdf - Accepted Version
Available under License Creative Commons Attribution-NonCommercial.
Download (14MB)
Abstract
Sequential recommendation has emerged as a fundamental task in various domains, aiming to predict a user’s next interaction based on historical behavior. Recent advances in deep sequence models, particularly Transformer-based architectures and the more recent Mamba, have substantially pushed the boundaries of sequential modeling performance. However, existing methods still face two critical challenges. First, many current approaches overlook the hierarchical structures and high-order dependencies among items, typically restricting representation learning to conventional Euclidean spaces, which limits their capacity to capture complex relational information. Second, although Mamba excels at long-range dependency modeling, its reliance on static Feed-Forward Networks (FFNs) hinders its ability to dynamically adapt to evolving user preferences across diverse contexts. To address these limitations, we propose a Hyperbolic-Enhanced Mixture-of-Experts Mamba recommender (HM2Rec) for sequential recommendation. HM2Rec first encodes user-item relationships through hyperbolic graph convolution to exploit hierarchical structure more effectively. Then, a Variational Graph Auto-Encoder (VGAE) is employed to reconstruct node embeddings, improving structural robustness. To further enhance sequential modeling, we integrate Rotary Positional Encoding (RoPE) into Mamba to better capture relative position dependencies, and replace the FFN with Mixture-of-Expert (MOE) module, enabling dynamic and personalized expert selection for each token. Our extensive experiments on four widely-used public datasets demonstrate that HM2Rec outperforms several advanced baseline models.