ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.04421
31
0

LONGER: Scaling Up Long Sequence Modeling in Industrial Recommenders

7 May 2025
Zheng Chai
Qin Ren
Xijun Xiao
H. Yang
Bo Han
S. Zhang
Di Chen
Hui Lu
Wenlin Zhao
Lele Yu
Xionghang Xie
Shiru Ren
Xiang Sun
Yaocheng Tan
Peng Xu
Yuchao Zheng
Di Wu
ArXivPDFHTML
Abstract

Modeling ultra-long user behavior sequences is critical for capturing both long- and short-term preferences in industrial recommender systems. Existing solutions typically rely on two-stage retrieval or indirect modeling paradigms, incuring upstream-downstream inconsistency and computational inefficiency. In this paper, we present LONGER, a Long-sequence Optimized traNsformer for GPU-Efficient Recommenders. LONGER incorporates (i) a global token mechanism for stabilizing attention over long contexts, (ii) a token merge module with lightweight InnerTransformers and hybrid attention strategy to reduce quadratic complexity, and (iii) a series of engineering optimizations, including training with mixed-precision and activation recomputation, KV cache serving, and the fully synchronous model training and serving framework for unified GPU-based dense and sparse parameter updates. LONGER consistently outperforms strong baselines in both offline metrics and online A/B testing in both advertising and e-commerce services at ByteDance, validating its consistent effectiveness and industrial-level scaling laws. Currently, LONGER has been fully deployed at more than 10 influential scenarios at ByteDance, serving billion users.

View on arXiv
@article{chai2025_2505.04421,
  title={ LONGER: Scaling Up Long Sequence Modeling in Industrial Recommenders },
  author={ Zheng Chai and Qin Ren and Xijun Xiao and Huizhi Yang and Bo Han and Sijun Zhang and Di Chen and Hui Lu and Wenlin Zhao and Lele Yu and Xionghang Xie and Shiru Ren and Xiang Sun and Yaocheng Tan and Peng Xu and Yuchao Zheng and Di Wu },
  journal={arXiv preprint arXiv:2505.04421},
  year={ 2025 }
}
Comments on this paper