ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.17505
37
0

L-MTP: Leap Multi-Token Prediction Beyond Adjacent Context for Large Language Models

23 May 2025
Xiaohao Liu
Xiaobo Xia
Weixiang Zhao
Manyi Zhang
Xianzhi Yu
Xiu Su
Shuo Yang
See-Kiong Ng
Tat-Seng Chua
    KELM
    LRM
ArXivPDFHTML
Abstract

Large language models (LLMs) have achieved notable progress. Despite their success, next-token prediction (NTP), the dominant method for LLM training and inference, is constrained in both contextual coverage and inference efficiency due to its inherently sequential process. To overcome these challenges, we propose leap multi-token prediction~(L-MTP), an innovative token prediction method that extends the capabilities of multi-token prediction (MTP) by introducing a leap-based mechanism. Unlike conventional MTP, which generates multiple tokens at adjacent positions, L-MTP strategically skips over intermediate tokens, predicting non-sequential ones in a single forward pass. This structured leap not only enhances the model's ability to capture long-range dependencies but also enables a decoding strategy specially optimized for non-sequential leap token generation, effectively accelerating inference. We theoretically demonstrate the benefit of L-MTP in improving inference efficiency. Experiments across diverse benchmarks validate its merit in boosting both LLM performance and inference speed. The source code will be publicly available.

View on arXiv
@article{liu2025_2505.17505,
  title={ L-MTP: Leap Multi-Token Prediction Beyond Adjacent Context for Large Language Models },
  author={ Xiaohao Liu and Xiaobo Xia and Weixiang Zhao and Manyi Zhang and Xianzhi Yu and Xiu Su and Shuo Yang and See-Kiong Ng and Tat-Seng Chua },
  journal={arXiv preprint arXiv:2505.17505},
  year={ 2025 }
}
Comments on this paper