ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.01405
26
3

On Expressive Power of Looped Transformers: Theoretical Analysis and Enhancement via Timestep Encoding

2 October 2024
Kevin Xu
Issei Sato
ArXivPDFHTML
Abstract

Looped Transformers provide advantages in parameter efficiency, computational capabilities, and generalization for reasoning tasks. However, their expressive power regarding function approximation remains underexplored. In this paper, we establish the approximation rate of Looped Transformers by defining the modulus of continuity for sequence-to-sequence functions. This reveals a limitation specific to the looped architecture. That is, the analysis prompts the incorporation of scaling parameters for each loop, conditioned on timestep encoding. Experiments validate the theoretical results, showing that increasing the number of loops enhances performance, with further gains achieved through the timestep encoding.

View on arXiv
@article{xu2025_2410.01405,
  title={ On Expressive Power of Looped Transformers: Theoretical Analysis and Enhancement via Timestep Encoding },
  author={ Kevin Xu and Issei Sato },
  journal={arXiv preprint arXiv:2410.01405},
  year={ 2025 }
}
Comments on this paper