ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.06708
  4. Cited By
Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free

Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free

10 May 2025
Z. Qiu
Z. Wang
Bo Zheng
Zeyu Huang
Kaiyue Wen
S. Yang
Rui Men
Le Yu
Fei Huang
Suozhi Huang
Dayiheng Liu
Jingren Zhou
Junyang Lin
    MoE
ArXivPDFHTML

Papers citing "Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free"

Title
No papers