ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2404.07519
  4. Cited By
LATTE: Low-Precision Approximate Attention with Head-wise Trainable
  Threshold for Efficient Transformer

LATTE: Low-Precision Approximate Attention with Head-wise Trainable Threshold for Efficient Transformer

11 April 2024
Jiing-Ping Wang
Ming-Guang Lin
An-Yeu Wu
Wu
ArXivPDFHTML

Papers citing "LATTE: Low-Precision Approximate Attention with Head-wise Trainable Threshold for Efficient Transformer"

1 / 1 papers shown
Title
Energon: Towards Efficient Acceleration of Transformers Using Dynamic
  Sparse Attention
Energon: Towards Efficient Acceleration of Transformers Using Dynamic Sparse Attention
Zhe Zhou
Junling Liu
Zhenyu Gu
Guangyu Sun
61
42
0
18 Oct 2021
1