Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2311.14652
Cited By
One Pass Streaming Algorithm for Super Long Token Attention Approximation in Sublinear Space
24 November 2023
Raghav Addanki
Chenyang Li
Zhao-quan Song
Chiwun Yang
Re-assign community
ArXiv
PDF
HTML
Papers citing
"One Pass Streaming Algorithm for Super Long Token Attention Approximation in Sublinear Space"
7 / 7 papers shown
Title
A Survey on Proactive Dialogue Systems: Problems, Methods, and Prospects
Yang Deng
Wenqiang Lei
W. Lam
Tat-Seng Chua
56
25
0
04 May 2023
Is Your Code Generated by ChatGPT Really Correct? Rigorous Evaluation of Large Language Models for Code Generation
Jiawei Liu
Chun Xia
Yuyao Wang
Lingming Zhang
ELM
ALM
161
388
0
02 May 2023
Sparks of Artificial General Intelligence: Early experiments with GPT-4
Sébastien Bubeck
Varun Chandrasekaran
Ronen Eldan
J. Gehrke
Eric Horvitz
...
Scott M. Lundberg
Harsha Nori
Hamid Palangi
Marco Tulio Ribeiro
Yi Zhang
ELM
AI4MH
AI4CE
ALM
197
2,232
0
22 Mar 2023
How Do Transformers Learn Topic Structure: Towards a Mechanistic Understanding
Yuchen Li
Yuan-Fang Li
Andrej Risteski
87
61
0
07 Mar 2023
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
301
11,730
0
04 Mar 2022
Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
Ofir Press
Noah A. Smith
M. Lewis
234
476
0
27 Aug 2021
Big Bird: Transformers for Longer Sequences
Manzil Zaheer
Guru Guruganesh
Kumar Avinava Dubey
Joshua Ainslie
Chris Alberti
...
Philip Pham
Anirudh Ravula
Qifan Wang
Li Yang
Amr Ahmed
VLM
246
1,982
0
28 Jul 2020
1