Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2410.08292
Cited By
Can Looped Transformers Learn to Implement Multi-step Gradient Descent for In-context Learning?
10 October 2024
Khashayar Gatmiry
Nikunj Saunshi
Sashank J. Reddi
Stefanie Jegelka
Sanjiv Kumar
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Can Looped Transformers Learn to Implement Multi-step Gradient Descent for In-context Learning?"
4 / 4 papers shown
Title
Looped ReLU MLPs May Be All You Need as Practical Programmable Computers
Yingyu Liang
Zhizhou Sha
Zhenmei Shi
Zhao-quan Song
Yufa Zhou
81
17
0
21 Feb 2025
Relaxed Recursive Transformers: Effective Parameter Sharing with Layer-wise LoRA
Sangmin Bae
Adam Fisch
Hrayr Harutyunyan
Ziwei Ji
Seungyeon Kim
Tal Schuster
KELM
61
5
0
28 Oct 2024
Context-Scaling versus Task-Scaling in In-Context Learning
Amirhesam Abedsoltan
Adityanarayanan Radhakrishnan
Jingfeng Wu
M. Belkin
ReLM
LRM
29
3
0
16 Oct 2024
On Expressive Power of Looped Transformers: Theoretical Analysis and Enhancement via Timestep Encoding
Kevin Xu
Issei Sato
37
3
0
02 Oct 2024
1