Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2410.03001
Cited By
Can Transformers Learn
n
n
n
-gram Language Models?
3 October 2024
Anej Svete
Nadav Borenstein
M. Zhou
Isabelle Augenstein
Ryan Cotterell
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Can Transformers Learn $n$-gram Language Models?"
2 / 2 papers shown
Title
Bigram Subnetworks: Mapping to Next Tokens in Transformer Language Models
Tyler A. Chang
Benjamin Bergen
38
0
0
21 Apr 2025
Better Estimation of the KL Divergence Between Language Models
Afra Amini
Tim Vieira
Ryan Cotterell
41
0
0
14 Apr 2025
1