Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2309.02373
Cited By
nanoT5: A PyTorch Framework for Pre-training and Fine-tuning T5-style Models with Limited Resources
5 September 2023
Piotr Nawrot
AI4CE
Re-assign community
ArXiv
PDF
HTML
Papers citing
"nanoT5: A PyTorch Framework for Pre-training and Fine-tuning T5-style Models with Limited Resources"
3 / 3 papers shown
Title
Towards Developmentally Plausible Rewards: Communicative Success as a Learning Signal for Interactive Language Models
Lennart Stöpler
Rufat Asadli
Mitja Nikolaus
Ryan Cotterell
Alex Warstadt
LRM
32
0
0
09 May 2025
No Train No Gain: Revisiting Efficient Training Algorithms For Transformer-based Language Models
Jean Kaddour
Oscar Key
Piotr Nawrot
Pasquale Minervini
Matt J. Kusner
13
41
0
12 Jul 2023
Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
Ofir Press
Noah A. Smith
M. Lewis
234
690
0
27 Aug 2021
1