Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2112.08583
Cited By
Does Pre-training Induce Systematic Inference? How Masked Language Models Acquire Commonsense Knowledge
16 December 2021
Ian Porada
Alessandro Sordoni
Jackie C.K. Cheung
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Does Pre-training Induce Systematic Inference? How Masked Language Models Acquire Commonsense Knowledge"
4 / 4 papers shown
Title
Frequency Effects on Syntactic Rule Learning in Transformers
Jason W. Wei
Dan Garrette
Tal Linzen
Ellie Pavlick
80
62
0
14 Sep 2021
Scaling Laws for Neural Language Models
Jared Kaplan
Sam McCandlish
T. Henighan
Tom B. Brown
B. Chess
R. Child
Scott Gray
Alec Radford
Jeff Wu
Dario Amodei
226
4,453
0
23 Jan 2020
Megatron-LM: Training Multi-Billion Parameter Language Models Using Model Parallelism
M. Shoeybi
M. Patwary
Raul Puri
P. LeGresley
Jared Casper
Bryan Catanzaro
MoE
243
1,817
0
17 Sep 2019
Language Models as Knowledge Bases?
Fabio Petroni
Tim Rocktaschel
Patrick Lewis
A. Bakhtin
Yuxiang Wu
Alexander H. Miller
Sebastian Riedel
KELM
AI4MH
406
2,584
0
03 Sep 2019
1