Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
1803.05457
Cited By
Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge
14 March 2018
Peter Clark
Isaac Cowhey
Oren Etzioni
Tushar Khot
Ashish Sabharwal
Carissa Schoenick
Oyvind Tafjord
ELM
RALM
LRM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge"
50 / 1,907 papers shown
CDT: A Comprehensive Capability Framework for Large Language Models Across Cognition, Domain, and Task
Haosi Mo
Xinyu Ma
Xuebo Liu
Yang Li
Yu Li
Jie Liu
Min Zhang
ELM
119
0
0
29 Sep 2025
From Score Distributions to Balance: Plug-and-Play Mixture-of-Experts Routing
Rana Shahout
Colin Cai
Yilun Du
Minlan Yu
Michael Mitzenmacher
MoE
MoMe
158
3
0
29 Sep 2025
Pretraining with hierarchical memories: separating long-tail and common knowledge
Hadi Pouransari
David Grangier
C Thomas
Michael Kirchhof
Oncel Tuzel
RALM
KELM
240
1
0
29 Sep 2025
LLM DNA: Tracing Model Evolution via Functional Representations
Zhaomin Wu
Haodong Zhao
Ziyang Wang
Jizhou Guo
Qian Wang
Bingsheng He
124
2
0
29 Sep 2025
Query Circuits: Explaining How Language Models Answer User Prompts
Tung-Yu Wu
Fazl Barez
ReLM
LRM
154
0
0
29 Sep 2025
Reasoning or Retrieval? A Study of Answer Attribution on Large Reasoning Models
Yuhui Wang
Changjiang Li
Guangke Chen
Jiacheng Liang
Ting Wang
ReLM
KELM
LRM
141
1
0
29 Sep 2025
Conda: Column-Normalized Adam for Training Large Language Models Faster
Junjie Wang
Pan Zhou
Yiming Dong
Huan Li
Jia Li
Xun Zhou
Qicheng Lao
Cong Fang
Zhouchen Lin
AI4CE
236
0
0
29 Sep 2025
ChunkLLM: A Lightweight Pluggable Framework for Accelerating LLMs Inference
Haojie Ouyang
Jianwei Lv
Lei Ren
Chen Wei
Xiaojie Wang
Fangxiang Feng
VLM
164
0
0
28 Sep 2025
Tequila: Trapping-free Ternary Quantization for Large Language Models
Hong Huang
Decheng Wu
Rui Cen
Guanghua Yu
Z. Li
Kai Liu
Jianchen Zhu
Peng Chen
Xue Liu
Dapeng Wu
MQ
237
2
0
28 Sep 2025
Sequential Diffusion Language Models
Yangzhou Liu
Yue Cao
Hao-Wen Li
Gen Luo
Z. Chen
...
Yuqiang Li
Tong Lu
Yu Qiao
Jifeng Dai
Wenhai Wang
104
5
0
28 Sep 2025
Beyond Benchmarks: Understanding Mixture-of-Experts Models through Internal Mechanisms
Jiahao Ying
Mingbao Lin
Qianru Sun
Yixin Cao
MoE
55
0
0
28 Sep 2025
Winning the Pruning Gamble: A Unified Approach to Joint Sample and Token Pruning for Efficient Supervised Fine-Tuning
Shaobo Wang
Jiaming Wang
Jiajun Zhang
C. Wang
Yue Min
...
Fei Huang
Huiqiang Jiang
Junyang Lin
Dayiheng Liu
Linfeng Zhang
147
5
0
28 Sep 2025
Bridging the Knowledge-Prediction Gap in LLMs on Multiple-Choice Questions
Yoonah Park
Haesung Pyun
Yohan Jo
KELM
368
0
0
28 Sep 2025
Don't Settle Too Early: Self-Reflective Remasking for Diffusion Language Models
Zemin Huang
Yuhang Wang
Zhiyang Chen
Guo-Jun Qi
97
4
0
28 Sep 2025
A2D: Any-Order, Any-Step Safety Alignment for Diffusion Language Models
Wonje Jeung
Sangyeon Yoon
Yoonjun Cho
Dongjae Jeon
Sangwoo Shin
Hyesoo Hong
Albert No
DiffM
138
0
0
27 Sep 2025
MoE-PHDS: One MoE checkpoint for flexible runtime sparsity
Lauren Hannah
Soheil Zibakhsh
K. Nishu
Arnav Kundu
Mohammad Samragh Razlighi
Mehrdad Farajtabar
Minsik Cho
MoE
96
0
0
27 Sep 2025
Text-Based Approaches to Item Difficulty Modeling in Large-Scale Assessments: A Systematic Review
Sydney Peters
Nan Zhang
Hong Jiao
Ming Li
Tianyi Zhou
Robert Lissitz
124
3
0
27 Sep 2025
Bridging the Gap Between Promise and Performance for Microscaling FP4 Quantization
Vage Egiazarian
Roberto L. Castro
Denis Kuznedelev
Andrei Panferov
Eldar Kurtic
...
Alexandre Marques
Mark Kurtz
Saleh Ashkboos
Torsten Hoefler
Dan Alistarh
MQ
228
1
0
27 Sep 2025
Quant-dLLM: Post-Training Extreme Low-Bit Quantization for Diffusion Large Language Models
Tianao Zhang
Zhiteng Li
Xianglong Yan
Haotong Qin
Yong Guo
Yulun Zhang
MQ
121
0
0
27 Sep 2025
Multiplayer Nash Preference Optimization
Fang Wu
X. Y. Huang
Weihao Xuan
Zhiwei Zhang
Yijia Xiao
...
Xiaomin Li
Bing Hu
Peng Xia
Jure Leskovec
Yejin Choi
137
2
0
27 Sep 2025
Train Once, Answer All: Many Pretraining Experiments for the Cost of One
Sebastian Bordt
Martin Pawelczyk
CLL
175
1
0
27 Sep 2025
SDQ-LLM: Sigma-Delta Quantization for 1-bit LLMs of any size
Junhao Xia
Ming Zhao
Limin Xiao
Xiujun Zhang
MQ
107
0
0
27 Sep 2025
PT
2
^2
2
-LLM: Post-Training Ternarization for Large Language Models
Xianglong Yan
Chengzhu Bao
Zhiteng Li
Tianao Zhang
Kaicheng Yang
Haotong Qin
Ruobing Xie
Xingwu Sun
Yulun Zhang
MQ
207
0
0
27 Sep 2025
Beyond Outliers: A Study of Optimizers Under Quantization
Georgios Vlassis
Saleh Ashkboos
Alexandra Volkova
Torsten Hoefler
Dan Alistarh
MQ
206
0
0
27 Sep 2025
DOoM: Difficult Olympiads of Math
Ilya Kuleshov
Ilin Pavel
Nikolay Kompanets
Ksenia Sycheva
Aleksandr Nikolich
AIMat
254
0
0
27 Sep 2025
Thinking in Many Modes: How Composite Reasoning Elevates Large Language Model Performance with Limited Data
Zishan Ahmad
Saisubramaniam Gopalakrishnan
LRM
90
0
0
26 Sep 2025
What Matters More For In-Context Learning under Matched Compute Budgets: Pretraining on Natural Text or Incorporating Targeted Synthetic Examples?
Mohammed Sabry
Anya Belz
98
0
0
26 Sep 2025
Stochastic activations
Maria Lomeli
Matthijs Douze
Gergely Szilvasy
Loic Cabannes
Jade Copet
Sainbayar Sukhbaatar
Jason Weston
Gabriel Synnaeve
Pierre-Emmanuel Mazaré
Hervé Jégou
LLMSV
268
0
0
26 Sep 2025
Tiny-QMoE
Jack Cashman
Jiaqi Nie
24
0
0
26 Sep 2025
Elastic MoE: Unlocking the Inference-Time Scalability of Mixture-of-Experts
Naibin Gu
Zhenyu Zhang
Yuchen Feng
Yilong Chen
Peng Fu
...
Shuohuan Wang
Yu Sun
Hua Wu
Weiping Wang
Haifeng Wang
MoE
85
0
0
26 Sep 2025
Erase or Hide? Suppressing Spurious Unlearning Neurons for Robust Unlearning
Nakyeong Yang
Dong-Kyum Kim
Jea Kwon
Minsung Kim
Kyomin Jung
M. Cha
MU
KELM
112
0
0
26 Sep 2025
COSPADI: Compressing LLMs via Calibration-Guided Sparse Dictionary Learning
Dmitriy Shopkhoev
Denis Makhov
Magauiya Zhussip
Ammar Ali
Stamatios Lefkimmiatis
186
0
0
26 Sep 2025
MindCraft: How Concept Trees Take Shape In Deep Models
Bowei Tian
Yexiao He
Wanghao Ye
Ziyao Wang
Meng Liu
Ang Li
LRM
102
0
0
26 Sep 2025
Towards Generalizable Implicit In-Context Learning with Attention Routing
Jiaqian Li
Yanshu Li
Ligong Han
Ruixiang Tang
Wenya Wang
141
0
0
26 Sep 2025
Context Parametrization with Compositional Adapters
Josip Jukić
Martin Tutek
Jan Snajder
123
0
0
26 Sep 2025
Meta-Awareness Enhances Reasoning Models: Self-Alignment Reinforcement Learning
Yoonjeon Kim
Doohyuk Jang
Eunho Yang
ReLM
AIFin
LRM
199
1
0
26 Sep 2025
Lightweight error mitigation strategies for post-training N:M activation sparsity in LLMs
Shirin Alanova
Kristina Kazistova
Ekaterina Galaeva
Alina Kostromina
Vladimir Smirnov
Redko Dmitry
Alexey Dontsov
Maxim Zhelnin
Evgeny Burnaev
Egor Shvetsov
137
0
0
26 Sep 2025
SBFA: Single Sneaky Bit Flip Attack to Break Large Language Models
Jingkai Guo
C. Chakrabarti
Deliang Fan
AAML
65
3
0
26 Sep 2025
IIET: Efficient Numerical Transformer via Implicit Iterative Euler Method
Xinyu Liu
Bei Li
Jiahao Liu
Junhao Ruan
Kechen Jiao
Hongyin Tang
Jingang Wang
Xiao Tong
Jingbo Zhu
179
0
0
26 Sep 2025
Rethinking RoPE Scaling in Quantized LLM: Theory, Outlier, and Channel-Band Analysis with Weight Rescaling
Ye Qiao
Haocheng Xu
Xiaofan Zhang
Sitao Huang
MQ
112
0
0
26 Sep 2025
Front-Loading Reasoning: The Synergy between Pretraining and Post-Training Data
Syeda Nahida Akter
Shrimai Prabhumoye
Eric Nyberg
M. Patwary
Mohammad Shoeybi
Yejin Choi
Bryan Catanzaro
AIFin
LRM
AI4CE
120
6
0
26 Sep 2025
JGU Mainz's Submission to the WMT25 Shared Task on LLMs with Limited Resources for Slavic Languages: MT and QA
Hossain Shaikh Saadi
Minh Duc Bui
Mario Sanz-Guerrero
Katharina von der Wense
83
1
0
26 Sep 2025
Blockwise Hadamard high-Rank Adaptation for Parameter-Efficient LLM Fine-Tuning
Feng Yu
Jia Hu
Geyong Min
153
0
0
25 Sep 2025
Predicting LLM Reasoning Performance with Small Proxy Model
Woosung Koh
Juyoung Suk
Sungjun Han
Se-Young Yun
Jay Shin
LRM
AI4CE
270
0
0
25 Sep 2025
Painless Activation Steering: An Automated, Lightweight Approach for Post-Training Large Language Models
Sasha Cui
Zhongren Chen
LLMSV
233
1
0
25 Sep 2025
Expanding Reasoning Potential in Foundation Model by Learning Diverse Chains of Thought Patterns
Xuemiao Zhang
Can Ren
Chengying Tu
Rongxiang Weng
Shuo Wang
Hongfei Yan
Jingang Wang
Xunliang Cai
LRM
AI4CE
210
1
0
25 Sep 2025
Mixture of Thoughts: Learning to Aggregate What Experts Think, Not Just What They Say
Jacob Fein-Ashley
Dhruv Parikh
Rajgopal Kannan
Viktor Prasanna
MoE
MoMe
LRM
175
1
0
25 Sep 2025
On Code-Induced Reasoning in LLMs
Abdul Waheed
Zhen Wu
Carolyn Rose
Daphne Ippolito
LRM
157
0
0
25 Sep 2025
SFT Doesn't Always Hurt General Capabilities: Revisiting Domain-Specific Fine-Tuning in LLMs
J. Lin
Zhongruo Wang
Kun Qian
Tian Wang
Arvind Srinivasan
...
Weiqi Zhang
Sujay Sanghavi
C. L. P. Chen
Hyokun Yun
Lihong Li
CLL
350
1
0
25 Sep 2025
Detoxifying Large Language Models via Autoregressive Reward Guided Representation Editing
Yisong Xiao
Aishan Liu
Siyuan Liang
Zonghao Ying
Xianglong Liu
Dacheng Tao
KELM
152
2
0
24 Sep 2025
Previous
1
2
3
...
6
7
8
...
37
38
39
Next