Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
1803.05457
Cited By
Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge
14 March 2018
Peter Clark
Isaac Cowhey
Oren Etzioni
Tushar Khot
Ashish Sabharwal
Carissa Schoenick
Oyvind Tafjord
ELM
RALM
LRM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge"
50 / 1,910 papers shown
RSAVQ: Riemannian Sensitivity-Aware Vector Quantization for Large Language Models
Zukang Xu
Yan Chen
Qiang Wu
Dawei Yang
MQ
235
0
0
24 Sep 2025
Enhancing Linear Attention with Residual Learning
Xunhao Lai
Jialiang Kang
Jianqiao Lu
Tong Lin
Pengyu Zhao
KELM
CLL
118
0
0
24 Sep 2025
Q-Palette: Fractional-Bit Quantizers Toward Optimal Bit Allocation for Efficient LLM Deployment
Deokjae Lee
Hyun Oh Song
MQ
208
0
0
24 Sep 2025
Soft Tokens, Hard Truths
Natasha Butt
Ariel Kwiatkowski
Ismail Labiad
Julia Kempe
Yann Ollivier
OffRL
CLL
LRM
165
1
0
23 Sep 2025
Prior-based Noisy Text Data Filtering: Fast and Strong Alternative For Perplexity
Yeongbin Seo
Gayoung Kim
Jaehyung Kim
Jinyoung Yeo
150
0
0
23 Sep 2025
HyperAdapt: Simple High-Rank Adaptation
Abel Gurung
Joseph Campbell
167
0
0
23 Sep 2025
CCQA: Generating Question from Solution Can Improve Inference-Time Reasoning in SLMs
Jin Young Kim
Ji Won Yoon
ReLM
LRM
157
0
0
23 Sep 2025
On-the-Fly Adaptation to Quantization: Configuration-Aware LoRA for Efficient Fine-Tuning of Quantized LLMs
Rongguang Ye
Ming Tang
Edith C. H. Ngai
MQ
96
0
0
22 Sep 2025
QWHA: Quantization-Aware Walsh-Hadamard Adaptation for Parameter-Efficient Fine-Tuning on Large Language Models
Hyesung Jeon
Seojune Lee
Beomseok Kang
Yulhwa Kim
Jae-Joon Kim
MQ
291
0
0
22 Sep 2025
TASO: Task-Aligned Sparse Optimization for Parameter-Efficient Model Adaptation
Daiye Miao
Yufang Liu
Jie Wang
Changzhi Sun
Yunke Zhang
Demei Yan
Shaokang Dong
Qi Zhang
Man Lan
88
1
0
22 Sep 2025
Diagnosing Model Editing via Knowledge Spectrum
Tsung-Hsuan Pan
Chung-Chi Chen
Hen-Hsen Huang
Hsin-Hsi Chen
KELM
117
0
0
22 Sep 2025
Training-free Truthfulness Detection via Value Vectors in LLMs
Runheng Liu
Heyan Huang
Xingchen Xiao
Zhijing Wu
93
0
0
22 Sep 2025
seqBench: A Tunable Benchmark to Quantify Sequential Reasoning Limits of LLMs
Mohammad Ramezanali
Mo Vazifeh
Paolo Santi
LRM
ELM
91
0
0
21 Sep 2025
Dynamic Expert Specialization: Towards Catastrophic Forgetting-Free Multi-Domain MoE Adaptation
Junzhuo Li
Bo Wang
Xiuze Zhou
Xuming Hu
MoMe
CLL
MoE
205
2
0
21 Sep 2025
MoEs Are Stronger than You Think: Hyper-Parallel Inference Scaling with RoE
Soheil Zibakhsh
Mohammad Samragh
K. Nishu
Lauren Hannah
Arnav Kundu
Minsik Cho
MoE
BDL
LRM
267
0
0
21 Sep 2025
PTQTP: Post-Training Quantization to Trit-Planes for Large Language Models
He Xiao
Runming Yang
Qingyao Yang
Wendong Xu
Zheng Li
Yupeng Su
Zhengwu Liu
Hongxia Yang
Ngai Wong
MQ
143
2
0
21 Sep 2025
EG-MLA: Embedding-Gated Multi-head Latent Attention for Scalable and Efficient LLMs
Zhengge Cai
Haowen Hou
70
0
0
20 Sep 2025
Rethinking the Role of Text Complexity in Language Model Pretraining
Dan John Velasco
M. R
215
2
0
20 Sep 2025
SABER: Uncovering Vulnerabilities in Safety Alignment via Cross-Layer Residual Connection
Maithili Joshi
Palash Nandi
Tanmoy Chakraborty
AAML
LLMSV
105
0
0
19 Sep 2025
Distribution-Aligned Decoding for Efficient LLM Task Adaptation
Senkang Hu
Xudong Han
Jinqi Jiang
Yihang Tao
Zihan Fang
Yong Dai
Sam Kwong
Yuguang Fang
239
2
0
19 Sep 2025
DiEP: Adaptive Mixture-of-Experts Compression through Differentiable Expert Pruning
Sikai Bai
Haoxi Li
Jie Zhang
Zicong Hong
Song Guo
MoE
118
1
0
19 Sep 2025
Concept Unlearning in Large Language Models via Self-Constructed Knowledge Triplets
Tomoya Yamashita
Yuuki Yamanaka
M. Yamada
Takayuki Miura
Toshiki Shibahara
Tomoharu Iwata
MU
98
1
0
19 Sep 2025
Pico: A Modular Framework for Hypothesis-Driven Small Language Model Research
Richard Diehl Martinez
David Demitri Africa
Yuval Weiss
Suchir Salhan
Ryan Daniels
P. Buttery
144
1
0
19 Sep 2025
Exploring Polyglot Harmony: On Multilingual Data Allocation for Large Language Models Pretraining
Ping Guo
Y. Ren
Binbin Liu
Fengze Liu
Haobin Lin
Yifan Zhang
Bingni Zhang
Taifeng Wang
Yin Zheng
155
1
0
19 Sep 2025
Pre-training under infinite compute
Konwoo Kim
Suhas Kotha
Abigail Z. Jacobs
Tatsunori Hashimoto
229
3
0
18 Sep 2025
CARGO: A Framework for Confidence-Aware Routing of Large Language Models
Amine Barrak
Yosr Fourati
Michael Olchawa
Emna Ksontini
Khalil Zoghlami
153
1
0
18 Sep 2025
Fair-GPTQ: Bias-Aware Quantization for Large Language Models
Irina Proskurina
Guillaume Metzler
Julien Velcin
MQ
130
0
0
18 Sep 2025
FURINA: Free from Unmergeable Router via LINear Aggregation of mixed experts
Jiayi Han
Liang Du
Yinda Chen
Xiao Kang
Weiyang Ding
Donghong Han
MoE
MoMe
128
0
0
18 Sep 2025
NIRVANA: Structured pruning reimagined for large language models compression
Mengting Ai
Tianxin Wei
Sirui Chen
Jingrui He
VLM
1.6K
1
0
17 Sep 2025
SBVR: Summation of BitVector Representation for Efficient LLM Quantization
Wonjun Bang
Jongseok Park
Hongseung Yu
Kyungmin Bin
Kyunghan Lee
MQ
152
0
0
17 Sep 2025
Synthetic bootstrapped pretraining
Zitong Yang
Aonan Zhang
Hong Liu
Tatsunori Hashimoto
Emmanuel Candès
Chong-Jun Wang
Ruoming Pang
SyDa
295
0
0
17 Sep 2025
DSFT: Inspiring Diffusion Large Language Models to Comprehend Mathematical and Logical Patterns
Ranfei Chen
Ming Chen
DiffM
AI4CE
81
0
0
17 Sep 2025
SteeringSafety: A Systematic Safety Evaluation Framework of Representation Steering in LLMs
Vincent Siu
Nicholas Crispino
David Park
Nathan W. Henry
Yu Yang
Yang Liu
Kurt Thomas
Chenguang Wang
LLMSV
333
1
0
16 Sep 2025
Preservation of Language Understanding Capabilities in Speech-aware Large Language Models
Marek Kubis
Paweł Skórzewski
Iwona Christop
Mateusz Czyżnikiewicz
Jakub Kubiak
Łukasz Bondaruk
Marcin Lewandowski
AuLLM
ELM
190
0
0
15 Sep 2025
AMQ: Enabling AutoML for Mixed-precision Weight-Only Quantization of Large Language Models
Sangjun Lee
Seung-taek Woo
Jungyu Jin
Changhun Lee
Eunhyeok Park
MQ
116
3
0
15 Sep 2025
NeuroStrike: Neuron-Level Attacks on Aligned LLMs
Lichao Wu
Sasha Behrouzi
Mohamadreza Rostami
Maximilian Thang
S. Picek
A. Sadeghi
AAML
240
1
0
15 Sep 2025
MORABLES: A Benchmark for Assessing Abstract Moral Reasoning in LLMs with Fables
Matteo Marcuzzo
A. Zangari
A. Albarelli
Jose Camacho-Collados
Mohammad Taher Pilehvar
216
3
0
15 Sep 2025
CBP-Tuning: Efficient Local Customization for Black-box Large Language Models
Jiaxuan Zhao
Naibin Gu
Yuchen Feng
Xiyu Liu
Peng Fu
Zheng Lin
Weiping Wang
112
0
0
15 Sep 2025
Fluid Language Model Benchmarking
Valentin Hofmann
David Heineman
Ian H. Magnusson
Kyle Lo
Jesse Dodge
Maarten Sap
Pang Wei Koh
Chun Wang
Hannaneh Hajishirzi
Noah A. Smith
136
7
0
14 Sep 2025
From Parameters to Performance: A Data-Driven Study on LLM Structure and Development
Suqing Wang
Zuchao Li
Luohe Shi
Bo Du
Hai Zhao
Yun Li
Qianren Wang
135
0
0
14 Sep 2025
Optimal Brain Restoration for Joint Quantization and Sparsification of LLMs
Hang Guo
Yawei Li
Luca Benini
MQ
215
0
0
14 Sep 2025
AQUA: Attention via QUery mAgnitudes for Memory and Compute Efficient Inference in LLMs
S. Shah
Saurav Prakash
Balaraman Ravindran
92
0
0
14 Sep 2025
Dropping Experts, Recombining Neurons: Retraining-Free Pruning for Sparse Mixture-of-Experts LLMs
Yixiao Zhou
Ziyu Zhao
Dongzhou Cheng
Zhiliang Wu
Jie Gui
Yi-feng Yang
Fei Wu
Yu Cheng
Hehe Fan
MoMe
MoE
164
5
0
12 Sep 2025
Test-Time Warmup for Multimodal Large Language Models
Nikita Rajaneesh
Thomas P. Zollo
R. Zemel
MLLM
VLM
LRM
209
0
0
12 Sep 2025
Automated MCQA Benchmarking at Scale: Evaluating Reasoning Traces as Retrieval Sources for Domain Adaptation of Small Language Models
Ozan Gokdemir
N. Getty
Robert Underwood
Sandeep Madireddy
Franck Cappello
Arvind Ramanathan
Ian Foster
R. Stevens
ELM
LRM
112
1
0
12 Sep 2025
GrACE: A Generative Approach to Better Confidence Elicitation in Large Language Models
Zhaohan Zhang
Ziquan Liu
Ioannis Patras
152
2
0
11 Sep 2025
TORSO: Template-Oriented Reasoning Towards General Tasks
Minhyuk Kim
Seungyoon Lee
Heuiseok Lim
LRM
189
0
0
11 Sep 2025
ButterflyQuant: Ultra-low-bit LLM Quantization through Learnable Orthogonal Butterfly Transforms
Bingxin Xu
Zhen Dong
Oussama Elachqar
Yuzhang Shang
MQ
192
1
0
11 Sep 2025
Open-sci-ref-0.01: open and reproducible reference baselines for language model and dataset comparison
Marianna Nezhurina
Jörg Franke
Taishi Nakamura
Timur Carstensen
Niccolò Ajroldi
Ville Komulainen
David Salinas
J. Jitsev
178
2
0
10 Sep 2025
Interpretable Physics Reasoning and Performance Taxonomy in Vision-Language Models
Pranav Pawar
Kavish Shah
Akshat Bhalani
Komal Kasat
Dev Mittal
Hadi Gala
Deepali Patil
Nikita Raichada
Monali Deshmukh
ReLM
LRM
80
0
0
10 Sep 2025
Previous
1
2
3
...
7
8
9
...
37
38
39
Next