ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1803.05457
  4. Cited By
Think you have Solved Question Answering? Try ARC, the AI2 Reasoning
  Challenge

Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge

14 March 2018
Peter Clark
Isaac Cowhey
Oren Etzioni
Tushar Khot
Ashish Sabharwal
Carissa Schoenick
Oyvind Tafjord
    ELMRALMLRM
ArXiv (abs)PDFHTML

Papers citing "Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge"

50 / 1,910 papers shown
Pimba: A Processing-in-Memory Acceleration for Post-Transformer Large Language Model Serving
Pimba: A Processing-in-Memory Acceleration for Post-Transformer Large Language Model Serving
Wonung Kim
Yubin Lee
Yoonsung Kim
Jinwoo Hwang
Seongryong Oh
...
Aziz Huseynov
Woong Gyu Park
Chang Hyun Park
Divya Mahajan
Jongse Park
634
3
0
14 Jul 2025
FusionFactory: Fusing LLM Capabilities with Multi-LLM Log Data
FusionFactory: Fusing LLM Capabilities with Multi-LLM Log Data
Tao Feng
Haozhen Zhang
Zijie Lei
Pengrui Han
M. Patwary
Mohammad Shoeybi
Bryan Catanzaro
Jiaxuan You
MoMe
203
0
0
14 Jul 2025
Advancing Large Language Models for Tibetan with Curated Data and Continual Pre-Training
Advancing Large Language Models for Tibetan with Curated Data and Continual Pre-Training
Leiyu Pan
Bojian Xiong
Lei Yang
Renren Jin
Shaowei Zhang
...
Tianyu Dong
Zhuowen Han
Zhuo Chen
Yuqi Ren
Deyi Xiong
CLL
373
0
0
12 Jul 2025
DATE-LM: Benchmarking Data Attribution Evaluation for Large Language Models
DATE-LM: Benchmarking Data Attribution Evaluation for Large Language Models
Cathy Jiao
Yijun Pan
Emily Xiao
Daisy Sheng
Niket Jain
H. C. Zhao
Ishita Dasgupta
Jiaqi W. Ma
Chenyan Xiong
216
0
0
12 Jul 2025
Pre-Training LLMs on a budget: A comparison of three optimizers
Pre-Training LLMs on a budget: A comparison of three optimizers
Joel Schlotthauer
Christian Kroos
Chris Hinze
Viktor Hangya
Luzian Hahn
Fabian Küch
203
0
0
11 Jul 2025
Lizard: An Efficient Linearization Framework for Large Language Models
Lizard: An Efficient Linearization Framework for Large Language Models
Chien Van Nguyen
Ruiyi Zhang
Hanieh Deilamsalehy
Puneet Mathur
Viet Dac Lai
...
Ryan Rossi
Trung H. Bui
N. Vlassis
Franck Dernoncourt
T. Nguyen
KELM
247
2
0
11 Jul 2025
KV Cache Steering for Controlling Frozen LLMs
KV Cache Steering for Controlling Frozen LLMs
Max Belitsky
D. J. Kopiczko
Michael Dorkenwald
M. Jehanzeb Mirza
James R. Glass
Cees G. M. Snoek
Yuki M. Asano
LLMSVLRM
267
0
0
11 Jul 2025
AbbIE: Autoregressive Block-Based Iterative Encoder for Efficient Sequence Modeling
AbbIE: Autoregressive Block-Based Iterative Encoder for Efficient Sequence Modeling
Preslav Aleksandrov
Meghdad Kurmanji
Fernando Garcia Redondo
David O'Shea
William F. Shen
Alex Iacob
Lorenzo Sani
Xinchi Qiu
Nicola Cancedda
Nicholas D. Lane
186
4
0
11 Jul 2025
SAS: Simulated Attention Score
SAS: Simulated Attention Score
Chuanyang Zheng
J. Sun
Yihang Gao
Yuehao Wang
Peihao Wang
...
Atlas Wang
Mac Schwager
Anderson Schneider
Xiaodong Liu
Jianfeng Gao
AI4TS
244
2
0
10 Jul 2025
FlexOlmo: Open Language Models for Flexible Data Use
FlexOlmo: Open Language Models for Flexible Data Use
Weijia Shi
Akshita Bhagia
Kevin Farhat
Niklas Muennighoff
Pete Walsh
...
Luke Zettlemoyer
Pang Wei Koh
Hannaneh Hajishirzi
Ali Farhadi
Sewon Min
MoE
398
4
0
09 Jul 2025
Train-before-Test Harmonizes Language Model Rankings
Train-before-Test Harmonizes Language Model Rankings
Guanhua Zhang
Ricardo Dominguez-Olmedo
Moritz Hardt
ALM
212
2
0
07 Jul 2025
Steering Information Utility in Key-Value Memory for Language Model Post-Training
Steering Information Utility in Key-Value Memory for Language Model Post-Training
Chunyuan Deng
Ruidi Chang
Hanjie Chen
LLMSV
369
0
0
07 Jul 2025
RAT: Bridging RNN Efficiency and Attention Accuracy via Chunk-based Sequence Modeling
RAT: Bridging RNN Efficiency and Attention Accuracy via Chunk-based Sequence ModelingIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2025
Xiuying Wei
Anunay Yadav
Razvan Pascanu
Çağlar Gülçehre
AI4TS
262
0
0
06 Jul 2025
LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization
LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization
Xujia Wang
Yunjia Qi
Bin Xu
262
0
0
06 Jul 2025
OrthoRank: Token Selection via Sink Token Orthogonality for Efficient LLM inference
OrthoRank: Token Selection via Sink Token Orthogonality for Efficient LLM inference
Seungjun Shin
Jaehoon Oh
Dokwan Oh
172
1
0
05 Jul 2025
Blending Supervised and Reinforcement Fine-Tuning with Prefix Sampling
Blending Supervised and Reinforcement Fine-Tuning with Prefix Sampling
Zeyu Huang
Tianhao Cheng
Zihan Qiu
Zili Wang
Yinghui Xu
Edoardo M. Ponti
Ivan Titov
340
16
0
02 Jul 2025
Eka-Eval: An Evaluation Framework for Low-Resource Multilingual Large Language Models
Eka-Eval: An Evaluation Framework for Low-Resource Multilingual Large Language Models
Samridhi Raj Sinha
Rajvee Sheth
Abhishek Upperwal
Mayank Singh
ELM
191
0
0
02 Jul 2025
Tuning without Peeking: Provable Generalization Bounds and Robust LLM Post-Training
Tuning without Peeking: Provable Generalization Bounds and Robust LLM Post-Training
Ismail Labiad
Mathurin Videau
Matthieu Kowalski
Marc Schoenauer
Alessandro Leite
Julia Kempe
O. Teytaud
AAML
291
0
0
02 Jul 2025
Semantic-guided Diverse Decoding for Large Language Model
Semantic-guided Diverse Decoding for Large Language Model
Weijie Shi
Yue Cui
Yaguang Wu
J. Fang
Shibo Zhang
Mengze Li
Sirui Han
Jia Zhu
Jiajie Xu
Xiaofang Zhou
220
0
0
30 Jun 2025
Data Uniformity Improves Training Efficiency and More, with a Convergence Framework Beyond the NTK Regime
Data Uniformity Improves Training Efficiency and More, with a Convergence Framework Beyond the NTK Regime
Yuqing Wang
Shangding Gu
216
0
0
30 Jun 2025
AutoMixer: Checkpoint Artifacts as Automatic Data Mixers
AutoMixer: Checkpoint Artifacts as Automatic Data MixersAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Ernie Chang
Yang Li
Patrick Huber
Vish Vogeti
David Kant
Yangyang Shi
Vikas Chandra
148
3
0
27 Jun 2025
DuoGPT: Training-free Dual Sparsity through Activation-aware Pruning in LLMs
DuoGPT: Training-free Dual Sparsity through Activation-aware Pruning in LLMs
Ruokai Yin
Yuhang Li
Donghyun Lee
Priyadarshini Panda
VLM
246
2
0
25 Jun 2025
Multi-Preference Lambda-weighted Listwise DPO for Small-Scale Model Alignment
Multi-Preference Lambda-weighted Listwise DPO for Small-Scale Model Alignment
Yuhui Sun
Xiyao Wang
Zixi Li
Zhenlong Yuan
Jinman Zhao
210
0
0
24 Jun 2025
AnTKV: Anchor Token-Aware Sub-Bit Vector Quantization for KV Cache in Large Language Models
AnTKV: Anchor Token-Aware Sub-Bit Vector Quantization for KV Cache in Large Language Models
Zeyu Li
Chuanfu Xiao
Yang Wang
Xiang Liu
Zhenheng Tang
Baotong Lu
Mao Yang
Xinyu Chen
Xiaowen Chu
MQ
138
1
0
24 Jun 2025
Revisiting LoRA through the Lens of Parameter Redundancy: Spectral Encoding Helps
Revisiting LoRA through the Lens of Parameter Redundancy: Spectral Encoding HelpsAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Jiashun Cheng
Chenyi Zi
Polydoros Giannouris
Ziqi Gao
Yuhan Li
Jia Li
Fugee Tsung
220
0
0
20 Jun 2025
SparseLoRA: Accelerating LLM Fine-Tuning with Contextual Sparsity
SparseLoRA: Accelerating LLM Fine-Tuning with Contextual Sparsity
Samir Khaki
Xiuyu Li
Junxian Guo
Ligeng Zhu
Chenfeng Xu
Konstantinos N. Plataniotis
Amir Yazdanbakhsh
Kurt Keutzer
Song Han
Zhijian Liu
217
4
0
19 Jun 2025
EvoLM: In Search of Lost Language Model Training Dynamics
EvoLM: In Search of Lost Language Model Training Dynamics
Zhenting Qi
Fan Nie
Alexandre Alahi
James Zou
Himabindu Lakkaraju
Yilun Du
Eric P. Xing
Sham Kakade
Hanlin Zhang
316
2
0
19 Jun 2025
Drag-and-Drop LLMs: Zero-Shot Prompt-to-Weights
Drag-and-Drop LLMs: Zero-Shot Prompt-to-Weights
Zhiyuan Liang
Dongwen Tang
Yuhao Zhou
Xuanlei Zhao
Mingjia Shi
...
Damian Borth
Michael M. Bronstein
Yang You
Zinan Lin
Kai Wang
OffRL
241
3
0
19 Jun 2025
Thunder-Tok: Minimizing Tokens per Word in Tokenizing Korean Texts for Generative Language Models
Thunder-Tok: Minimizing Tokens per Word in Tokenizing Korean Texts for Generative Language Models
Gyeongje Cho
Yeonkyoun So
Chanwoo Park
Sangmin Lee
Sungmok Jung
Jaejin Lee
VLM
224
0
0
18 Jun 2025
Finance Language Model Evaluation (FLaME)
Finance Language Model Evaluation (FLaME)
Glenn Matlin
Mika Okamoto
Huzaifa Pardawala
Yang Yang
Sudheer Chava
AIFinLRM
190
1
0
18 Jun 2025
RATTENTION: Towards the Minimal Sliding Window Size in Local-Global Attention Models
RATTENTION: Towards the Minimal Sliding Window Size in Local-Global Attention Models
Bailin Wang
Chang Lan
Chong-Jun Wang
Ruoming Pang
257
2
0
18 Jun 2025
CC-LEARN: Cohort-based Consistency Learning
CC-LEARN: Cohort-based Consistency Learning
Xiao Ye
Shaswat Shrivastava
Zhaonan Li
Jacob Dineen
Shijie Lu
Avneet Ahuja
Ming shen
Zhikun Xu
Ben Zhou
OffRLLRM
340
2
0
18 Jun 2025
SLR: Automated Synthesis for Scalable Logical Reasoning
SLR: Automated Synthesis for Scalable Logical Reasoning
Lukas Helff
Ahmad Omar
Felix Friedrich
Antonia Wüst
Hikaru Shindo
Tim Woydt
Rupert Mitchell
P. Schramowski
Wolfgang Stammer
Kristian Kersting
LRM
364
0
0
18 Jun 2025
Instruction Tuning with and without Context: Behavioral Shifts and Downstream Impact
Instruction Tuning with and without Context: Behavioral Shifts and Downstream Impact
Hyunji Lee
Seunghyun Yoon
Yunjae Won
Hanseok Oh
Geewook Kim
Trung H. Bui
Franck Dernoncourt
Elias Stengel-Eskin
Mohit Bansal
Minjoon Seo
LRM
249
2
0
18 Jun 2025
SFT-GO: Supervised Fine-Tuning with Group Optimization for Large Language Models
SFT-GO: Supervised Fine-Tuning with Group Optimization for Large Language Models
Gyuhak Kim
Sumiran Thakur
Su Min Park
Wei Wei
Yujia Bao
154
2
0
17 Jun 2025
Improving LoRA with Variational Learning
Improving LoRA with Variational Learning
Bai Cong
Nico Daheim
Yuesong Shen
Rio Yokota
Mohammad Emtiyaz Khan
Thomas Möllenhoff
232
1
0
17 Jun 2025
Massive Supervised Fine-tuning Experiments Reveal How Data, Layer, and Training Factors Shape LLM Alignment Quality
Massive Supervised Fine-tuning Experiments Reveal How Data, Layer, and Training Factors Shape LLM Alignment Quality
Yuto Harada
Yusuke Yamauchi
Yusuke Oda
Yohei Oseki
Yusuke Miyao
Yu Takagi
ALM
262
5
0
17 Jun 2025
MoORE: SVD-based Model MoE-ization for Conflict- and Oblivion-Resistant Multi-Task Adaptation
MoORE: SVD-based Model MoE-ization for Conflict- and Oblivion-Resistant Multi-Task Adaptation
Shen Yuan
Yin Zheng
Taifeng Wang
Binbin Liu
Hongteng Xu
MoMe
388
1
0
17 Jun 2025
ROSAQ: Rotation-based Saliency-Aware Weight Quantization for Efficiently Compressing Large Language Models
ROSAQ: Rotation-based Saliency-Aware Weight Quantization for Efficiently Compressing Large Language Models
Junho Yoon
Geom Lee
Donghyeon Jeon
Inho Kang
Seung-Hoon Na
MQVLM
227
0
0
16 Jun 2025
TensorSLM: Energy-efficient Embedding Compression of Sub-billion Parameter Language Models on Low-end Devices
TensorSLM: Energy-efficient Embedding Compression of Sub-billion Parameter Language Models on Low-end Devices
Mingxue Xu
Y. Xu
Danilo Mandic
187
0
0
16 Jun 2025
Understand the Implication: Learning to Think for Pragmatic Understanding
Understand the Implication: Learning to Think for Pragmatic UnderstandingAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
S. Sravanthi
Kishan Maharaj
Sravani Gunnu
Abhijit Mishra
Pushpak Bhattacharyya
ReLMLRM
204
0
0
16 Jun 2025
EAQuant: Enhancing Post-Training Quantization for MoE Models via Expert-Aware Optimization
EAQuant: Enhancing Post-Training Quantization for MoE Models via Expert-Aware Optimization
Zhongqian Fu
Ning Ding
Kai Han
Xianzhi Yu
Xiaosong Li
Xinghao Chen
Yehui Tang
Yunhe Wang
MQMoE
252
1
0
16 Jun 2025
Load Balancing Mixture of Experts with Similarity Preserving Routers
Load Balancing Mixture of Experts with Similarity Preserving Routers
Nabil Omi
S. Sen
Ali Farhadi
MoE
284
7
0
16 Jun 2025
Just Go Parallel: Improving the Multilingual Capabilities of Large Language Models
Just Go Parallel: Improving the Multilingual Capabilities of Large Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Muhammad Reza Qorib
Junyi Li
Hwee Tou Ng
LRM
251
4
0
16 Jun 2025
BOW: Reinforcement Learning for Bottlenecked Next Word Prediction
BOW: Reinforcement Learning for Bottlenecked Next Word Prediction
Ming shen
Zhikun Xu
Xiao Ye
Jacob Dineen
Ben Zhou
OffRLLRM
227
1
0
16 Jun 2025
Unveiling the Learning Mind of Language Models: A Cognitive Framework and Empirical Study
Unveiling the Learning Mind of Language Models: A Cognitive Framework and Empirical Study
Zhengyu Hu
Jianxun Lian
Zheyuan Xiao
Seraphina Zhang
Tianfu Wang
Nicholas Jing Yuan
Xing Xie
Hui Xiong
ELMLRM
229
3
0
16 Jun 2025
Mixture of Weight-shared Heterogeneous Group Attention Experts for Dynamic Token-wise KV Optimization
Mixture of Weight-shared Heterogeneous Group Attention Experts for Dynamic Token-wise KV Optimization
Guanghui Song
Dongping Liao
Yiren Zhao
Kejiang Ye
Cheng-zhong Xu
X. Gao
MoE
182
0
0
16 Jun 2025
Assessing the Role of Data Quality in Training Bilingual Language Models
Assessing the Role of Data Quality in Training Bilingual Language Models
Skyler Seto
Maartje ter Hoeve
Maureen de Seyssel
David Grangier
162
0
0
15 Jun 2025
GTA: Grouped-head latenT Attention
GTA: Grouped-head latenT Attention
Luoyang Sun
Cheng Deng
Jiwen Jiang
Xinjian Wu
Haifeng Zhang
Lei Chen
Lionel M. Ni
Ning Yang
174
1
0
15 Jun 2025
Infini-gram mini: Exact n-gram Search at the Internet Scale with FM-Index
Infini-gram mini: Exact n-gram Search at the Internet Scale with FM-Index
Hao Xu
Hamish Ivison
Yejin Choi
Noah A. Smith
Hannaneh Hajishirzi
258
2
0
13 Jun 2025
Previous
123...111213...373839
Next
Page 12 of 39
Pageof 39