ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1803.05457
  4. Cited By
Think you have Solved Question Answering? Try ARC, the AI2 Reasoning
  Challenge

Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge

14 March 2018
Peter Clark
Isaac Cowhey
Oren Etzioni
Tushar Khot
Ashish Sabharwal
Carissa Schoenick
Oyvind Tafjord
    ELMRALMLRM
ArXiv (abs)PDFHTML

Papers citing "Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge"

50 / 1,614 papers shown
Title
From Prompts to Power: Measuring the Energy Footprint of LLM Inference
From Prompts to Power: Measuring the Energy Footprint of LLM Inference
Francisco Caravaca
Ángel Cuevas
R. Cuevas
64
0
0
05 Nov 2025
In Good GRACEs: Principled Teacher Selection for Knowledge Distillation
In Good GRACEs: Principled Teacher Selection for Knowledge DistillationIEEE computer architecture letters (CAL), 2025
A. Panigrahi
Bingbin Liu
Sadhika Malladi
Sham Kakade
Surbhi Goel
116
0
0
04 Nov 2025
Zero-shot data citation function classification using transformer-based large language models (LLMs)
Zero-shot data citation function classification using transformer-based large language models (LLMs)
Neil Byers
Ali Zaidi
Valerie Skye
Chris Beecroft
Kjiersten Fagnan
44
0
0
04 Nov 2025
IG-Pruning: Input-Guided Block Pruning for Large Language Models
IG-Pruning: Input-Guided Block Pruning for Large Language Models
Kangyu Qiao
Shaolei Zhang
Yang Feng
VLM
161
0
0
04 Nov 2025
The Ouroboros of Benchmarking: Reasoning Evaluation in an Era of Saturation
The Ouroboros of Benchmarking: Reasoning Evaluation in an Era of Saturation
İbrahim Ethem Deveci
Duygu Ataman
ReLMALMELMLRM
135
0
0
03 Nov 2025
CryptoMoE: Privacy-Preserving and Scalable Mixture of Experts Inference via Balanced Expert Routing
CryptoMoE: Privacy-Preserving and Scalable Mixture of Experts Inference via Balanced Expert RoutingIEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2025
Yifan Zhou
Tianshi Xu
Jue Hong
Ye Wu
Meng Li
MoE
397
0
0
03 Nov 2025
Why Should the Server Do It All?: A Scalable, Versatile, and Model-Agnostic Framework for Server-Light DNN Inference over Massively Distributed Clients via Training-Free Intermediate Feature Compression
Why Should the Server Do It All?: A Scalable, Versatile, and Model-Agnostic Framework for Server-Light DNN Inference over Massively Distributed Clients via Training-Free Intermediate Feature Compression
Mingyu Sung
Suhwan Im
Daeho Bang
Il-Min Kim
Sangseok Yun
Jae-Mo Kang
36
0
0
03 Nov 2025
Optimizing Native Sparse Attention with Latent Attention and Local Global Alternating Strategies
Optimizing Native Sparse Attention with Latent Attention and Local Global Alternating Strategies
Yuxuan Hu
Jianchao Tan
Jiaqi Zhang
Wen Zan
Pingwei Sun
Yifan Lu
Yerui Sun
Yuchen Xie
Xunliang Cai
Jing Zhang
180
0
0
02 Nov 2025
Improving Romanian LLM Pretraining Data using Diversity and Quality Filtering
Improving Romanian LLM Pretraining Data using Diversity and Quality Filtering
Vlad Negoita
Mihai Masala
Traian Rebedea
62
0
0
02 Nov 2025
FlashEVA: Accelerating LLM inference via Efficient Attention
FlashEVA: Accelerating LLM inference via Efficient Attention
Juan Gabriel Kostelec
Qinghai Guo
85
0
0
01 Nov 2025
OmniEduBench: A Comprehensive Chinese Benchmark for Evaluating Large Language Models in Education
OmniEduBench: A Comprehensive Chinese Benchmark for Evaluating Large Language Models in Education
Min Zhang
Hao Chen
Hao Chen
Wenqi Zhang
Didi Zhu
Xin Lin
Bo Jiang
Aimin Zhou
Fei Wu
Kun Kuang
ELM
116
0
0
30 Oct 2025
MossNet: Mixture of State-Space Experts is a Multi-Head Attention
MossNet: Mixture of State-Space Experts is a Multi-Head Attention
Shikhar Tuli
James Smith
Haris Jeelani
Chi-Heng Lin
Abhishek Patel
Vasili Ramanishka
Yen-Chang Hsu
Hongxia Jin
MoE
235
0
0
30 Oct 2025
EdgeRunner 20B: Military Task Parity with GPT-5 while Running on the Edge
EdgeRunner 20B: Military Task Parity with GPT-5 while Running on the Edge
Jack FitzGerald
Aristotelis Lazaridis
Dylan Bates
Aman Sharma
Jonnathan Castillo
...
Dave Anderson
Jonathan Beck
Jamie Cuticello
Colton Malkerson
Tyler Saltsman
ELM
254
0
0
30 Oct 2025
From Amateur to Master: Infusing Knowledge into LLMs via Automated Curriculum Learning
From Amateur to Master: Infusing Knowledge into LLMs via Automated Curriculum Learning
Nishit Neema
Srinjoy Mukherjee
Sapan Shah
Gokul Ramakrishnan
Ganesh Venkatesh
CLL
204
0
0
30 Oct 2025
Kimi Linear: An Expressive, Efficient Attention Architecture
Kimi Linear: An Expressive, Efficient Attention Architecture
Kimi Team
Yu Zhang
Zongyu Lin
Xingcheng Yao
J. Hu
...
Guokun Lai
Yuxin Wu
Xinyu Zhou
Zhilin Yang
Yulun Du
88
2
0
30 Oct 2025
Angular Steering: Behavior Control via Rotation in Activation Space
Angular Steering: Behavior Control via Rotation in Activation Space
Hieu M. Vu
T. Nguyen
LLMSV
252
3
0
30 Oct 2025
NeuronMM: High-Performance Matrix Multiplication for LLM Inference on AWS Trainium
NeuronMM: High-Performance Matrix Multiplication for LLM Inference on AWS Trainium
Dinghong Song
Jierui Xu
Weichu Yang
Pengfei Su
Dong Li
78
0
0
29 Oct 2025
A Survey on Unlearning in Large Language Models
A Survey on Unlearning in Large Language Models
Ruichen Qiu
Jiajun Tan
Jiayue Pu
Honglin Wang
Xiao-Shan Gao
Fei Sun
MUAILawPILM
522
0
0
29 Oct 2025
Information-Theoretic Discrete Diffusion
Information-Theoretic Discrete Diffusion
Moongyu Jeon
Sangwoo Shin
Dongjae Jeon
Albert No
DiffMFedML
127
0
0
28 Oct 2025
MISA: Memory-Efficient LLMs Optimization with Module-wise Importance Sampling
MISA: Memory-Efficient LLMs Optimization with Module-wise Importance Sampling
Yuxi Liu
Renjia Deng
Yutong He
Xue Wang
Tao Yao
Kun Yuan
82
0
0
28 Oct 2025
From Cross-Task Examples to In-Task Prompts: A Graph-Based Pseudo-Labeling Framework for In-context Learning
From Cross-Task Examples to In-Task Prompts: A Graph-Based Pseudo-Labeling Framework for In-context Learning
Zihan Chen
Song Wang
Xingbo Fu
Chengshuai Shi
Zhenyu Lei
Cong Shen
Jundong Li
88
1
0
28 Oct 2025
ChessQA: Evaluating Large Language Models for Chess Understanding
ChessQA: Evaluating Large Language Models for Chess Understanding
Qianfeng Wen
Zhenwei Tang
Ashton Anderson
ELMLRM
149
0
0
28 Oct 2025
Charting the European LLM Benchmarking Landscape: A New Taxonomy and a Set of Best Practices
Charting the European LLM Benchmarking Landscape: A New Taxonomy and a Set of Best Practices
Špela Vintar
Taja Kuzman Pungeršek
Mojca Brglez
Nikola Ljubešić
139
0
0
28 Oct 2025
Calibrating and Rotating: A Unified Framework for Weight Conditioning in PEFT
Calibrating and Rotating: A Unified Framework for Weight Conditioning in PEFT
Da Chang
Peng Xue
Yu Li
Yongxiang Liu
P. Xu
Shixun Zhang
136
0
0
28 Oct 2025
Optimizing Retrieval for RAG via Reinforced Contrastive Learning
Optimizing Retrieval for RAG via Reinforced Contrastive Learning
Jiawei Zhou
Lei Chen
103
1
0
28 Oct 2025
FALQON: Accelerating LoRA Fine-tuning with Low-Bit Floating-Point Arithmetic
FALQON: Accelerating LoRA Fine-tuning with Low-Bit Floating-Point Arithmetic
Kanghyun Choi
Hyeyoon Lee
S. Park
Dain Kwon
Jinho Lee
MQ
112
0
0
28 Oct 2025
A Survey on LLM Mid-Training
A Survey on LLM Mid-Training
Chengying Tu
Xuemiao Zhang
Rongxiang Weng
Rumei Li
Chen Zhang
Yang Bai
Hongfei Yan
Jingang Wang
Xunliang Cai
OffRLLRM
181
0
0
27 Oct 2025
Multi-Agent Evolve: LLM Self-Improve through Co-evolution
Multi-Agent Evolve: LLM Self-Improve through Co-evolution
Yixing Chen
Yiding Wang
Siqi Zhu
Haofei Yu
Tao Feng
Muhan Zhang
M. Patwary
Jiaxuan You
LLMAGLRM
251
4
0
27 Oct 2025
Offline Preference Optimization via Maximum Marginal Likelihood Estimation
Offline Preference Optimization via Maximum Marginal Likelihood Estimation
Saeed Najafi
Alona Fyshe
OffRL
112
0
0
27 Oct 2025
RaCoT: Plug-and-Play Contrastive Example Generation Mechanism for Enhanced LLM Reasoning Reliability
RaCoT: Plug-and-Play Contrastive Example Generation Mechanism for Enhanced LLM Reasoning Reliability
Kaitong Cai
Jusheng Zhang
Yijia Fan
Jing Yang
Keze Wang
LRM
96
1
0
26 Oct 2025
SeeDNorm: Self-Rescaled Dynamic Normalization
SeeDNorm: Self-Rescaled Dynamic Normalization
Wenrui Cai
Defa Zhu
Qingjie Liu
Qiyang Min
96
0
0
26 Oct 2025
Frustratingly Easy Task-aware Pruning for Large Language Models
Frustratingly Easy Task-aware Pruning for Large Language Models
Yuanhe Tian
Junjie Liu
Xican Yang
Haishan Ye
Yan Song
85
0
0
26 Oct 2025
TELL-TALE: Task Efficient LLMs with Task Aware Layer Elimination
TELL-TALE: Task Efficient LLMs with Task Aware Layer Elimination
Omar Naim
Krish Sharma
Nicholas M. Asher
64
0
0
26 Oct 2025
The Structural Scalpel: Automated Contiguous Layer Pruning for Large Language Models
The Structural Scalpel: Automated Contiguous Layer Pruning for Large Language Models
Yao Lu
Yuqi Li
Wenbin Xie
Shanqing Yu
Qi Xuan
Zhaowei Zhu
Shiping Wen
52
1
0
25 Oct 2025
Plan Then Retrieve: Reinforcement Learning-Guided Complex Reasoning over Knowledge Graphs
Plan Then Retrieve: Reinforcement Learning-Guided Complex Reasoning over Knowledge Graphs
Yanlin Song
Ben Liu
Víctor Gutiérrez-Basulto
Zhiwei Hu
Qianqian Xie
Min Peng
Sophia Ananiadou
Jeff Z. Pan
RALMReLMLRM
223
0
0
23 Oct 2025
Context-level Language Modeling by Learning Predictive Context Embeddings
Context-level Language Modeling by Learning Predictive Context Embeddings
Beiya Dai
Y. Liu
Daozheng Xue
Qipeng Guo
Kai Chen
Xinbing Wang
Bowen Zhou
Zhouhan Lin
LRM
115
0
0
23 Oct 2025
What Does It Take to Build a Performant Selective Classifier?
What Does It Take to Build a Performant Selective Classifier?
Stephan Rabanser
Nicolas Papernot
162
0
0
23 Oct 2025
ELUTQ: Efficient LUT-Aware Quantization for Deploying Large Language Models on Edge Devices
ELUTQ: Efficient LUT-Aware Quantization for Deploying Large Language Models on Edge Devices
Xin Nie
Liang Dong
H. Zhang
JiaWang Xiao
G. Sun
MQ
312
0
0
22 Oct 2025
Latent Space Factorization in LoRA
Latent Space Factorization in LoRA
Shashi Kumar
Yacouba Kaloga
John Mitros
P. Motlícek
Ina Kodrasi
60
0
0
22 Oct 2025
GaLLoP: Gradient-based Sparse Learning on Low-Magnitude Parameters
GaLLoP: Gradient-based Sparse Learning on Low-Magnitude Parameters
Anand Choudhary
Yasser Sulaıman
Lukas Mauch
G. B. Hacene
Fabien Cardinaux
Antoine Bosselut
100
0
0
22 Oct 2025
Zhyper: Factorized Hypernetworks for Conditioned LLM Fine-Tuning
Zhyper: Factorized Hypernetworks for Conditioned LLM Fine-Tuning
M. H. I. Abdalla
Zhipin Wang
Christian M. M. Frey
Steffen Eger
Josif Grabocka
107
0
0
22 Oct 2025
Loopholing Discrete Diffusion: Deterministic Bypass of the Sampling Wall
Loopholing Discrete Diffusion: Deterministic Bypass of the Sampling Wall
Mingyu Jo
Jaesik Yoon
Justin Deschenaux
Çağlar Gülçehre
Sungjin Ahn
DiffM
156
0
0
22 Oct 2025
ARA: Adaptive Rank Allocation for Efficient Large Language Model SVD Compression
ARA: Adaptive Rank Allocation for Efficient Large Language Model SVD Compression
Lin Xv
Jingsheng Gao
Xian Gao
Ting Liu
Yuzhuo Fu
64
0
0
22 Oct 2025
Data-Centric Lessons To Improve Speech-Language Pretraining
Data-Centric Lessons To Improve Speech-Language Pretraining
Vishaal Udandarao
Zhiyun Lu
Xuankai Chang
Yongqiang Wang
Violet Z. Yao
Albin Madapally Jose
Fartash Faghri
Josh Gardner
Chung-Cheng Chiu
108
0
0
22 Oct 2025
Energy-Efficient and Dequantization-Free Q-LLMs: A Spiking Neural Network Approach to Salient Value Mitigation
Energy-Efficient and Dequantization-Free Q-LLMs: A Spiking Neural Network Approach to Salient Value Mitigation
Chenyu Wang
Zhanglu Yan
Zhi Zhou
Xu Chen
Weng-Fai Wong
MQ
116
0
0
22 Oct 2025
Scaling Laws Meet Model Architecture: Toward Inference-Efficient LLMs
Scaling Laws Meet Model Architecture: Toward Inference-Efficient LLMs
S. Bian
Tao Yu
Shivaram Venkataraman
Youngsuk Park
62
0
0
21 Oct 2025
Pay Attention to the Triggers: Constructing Backdoors That Survive Distillation
Pay Attention to the Triggers: Constructing Backdoors That Survive Distillation
Giovanni De Muri
Mark Vero
Robin Staab
Martin Vechev
107
0
0
21 Oct 2025
ScaleNet: Scaling up Pretrained Neural Networks with Incremental Parameters
ScaleNet: Scaling up Pretrained Neural Networks with Incremental Parameters
Zhiwei Hao
Jianyuan Guo
Li Shen
Kai Han
Yehui Tang
Han Hu
Yunhe Wang
147
0
0
21 Oct 2025
Binary Quadratic Quantization: Beyond First-Order Quantization for Real-Valued Matrix Compression
Binary Quadratic Quantization: Beyond First-Order Quantization for Real-Valued Matrix Compression
Kyo Kuroki
Yasuyuki Okoshi
Thiem Van Chu
Kazushi Kawamura
Masato Motomura
MQ
152
0
0
21 Oct 2025
ssToken: Self-modulated and Semantic-aware Token Selection for LLM Fine-tuning
ssToken: Self-modulated and Semantic-aware Token Selection for LLM Fine-tuning
Xiaohan Qin
Xiaoxing Wang
Ning Liao
Cancheng Zhang
Xiangdong Zhang
Mingquan Feng
Jingzhi Wang
Junchi Yan
110
0
0
21 Oct 2025
Previous
12345...313233
Next