ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.04210
  4. Cited By
Does Thinking More always Help? Mirage of Test-Time Scaling in Reasoning Models
v1v2v3 (latest)

Does Thinking More always Help? Mirage of Test-Time Scaling in Reasoning Models

4 June 2025
Soumya Suvra Ghosal
Souradip Chakraborty
Avinash Reddy
Yifu Lu
Mengdi Wang
Dinesh Manocha
Furong Huang
Mohammad Ghavamzadeh
Amrit Singh Bedi
    ReLMLRM
ArXiv (abs)PDFHTML

Papers citing "Does Thinking More always Help? Mirage of Test-Time Scaling in Reasoning Models"

16 / 16 papers shown
Title
The Zero-Step Thinking: An Empirical Study of Mode Selection as Harder Early Exit in Reasoning Models
The Zero-Step Thinking: An Empirical Study of Mode Selection as Harder Early Exit in Reasoning Models
Yuqiao Tan
Shizhu He
Kang Liu
Jun Zhao
LRM
82
1
0
22 Oct 2025
MoReBench: Evaluating Procedural and Pluralistic Moral Reasoning in Language Models, More than Outcomes
MoReBench: Evaluating Procedural and Pluralistic Moral Reasoning in Language Models, More than Outcomes
Yu Ying Chiu
Michael S. Lee
Rachel Calcott
Brandon Handoko
Paul de Font-Reaulx
...
Mantas Mazeika
Bing Liu
Yejin Choi
Mitchell L. Gordon
Sydney Levine
ELMLRM
121
0
0
18 Oct 2025
ThinkPilot: Steering Reasoning Models via Automated Think-prefixes Optimization
ThinkPilot: Steering Reasoning Models via Automated Think-prefixes Optimization
Sunzhu Li
Zhiyu Lin
Shuling Yang
Jiale Zhao
Wei Chen
LRM
85
0
0
14 Oct 2025
Learning What's Missing: Attention Dispersion and EMA Stabilization in Length Generalization
Learning What's Missing: Attention Dispersion and EMA Stabilization in Length Generalization
Pál Zsámboki
Benjamin Levi
David Ansel Josef Smith
Mitansh Kagalwala
Arlington Kell
Samuel Liechty
Cong Wang
64
0
0
09 Oct 2025
DRPO: Efficient Reasoning via Decoupled Reward Policy Optimization
DRPO: Efficient Reasoning via Decoupled Reward Policy Optimization
Gang Li
Yan Chen
Ming Lin
Tianbao Yang
OffRLLRM
121
0
1
06 Oct 2025
GuidedSampling: Steering LLMs Towards Diverse Candidate Solutions at Inference-Time
GuidedSampling: Steering LLMs Towards Diverse Candidate Solutions at Inference-Time
Divij Handa
Mihir Parmar
Aswin Rrv
Md Nayem Uddin
Hamid Palangi
Chitta Baral
87
0
0
04 Oct 2025
Training Large Language Models To Reason In Parallel With Global Forking Tokens
Training Large Language Models To Reason In Parallel With Global Forking Tokens
Sheng Jia
Xiao Wang
Shiva Prasad Kasiviswanathan
LRM
125
1
0
01 Oct 2025
Rethinking Thinking Tokens: LLMs as Improvement Operators
Rethinking Thinking Tokens: LLMs as Improvement Operators
Lovish Madaan
Aniket Didolkar
Suchin Gururangan
John Quan
Ruan Silva
Ruslan Salakhutdinov
Manzil Zaheer
Sanjeev Arora
Anirudh Goyal
ReLMLRM
178
1
1
01 Oct 2025
Overthinking Reduction with Decoupled Rewards and Curriculum Data Scheduling
Overthinking Reduction with Decoupled Rewards and Curriculum Data Scheduling
Shuyang Jiang
Yusheng Liao
Ya Zhang
Yanfeng Wang
Y. Wang
OffRLLRM
88
0
0
30 Sep 2025
RADAR: Reasoning-Ability and Difficulty-Aware Routing for Reasoning LLMs
RADAR: Reasoning-Ability and Difficulty-Aware Routing for Reasoning LLMs
Nigel Fernandez
Branislav Kveton
Ryan Rossi
Andrew Lan
Zichao Wang
LRM
202
0
0
29 Sep 2025
What Characterizes Effective Reasoning? Revisiting Length, Review, and Structure of CoT
What Characterizes Effective Reasoning? Revisiting Length, Review, and Structure of CoT
Yunzhen Feng
Julia Kempe
Cheng Zhang
Parag Jain
Anthony Hartshorn
LRM
106
4
0
23 Sep 2025
CoreThink: A Symbolic Reasoning Layer to reason over Long Horizon Tasks with LLMs
CoreThink: A Symbolic Reasoning Layer to reason over Long Horizon Tasks with LLMs
Jay Vaghasiya
Omkar Ghugarkar
Vishvesh Bhat
Vipul Dholaria
Julian McAuley
LLMAGReLMLRM
198
1
0
31 Aug 2025
ParaThinker: Native Parallel Thinking as a New Paradigm to Scale LLM Test-time Compute
ParaThinker: Native Parallel Thinking as a New Paradigm to Scale LLM Test-time Compute
Hao Wen
Yifan Su
Feifei Zhang
Yunxin Liu
Yunhao Liu
Y. Zhang
Yuanchun Li
ReLMLRM
150
10
0
30 Aug 2025
PiCSAR: Probabilistic Confidence Selection And Ranking
PiCSAR: Probabilistic Confidence Selection And Ranking
Joshua Ong Jun Leang
Zheng Zhao
Aryo Pradipta Gema
Sohee Yang
Wai-Chung Kwan
Xuanli He
Wenda Li
Pasquale Minervini
Eleonora Giunchiglia
Shay B. Cohen
ReLMBDLLRM
189
3
0
29 Aug 2025
Don't Overthink It: A Survey of Efficient R1-style Large Reasoning Models
Don't Overthink It: A Survey of Efficient R1-style Large Reasoning Models
Linan Yue
Yichao Du
Yizhi Wang
W. Gao
Fangzhou Yao
...
Ye Liu
Ziyu Xu
Qi Liu
Shimin Di
Xiaoshi Zhong
LRM
191
15
0
04 Aug 2025
Efficiency-Effectiveness Reranking FLOPs for LLM-based Rerankers
Efficiency-Effectiveness Reranking FLOPs for LLM-based Rerankers
Zhiyuan Peng
Ting-ruen Wei
Tingyu Song
Yilun Zhao
190
0
0
08 Jul 2025
1