Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2505.20674
Cited By
v1
v2 (latest)
Pretraining Language Models to Ponder in Continuous Space
27 May 2025
Boyi Zeng
Shixiang Song
Siyuan Huang
Yixuan Wang
He Li
Ziwei He
Xinbing Wang
Zhiyu Li
Zhouhan Lin
LRM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Pretraining Language Models to Ponder in Continuous Space"
5 / 5 papers shown
Title
Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
Yang Yue
Zhiqi Chen
Rui Lu
Andrew Zhao
Zhaokai Wang
Yang Yue
Shiji Song
Gao Huang
ReLM
LRM
233
128
0
18 Apr 2025
Chain-of-Thought Matters: Improving Long-Context Language Models with Reasoning Path Supervision
Dawei Zhu
Xiyu Wei
Guangxiang Zhao
Wenhao Wu
Haosheng Zou
Junfeng Ran
Xun Wang
Lin Sun
Xiangzheng Zhang
Sujian Li
LRM
134
3
0
28 Feb 2025
BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation
Bo Pang
Hanze Dong
Jiacheng Xu
Siyang Song
Yingbo Zhou
Caiming Xiong
KELM
LRM
174
10
0
06 Feb 2025
T1: Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling
Zhenyu Hou
Xin Lv
Rui Lu
Jing Zhang
Yongqian Li
Zijun Yao
Juanzi Li
J. Tang
Yuxiao Dong
OffRL
LRM
ReLM
151
33
0
20 Jan 2025
Variational Best-of-N Alignment
Afra Amini
Tim Vieira
Ryan Cotterell
Ryan Cotterell
BDL
107
23
0
08 Jul 2024
1