ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.01560
  4. Cited By
OpenMathInstruct-2: Accelerating AI for Math with Massive Open-Source
  Instruction Data
v1v2 (latest)

OpenMathInstruct-2: Accelerating AI for Math with Massive Open-Source Instruction Data

International Conference on Learning Representations (ICLR), 2024
2 October 2024
Shubham Toshniwal
Wei Du
Ivan Moshkov
Branislav Kisacanin
Alexan Ayrapetyan
Igor Gitman
    LRM
ArXiv (abs)PDFHTMLHuggingFace (4 upvotes)

Papers citing "OpenMathInstruct-2: Accelerating AI for Math with Massive Open-Source Instruction Data"

50 / 83 papers shown
Not-a-Bandit: Provably No-Regret Drafter Selection in Speculative Decoding for LLMs
Not-a-Bandit: Provably No-Regret Drafter Selection in Speculative Decoding for LLMs
Hongyi Liu
Jiaji Huang
Zhen Jia
Youngsuk Park
Yu Wang
OffRL
138
2
0
22 Oct 2025
ECG-LLM-- training and evaluation of domain-specific large language models for electrocardiography
ECG-LLM-- training and evaluation of domain-specific large language models for electrocardiography
Lara Ahrens
Wilhelm Haverkamp
Nils Strodthoff
128
0
0
21 Oct 2025
Pay Attention to the Triggers: Constructing Backdoors That Survive Distillation
Pay Attention to the Triggers: Constructing Backdoors That Survive Distillation
Giovanni De Muri
Mark Vero
Robin Staab
Martin Vechev
165
0
0
21 Oct 2025
FineVision: Open Data Is All You Need
FineVision: Open Data Is All You Need
Luis Wiedmann
Orr Zohar
Amir Mahla
Xiaohan Wang
Rui Li
Thibaud Frere
Leandro von Werra
Aritra Roy Gosthipaty
Andrés Marafioti
VLM
195
13
0
20 Oct 2025
QueST: Incentivizing LLMs to Generate Difficult Problems
QueST: Incentivizing LLMs to Generate Difficult Problems
Hanxu Hu
Xingxing Zhang
Jannis Vamvas
Rico Sennrich
Furu Wei
AIMatSyDaMQLRM
255
0
0
20 Oct 2025
To Infinity and Beyond: Tool-Use Unlocks Length Generalization in State Space Models
To Infinity and Beyond: Tool-Use Unlocks Length Generalization in State Space Models
Eran Malach
Omid Saremi
Sinead Williamson
Arwen Bradley
Aryo Lotfi
Emmanuel Abbe
J. Susskind
Etai Littwin
164
0
0
16 Oct 2025
HoneyBee: Data Recipes for Vision-Language Reasoners
HoneyBee: Data Recipes for Vision-Language Reasoners
Hritik Bansal
Devandra Singh Sachan
Kai-Wei Chang
Aditya Grover
Gargi Ghosh
Wen-tau Yih
Ramakanth Pasunuru
VLMLRM
161
3
0
14 Oct 2025
Pushing on Multilingual Reasoning Models with Language-Mixed Chain-of-Thought
Pushing on Multilingual Reasoning Models with Language-Mixed Chain-of-Thought
Guijin Son
Donghun Yang
Hitesh Laxmichand Patel
Amit Agarwal
Hyunwoo Ko
...
Minhyuk Kim
Nikunj Drolia
Dasol Choi
Kyong-Ha Lee
Youngjae Yu
LRM
151
1
0
05 Oct 2025
Principled and Tractable RL for Reasoning with Diffusion Language Models
Principled and Tractable RL for Reasoning with Diffusion Language Models
Anthony Zhan
DiffMAI4CE
111
2
0
05 Oct 2025
GuidedSampling: Steering LLMs Towards Diverse Candidate Solutions at Inference-Time
GuidedSampling: Steering LLMs Towards Diverse Candidate Solutions at Inference-Time
Divij Handa
Mihir Parmar
Aswin Rrv
Md Nayem Uddin
Hamid Palangi
Chitta Baral
93
0
0
04 Oct 2025
Graph2Eval: Automatic Multimodal Task Generation for Agents via Knowledge Graphs
Graph2Eval: Automatic Multimodal Task Generation for Agents via Knowledge Graphs
Yurun Chen
Xavier Hu
Y. Liu
Ziqi Wang
Zeyi Liao
...
Feng Wei
Yuxi Qian
Bo Zheng
Keting Yin
Shengyu Zhang
LLMAG
237
1
0
01 Oct 2025
Beyond English-Centric Training: How Reinforcement Learning Improves Cross-Lingual Reasoning in LLMs
Beyond English-Centric Training: How Reinforcement Learning Improves Cross-Lingual Reasoning in LLMs
Shulin Huang
Yiran Ding
Junshu Pan
Yue Zhang
OffRLLRM
130
1
0
28 Sep 2025
Learning More with Less: A Dynamic Dual-Level Down-Sampling Framework for Efficient Policy Optimization
Learning More with Less: A Dynamic Dual-Level Down-Sampling Framework for Efficient Policy Optimization
Chao Wang
Tao Yang
Hongtao Tian
Yunsheng Shi
Qiyao Ma
Xiaotao Liu
Ting Yao
Wenbo Ding
OffRL
121
0
0
26 Sep 2025
Exploring Solution Divergence and Its Effect on Large Language Model Problem Solving
Exploring Solution Divergence and Its Effect on Large Language Model Problem Solving
Hang Li
Kaiqi Yang
Yucheng Chu
Hui Liu
Shucheng Zhou
MoMeLRM
121
1
0
26 Sep 2025
ScaleDiff: Scaling Difficult Problems for Advanced Mathematical Reasoning
ScaleDiff: Scaling Difficult Problems for Advanced Mathematical Reasoning
Qizhi Pei
Zhuoshi Pan
Honglin Lin
Xin Gao
Yu Li
Zinan Tang
Conghui He
Rui Yan
Lijun Wu
AIMatOffRLLRM
225
2
0
25 Sep 2025
Expanding Reasoning Potential in Foundation Model by Learning Diverse Chains of Thought Patterns
Expanding Reasoning Potential in Foundation Model by Learning Diverse Chains of Thought Patterns
Xuemiao Zhang
Can Ren
Chengying Tu
Rongxiang Weng
Shuo Wang
Hongfei Yan
Jingang Wang
Xunliang Cai
LRMAI4CE
216
1
0
25 Sep 2025
CogAtom: From Cognitive Atoms to Olympiad-level Mathematical Reasoning in Large Language Models
CogAtom: From Cognitive Atoms to Olympiad-level Mathematical Reasoning in Large Language Models
Zhuofan Chen
Jiyuan He
Yichi Zhang
Xing Hu
Haoxing Wen
Jun Bai
Wenge Rong
LRM
239
0
0
22 Sep 2025
SAIL-VL2 Technical Report
SAIL-VL2 Technical Report
Weijie Yin
Yongjie Ye
Fangxun Shu
Yue Liao
Zijian Kang
...
Han Wang
Wenzhuo Liu
Xiao Liang
Shuicheng Yan
Chao Feng
LRMVLM
297
4
0
17 Sep 2025
Optimal Sparsity of Mixture-of-Experts Language Models for Reasoning Tasks
Optimal Sparsity of Mixture-of-Experts Language Models for Reasoning Tasks
Taishi Nakamura
Satoki Ishikawa
Masaki Kawamura
Takumi Okamoto
Daisuke Nohara
Jun Suzuki
Rio Yokota
MoELRM
175
0
0
26 Aug 2025
Can Structured Templates Facilitate LLMs in Tackling Harder Tasks? : An Exploration of Scaling Laws by Difficulty
Can Structured Templates Facilitate LLMs in Tackling Harder Tasks? : An Exploration of Scaling Laws by Difficulty
Zhichao Yang
Zhaoxin Fan
Gen Li
Yuanze Hu
Xinyu Wang
Ye Qiu
Xin Wang
Yifan Sun
Wenjun Wu
LRM
85
0
0
26 Aug 2025
NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model
NVIDIA Nemotron Nano 2: An Accurate and Efficient Hybrid Mamba-Transformer Reasoning Model
Nvidia
Aarti Basant
Abhijit Khairnar
Abhijit Paithankar
Abhinav Khattar
...
Keith Wyss
Keshav Santhanam
Kezhi Kong
Krzysztof Pawelec
Kumar Anik
LRM
298
0
0
20 Aug 2025
Data Mixing Optimization for Supervised Fine-Tuning of Large Language Models
Data Mixing Optimization for Supervised Fine-Tuning of Large Language Models
Yuan Li
Zhengzhong Liu
Eric P. Xing
139
1
0
16 Aug 2025
Apriel-Nemotron-15B-Thinker
Apriel-Nemotron-15B-Thinker
Shruthan Radhakrishna
S. Parikh
Gopal Sarda
Anil Turkkan
Quaizar Vohra
...
Sathwik Tejaswi Madhusudhan
Torsten Scholak
Sébastien Paquet
Sagar Davasam
Srinivas Sunkara
LLMAGMoELRM
205
2
0
13 Aug 2025
MathSmith: Towards Extremely Hard Mathematical Reasoning by Forging Synthetic Problems with a Reinforced Policy
MathSmith: Towards Extremely Hard Mathematical Reasoning by Forging Synthetic Problems with a Reinforced Policy
Shaoxiong Zhan
Yanlin Lai
Ziyu Lu
Dahua Lin
Ziqing Yang
Fei Tang
LRM
124
10
0
07 Aug 2025
WarriorMath: Enhancing the Mathematical Ability of Large Language Models with a Defect-aware Framework
WarriorMath: Enhancing the Mathematical Ability of Large Language Models with a Defect-aware Framework
Yue Chen
Minghua He
Fangkai Yang
Pu Zhao
Lu Wang
...
Yuefeng Zhan
Hao Sun
Qingwei Lin
Saravan Rajmohan
Dongmei Zhang
181
2
0
02 Aug 2025
SAND-Math: Using LLMs to Generate Novel, Difficult and Useful Mathematics Questions and Answers
SAND-Math: Using LLMs to Generate Novel, Difficult and Useful Mathematics Questions and Answers
Chaitanya Manem
Pratik Prabhanjan Brahma
Prakamya Mishra
Zicheng Liu
Emad Barsoum
AIMatLRM
344
4
0
28 Jul 2025
Diversity-Enhanced Reasoning for Subjective Questions
Diversity-Enhanced Reasoning for Subjective Questions
Yumeng Wang
Zhiyuan Fan
Jiayu Liu
J. Huang
Yi R. Fung
LRM
493
6
0
27 Jul 2025
PITA: Preference-Guided Inference-Time Alignment for LLM Post-Training
PITA: Preference-Guided Inference-Time Alignment for LLM Post-Training
Sarat Chandra Bobbili
Ujwal Dinesha
Dheeraj Narasimha
S. Shakkottai
165
2
0
26 Jul 2025
GenSelect: A Generative Approach to Best-of-N
GenSelect: A Generative Approach to Best-of-N
Shubham Toshniwal
Ivan Sorokin
Aleksander Ficek
Ivan Moshkov
Igor Gitman
LRM
142
7
0
23 Jul 2025
MegaScience: Pushing the Frontiers of Post-Training Datasets for Science Reasoning
MegaScience: Pushing the Frontiers of Post-Training Datasets for Science Reasoning
Run-Ze Fan
Zengzhi Wang
Pengfei Liu
LRM
326
15
0
22 Jul 2025
EvoLM: In Search of Lost Language Model Training Dynamics
EvoLM: In Search of Lost Language Model Training Dynamics
Zhenting Qi
Fan Nie
Alexandre Alahi
James Zou
Himabindu Lakkaraju
Yilun Du
Eric P. Xing
Sham Kakade
Hanlin Zhang
316
2
0
19 Jun 2025
Test-Time-Scaling for Zero-Shot Diagnosis with Visual-Language Reasoning
Test-Time-Scaling for Zero-Shot Diagnosis with Visual-Language Reasoning
Ji Young Byun
Young-Jin Park
Navid Azizan
Rama Chellappa
LM&MALRM
161
1
0
11 Jun 2025
TaskCraft: Automated Generation of Agentic Tasks
TaskCraft: Automated Generation of Agentic Tasks
Dingfeng Shi
Jingyi Cao
Qianben Chen
W. Sun
W. Li
...
Jiaheng Liu
Changwang Zhang
Jun Wang
Yuchen Eleanor Jiang
Wangchunshu Zhou
305
20
0
11 Jun 2025
Reinforce LLM Reasoning through Multi-Agent Reflection
Yurun Yuan
Tengyang Xie
LRM
317
16
0
10 Jun 2025
A Survey on Large Language Models for Mathematical Reasoning
Peng-Yuan Wang
Tian-Shuo Liu
Chenyang Wang
Yi-Di Wang
Shu Yan
...
Xu-Hui Liu
Xin-Wei Chen
Jia-Cheng Xu
Ziniu Li
Yang Yu
LRM
279
21
0
10 Jun 2025
Improving Large Language Models with Concept-Aware Fine-Tuning
Improving Large Language Models with Concept-Aware Fine-Tuning
Michael K. Chen
Xikun Zhang
Jiaxing Huang
Dacheng Tao
283
1
0
09 Jun 2025
SPARQ: Synthetic Problem Generation for Reasoning via Quality-Diversity Algorithms
SPARQ: Synthetic Problem Generation for Reasoning via Quality-Diversity Algorithms
Alex Havrilla
Edward Hughes
Mikayel Samvelyan
Jacob Abernethy
SyDaLRM
324
5
0
06 Jun 2025
Establishing Trustworthy LLM Evaluation via Shortcut Neuron Analysis
Establishing Trustworthy LLM Evaluation via Shortcut Neuron AnalysisAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Kejian Zhu
Shangqing Tu
Zhuoran Jin
Lei Hou
Juanzi Li
Jun Zhao
KELM
226
0
0
04 Jun 2025
DeepTheorem: Advancing LLM Reasoning for Theorem Proving Through Natural Language and Reinforcement Learning
DeepTheorem: Advancing LLM Reasoning for Theorem Proving Through Natural Language and Reinforcement Learning
Ziyin Zhang
Jiahao Xu
Zhiwei He
Tian Liang
Qiuzhi Liu
...
Zhuosheng Zhang
Rui Wang
Zhaopeng Tu
Haitao Mi
Dong Yu
OffRLLRM
310
10
0
29 May 2025
Benchmarking Abstract and Reasoning Abilities Through A Theoretical Perspective
Benchmarking Abstract and Reasoning Abilities Through A Theoretical Perspective
Qingchuan Ma
Yuhang Wu
Xiawu Zheng
Rongrong Ji
202
1
0
28 May 2025
LASER: Stratified Selective Sampling for Instruction Tuning with Dedicated Scoring Strategy
LASER: Stratified Selective Sampling for Instruction Tuning with Dedicated Scoring Strategy
Paramita Mirza
Lucas Weber
Fabian Küch
287
0
0
28 May 2025
ReCopilot: Reverse Engineering Copilot in Binary Analysis
ReCopilot: Reverse Engineering Copilot in Binary Analysis
Guoqiang Chen
Huiqi Sun
Daguang Liu
Zhiqi Wang
Qiang Wang
Bin Yin
Lu Liu
Lingyun Ying
217
6
0
22 May 2025
Watch your steps: Dormant Adversarial Behaviors that Activate upon LLM Finetuning
Watch your steps: Dormant Adversarial Behaviors that Activate upon LLM Finetuning
Thibaud Gloaguen
Mark Vero
Robin Staab
Martin Vechev
AAML
483
0
0
22 May 2025
Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert Models
Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert Models
Jingcong Liang
Siyuan Wang
Miren Tian
Yitong Li
Duyu Tang
Zhongyu Wei
MoE
321
0
0
21 May 2025
Trust, But Verify: A Self-Verification Approach to Reinforcement Learning with Verifiable Rewards
Trust, But Verify: A Self-Verification Approach to Reinforcement Learning with Verifiable Rewards
Xiaoyuan Liu
Tian Liang
Zhiwei He
Jiahao Xu
Wenxuan Wang
Pinjia He
Zhaopeng Tu
Haitao Mi
Dong Yu
OffRLReLMLRM
347
15
0
19 May 2025
Multi-Token Prediction Needs Registers
Multi-Token Prediction Needs Registers
Anastasios Gerontopoulos
Spyros Gidaris
N. Komodakis
377
3
0
15 May 2025
FineScope : Precision Pruning for Domain-Specialized Large Language Models Using SAE-Guided Self-Data Cultivation
FineScope : Precision Pruning for Domain-Specialized Large Language Models Using SAE-Guided Self-Data Cultivation
Chaitali Bhattacharyya
Hyunsei Lee
Junyoung Lee
Shinhyoung Jang
Il hong Suh
Yeseong Kim
309
1
0
01 May 2025
Accurate and Diverse LLM Mathematical Reasoning via Automated PRM-Guided GFlowNets
Accurate and Diverse LLM Mathematical Reasoning via Automated PRM-Guided GFlowNets
Adam Younsi
Abdalgader Abubaker
Abdalgader Abubaker
Hakim Hacid
Hakim Hacid
Salem Lahlou
LRM
553
6
0
28 Apr 2025
AIMO-2 Winning Solution: Building State-of-the-Art Mathematical Reasoning Models with OpenMathReasoning dataset
AIMO-2 Winning Solution: Building State-of-the-Art Mathematical Reasoning Models with OpenMathReasoning dataset
Ivan Moshkov
Darragh Hanley
Ivan Sorokin
Shubham Toshniwal
Christof Henkel
Benedikt Schifferer
Wei Du
Igor Gitman
ReLMLRM
302
78
0
23 Apr 2025
M1: Towards Scalable Test-Time Compute with Mamba Reasoning Models
M1: Towards Scalable Test-Time Compute with Mamba Reasoning Models
Junxiong Wang
Wen-Ding Li
Daniele Paliotta
Daniel Ritter
Alexander M. Rush
Tri Dao
LRM
352
12
0
14 Apr 2025
12
Next