ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.15748
24
0

Alchemy: Amplifying Theorem-Proving Capability through Symbolic Mutation

21 October 2024
Shaonan Wu
Shuai Lu
Y. Gong
Nan Duan
Ping Wei
    AIMat
ArXivPDFHTML
Abstract

Formal proofs are challenging to write even for experienced experts. Recent progress in Neural Theorem Proving (NTP) shows promise in expediting this process. However, the formal corpora available on the Internet are limited compared to the general text, posing a significant data scarcity challenge for NTP. To address this issue, this work proposes Alchemy, a general framework for data synthesis that constructs formal theorems through symbolic mutation. Specifically, for each candidate theorem in Mathlib, we identify all invocable theorems that can be used to rewrite or apply to it. Subsequently, we mutate the candidate theorem by replacing the corresponding term in the statement with its equivalent form or antecedent. As a result, our method increases the number of theorems in Mathlib by an order of magnitude, from 110k to 6M. Furthermore, we perform continual pretraining and supervised finetuning on this augmented corpus for large language models. Experimental results demonstrate the effectiveness of our approach, achieving a 4.70% absolute performance improvement on Leandojo benchmark. Additionally, our approach achieves a 2.47% absolute performance gain on the out-of-distribution miniF2F benchmark based on the syntheticthis http URLprovide further insights, we conduct a comprehensive analysis of synthetic data composition and the training paradigm, offering valuable guidance for developing a strong theorem prover.

View on arXiv
@article{wu2025_2410.15748,
  title={ Alchemy: Amplifying Theorem-Proving Capability through Symbolic Mutation },
  author={ Shaonan Wu and Shuai Lu and Yeyun Gong and Nan Duan and Ping Wei },
  journal={arXiv preprint arXiv:2410.15748},
  year={ 2025 }
}
Comments on this paper