ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.08681
59
0

Self-Taught Self-Correction for Small Language Models

11 March 2025
Viktor Moskvoretskii
Chris Biemann
Irina Nikishina
    LRM
ArXivPDFHTML
Abstract

Although large language models (LLMs) have achieved remarkable performance across various tasks, they remain prone to errors. A key challenge is enabling them to self-correct. While prior research has relied on external tools or large proprietary models, this work explores self-correction in small language models (SLMs) through iterative fine-tuning using solely self-generated data. We introduce the Self-Taught Self-Correction (STaSC) algorithm, which incorporates multiple algorithmic design choices. Experimental results on a question-answering task demonstrate that STaSC effectively learns self-correction, leading to significant performance improvements. Our analysis further provides insights into the mechanisms of self-correction and the impact of different design choices on learning dynamics and overall performance. To support future research, we release our user-friendly codebase and lightweight models.

View on arXiv
@article{moskvoretskii2025_2503.08681,
  title={ Self-Taught Self-Correction for Small Language Models },
  author={ Viktor Moskvoretskii and Chris Biemann and Irina Nikishina },
  journal={arXiv preprint arXiv:2503.08681},
  year={ 2025 }
}
Comments on this paper