170

Agnostics: Learning to Code in Any Programming Language via Reinforcement with a Universal Learning Environment

Main:10 Pages
21 Figures
Bibliography:2 Pages
8 Tables
Appendix:18 Pages
Abstract

Large language models (LLMs) already excel at writing code in high-resource languages such as Python and JavaScript, yet stumble on low-resource languages that remain essential to science and engineering. Besides the obvious shortage of pre-training data, post-training itself is a bottleneck: every new language seems to require new datasets, test harnesses, and reinforcement-learning (RL) infrastructure.

View on arXiv
Comments on this paper