CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning
Main:34 Pages
2 Figures
Bibliography:2 Pages
19 Tables
Abstract
The exponential growth in demand for GPU computing resources, driven by the rapid advancement of Large Language Models, has created an urgent need for automated CUDA optimization strategies. While recent advances in LLMs show promise for code generation, current SOTA models (e.g. R1, o1) achieve low success rates in improving CUDA speed. In this paper, we introduce CUDA-L1, an automated reinforcement learning framework for CUDA optimization.
View on arXivComments on this paper
