481

CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning

Main:34 Pages
2 Figures
Bibliography:2 Pages
19 Tables
Abstract

The exponential growth in demand for GPU computing resources, driven by the rapid advancement of Large Language Models, has created an urgent need for automated CUDA optimization strategies. While recent advances in LLMs show promise for code generation, current SOTA models (e.g. R1, o1) achieve low success rates in improving CUDA speed. In this paper, we introduce CUDA-L1, an automated reinforcement learning framework for CUDA optimization.

View on arXiv
Comments on this paper