AI power demand is growing unprecedentedly thanks to the high power density of AI compute and the emerging inferencing workload. On the supply side, abundant wind power is waiting for grid access in interconnection queues. In this light, this paper argues bringing AI workload to modular compute clusters co-located in wind farms. Our deployment right-sizing strategy makes it economically viable to deploy more than 6 million high-end GPUs today that could consume cheap, green power at its source. We built Heron, a cross-site software router, that could efficiently leverage the complementarity of power generation across wind farms by routing AI inferencing workload around power drops. Using 1-week ofcoding and conversation production traces from Azure and (real) variable wind power traces, we show how Heron improves aggregate goodput of AI compute by up to 80% compared to the state-of-the-art.
View on arXiv@article{reddy2025_2505.09989, title={ AI Greenferencing: Routing AI Inferencing to Green Modular Data Centers with Heron }, author={ Tella Rajashekhar Reddy and Palak and Rohan Gandhi and Anjaly Parayil and Chaojie Zhang and Mike Shepperd and Liangcheng Yu and Jayashree Mohan and Srinivasan Iyengar and Shivkumar Kalyanaraman and Debopam Bhattacherjee }, journal={arXiv preprint arXiv:2505.09989}, year={ 2025 } }