Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
All Papers
0 / 0 papers shown
Title
Home
Papers
2407.04014
Cited By
Offline Energy-Optimal LLM Serving: Workload-Based Energy Models for LLM Inference on Heterogeneous Systems
4 July 2024
Grant Wilkins
Srinivasan Keshav
Richard Mortier
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Offline Energy-Optimal LLM Serving: Workload-Based Energy Models for LLM Inference on Heterogeneous Systems"
4 / 4 papers shown
Title
From Prompts to Power: Measuring the Energy Footprint of LLM Inference
Francisco Caravaca
Ángel Cuevas
R. Cuevas
76
0
0
05 Nov 2025
GPT-OSS-20B: A Comprehensive Deployment-Centric Analysis of OpenAI's Open-Weight Mixture of Experts Model
Deepak Kumar
Divakar Yadav
Yash Patel
MoE
152
3
0
22 Aug 2025
Energy Considerations of Large Language Model Inference and Efficiency Optimizations
Annual Meeting of the Association for Computational Linguistics (ACL), 2025
Jared Fernandez
Clara Na
Vashisth Tiwari
Yonatan Bisk
Sasha Luccioni
Emma Strubell
429
15
0
24 Apr 2025
On-Device Language Models: A Comprehensive Review
Jiajun Xu
Zhiyuan Li
Wei Chen
Qun Wang
Xin Gao
Qi Cai
Ziyuan Ling
468
90
0
26 Aug 2024
1