Back to Search
Start Over
SLO-Aware GPU DVFS for Energy-Efficient LLM Inference Serving.
- Source :
- IEEE Computer Architecture Letters; Jul-Dec2024, Vol. 23 Issue 2, p150-153, 4p
- Publication Year :
- 2024
-
Abstract
- The increasing popularity of LLM-based chatbots combined with their reliance on power-hungry GPU infrastructure forms a critical challenge for providers: minimizing energy consumption under Service-Level Objectives (SLOs) that ensure optimal user experience. Traditional energy optimization methods fall short for LLM inference due to their autoregressive architecture, which renders them incapable of meeting a predefined SLO without energy overprovisioning. This autoregressive nature however, allows for iteration-level adjustments, enabling continuous fine-tuning of the system throughout the inference process. In this letter, we propose a solution based on iteration-level GPU Dynamic Voltage Frequency Scaling (DVFS), aiming to reduce the energy impact of LLM serving, an approach that has the potential for more than 22.8% and up to 45.5% energy gains when tested in real world situations under varying SLO constraints. Our approach works on top of existing LLM hosting services, requires minimal profiling and no intervention to the inference serving system. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 15566056
- Volume :
- 23
- Issue :
- 2
- Database :
- Complementary Index
- Journal :
- IEEE Computer Architecture Letters
- Publication Type :
- Academic Journal
- Accession number :
- 180071839
- Full Text :
- https://doi.org/10.1109/LCA.2024.3406038