Green-Code: Learning to Optimize Energy Efficiency in LLM-Based Code Generation

Open Access
Authors
  • S. Ilager
  • Lukas Florian Briem
  • Ivona Brandic
Publication date 2025
Book title 2025 IEEE 25th International Symposium on Cluster, Cloud and Internet Computing
Book subtitle CCGrid 2025 : Tromsø, Norway, 19-22 May 2025 : proceedings
ISBN
  • 9798331509354
ISBN (electronic)
  • 9798331509347
Event 25th IEEE/ACM International Symposium on Cluster, Cloud and Internet Computing, CCGrid 2025
Pages (from-to) 559-569
Publisher Los Alamitos, California: IEEE Computer Society
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract

Large Language Models (LLMs) are becoming integral to daily life, showcasing their vast potential across various Natural Language Processing (NLP) tasks. Beyond NLP, LLMs are increasingly used in software development tasks, such as code completion, modification, bug fixing, and code translation. Software engineers widely use tools like GitHub Copilot and Amazon Q, streamlining workflows and automating tasks with high accuracy. While the resource and energy intensity of LLM training is often highlighted, inference can be even more resourceintensive over time, as it's a continuous process with a high number of invocations. Therefore, developing resource-efficient alternatives for LLM inference is crucial for sustainability. This work proposes GREEN-CODE, a framework for energy-aware code generation in LLMs. GREEN-CODE performs dynamic early exit during LLM inference. We train a Reinforcement Learning (RL) agent that learns to balance the trade-offs between accuracy, latency, and energy consumption. Our approach is evaluated on two open-source LLMs, Llama 3.2 3B and OPT 2.7 B, using the JavaCorpus and PY150 datasets. Results show that our method reduces the energy consumption between 2350 % on average for code generation tasks without significantly affecting accuracy.

Document type Conference contribution
Language English
Published at https://doi.org/10.1109/CCGRID64434.2025.00068
Other links https://www.proceedings.com/80818.html
Downloads
Permalink to this page
Back