Language Models in Software Development Tasks: An Experimental Analysis of Energy and Accuracy

  • Negar Alizadeh*
  • , Boris Belchev
  • , Nishant Saurabh
  • , Patricia Kelbert
  • , Fernando Castor
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

1 Citation (Scopus)

Abstract

The use of generative AI-based coding assistants like ChatGPT and Github Copilot is a reality in contemporary software development. Many of these tools are provided as remote APIs. Using third-party APIs raises data privacy and security concerns for client companies, which motivates the use of locallydeployed language models. In this study, we explore the tradeoff between model accuracy and energy consumption, aiming to provide valuable insights to help developers make informed decisions when selecting a language model. We investigate the performance of 18 families of LLMs in typical software development tasks on two real-world infrastructures, a commodity GPU and a powerful AI-specific GPU. Given that deploying LLMs locally requires powerful infrastructure which might not be affordable for everyone, we consider both full-precision and quantized models. Our findings reveal that employing a big LLM with a higher energy budget does not always translate to significantly improved accuracy. Additionally, quantized versions of large models generally offer better efficiency and accuracy compared to full-precision versions of medium-sized ones. Apart from that, not a single model is suitable for all types of software development tasks.

Original languageEnglish
Title of host publication2025 IEEE/ACM 22nd International Conference on Mining Software Repositories, MSR 2025
PublisherIEEE
Pages725-736
Number of pages12
ISBN (Electronic)9798331501839
DOIs
Publication statusPublished - 13 Jun 2025
Event22nd IEEE/ACM International Conference on Mining Software Repositories, MSR 2025 - Ottawa, Canada
Duration: 27 Apr 202529 Apr 2025
Conference number: 22

Conference

Conference22nd IEEE/ACM International Conference on Mining Software Repositories, MSR 2025
Abbreviated titleMSR 2025
Country/TerritoryCanada
CityOttawa
Period27/04/2529/04/25

Keywords

  • 2025 OA procedure
  • Energy Efficiency
  • LLMs
  • Model Quantization
  • Software Development
  • Trade-Offs
  • Coding Assistant

Fingerprint

Dive into the research topics of 'Language Models in Software Development Tasks: An Experimental Analysis of Energy and Accuracy'. Together they form a unique fingerprint.

Cite this