AthenaLLM: Supporting Experiments with Large Language Models in Software Development

Benedito De Oliveira, Fernando Castor

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

36 Downloads (Pure)

Abstract

Existing studies on the use of Large Language Models (LLMs) in software development leverage methodologies that limit their scalability and require intensive manual data collection and analysis, for example, due to the use of video data or think-aloud protocols. We propose the use of a specialized tool capable of automatically collecting fine-grained, relevant data during experiments and case studies. It enables researchers to understand for example how often participants accept or reject suggestions made by LLMs and what kinds of prompts are more likely to trigger accepted suggestions, even in studies targeting a large number of participants. We implement this idea as a Visual Studio Code plugin named AthenaLLM1. It mimics the functionalities of GitHub Copilot and offers seamless integration with OpenAI API models like GPT-4 and GPT-3.5, and compatibility with other models providing an OpenAI-compatible API, e.g., Vicuna [6]. It automatically collects data at a fine level of granularity and covers both the interactions of developers with their IDE, e.g., all changes made in the code, and the products of such interactions, e.g., the generated code, when accepted. Thus, the proposed approach also reduces bias that the experimental process itself may introduce, e.g., due to the need for participants to verbalize their thoughts. In this paper we discuss how AthenaLLM could enable researchers to go both broader (in terms of number of participants) and deeper (in terms of the kinds of research questions that can be tackled).1Extension available at https://github.com/nandooliveira/athena_llm_extension
Original languageEnglish
Title of host publication2024 IEEE/ACM 32nd International Conference on Program Comprehension (ICPC)
PublisherIEEE
Pages69-73
Number of pages5
ISBN (Print)979-8-3503-5227-6
DOIs
Publication statusPublished - 15 Apr 2024
Event32nd IEEE/ACM International Conference on Program Comprehension, ICPC 2024 - Lisbon, Portugal
Duration: 15 Apr 202416 Apr 2024
Conference number: 32

Conference

Conference32nd IEEE/ACM International Conference on Program Comprehension, ICPC 2024
Abbreviated titleICPC 2024
Country/TerritoryPortugal
CityLisbon
Period15/04/2416/04/24

Keywords

  • 2024 OA procedure
  • Adaptation models
  • Codes
  • Protocols
  • Scalability
  • MIMICs
  • Manuals
  • Visualization

Fingerprint

Dive into the research topics of 'AthenaLLM: Supporting Experiments with Large Language Models in Software Development'. Together they form a unique fingerprint.

Cite this