Evaluating explanations for software patches generated by large language models

Dec 4, 2023·
Dominik Sobania
,
Alina Geiger
,
James Callan
,
Alexander Brownlee
,
Carol Hanna
,
Rebecca Moussa
,
Mar Zamorano López
,
Justyna Petke
,
Federica Sarro
· 0 min read
Abstract
Large language models (LLMs) have recently been integrated in a variety of applications including software engineering tasks. In this work, we study the use of LLMs to enhance the explainability of software patches. In particular, we evaluate the performance of GPT 3.5 in explaining patches generated by the search-based automated program repair system ARJA-e for 30 bugs from the popular Defects4J benchmark. We also investigate the performance achieved when explaining the corresponding patches written by software developers. We find that on average 84% of the LLM explanations for machine-generated patches were correct and 54% were complete for the studied categories in at least 1 out of 3 runs. Furthermore, we find that the LLM generates more accurate explanations for machine-generated patches than for human-written ones.
Type
Publication
In International Symposium on Search Based Software Engineering
publications