Int J Performability Eng ›› 2025, Vol. 21 ›› Issue (4): 179-187.doi: 10.23940/ijpe.25.04.p1.179187

    Next Articles

Leveraging Large Language Models for Iterative Software Error Tracking: A Case Study with VirtualBox

Pan Liua,*, Zhongze Yanga, Ruyi Luoa, and Yihao Lib   

  1. aFaculty of Business Information, Shanghai Business School, Shanghai, China;
    bSchool of Information and Electrical Engineering, Ludong University, Yantai, China
  • Submitted on ; Revised on ; Accepted on
  • Contact: *E-mail address: panl008@163.com

Abstract: This paper presents a case study exploring the application of large language models (LLMs) in tracing the origins of software errors. We developed an iterative software error tracking framework that combines the analytical capabilities of LLMs with expert human reasoning. Using this framework, we successfully identified and resolved a software error encountered in VirtualBox. The tracking process involved three key stages: generating outputs with the LLM, conducting human analysis of these outputs, and refining prompts to improve the accuracy of the LLM's responses. This study demonstrates the effectiveness of LLMs in software anomaly detection while emphasizing the critical role of human expertise in guiding the process. The findings offer valuable insights for software testing practitioners on leveraging LLMs to track and resolve runtime anomalies.

Key words: large language models, software error tracking, multi-round interactive Q&A, software testing, human analysis