- The paper demonstrates that a simplified Top-5 metric set effectively preserves prediction accuracy in both within-project and cross-project scenarios.
- It shows that simple classifiers, especially Naïve Bayes, maintain a strong balance between precision and recall using reduced metrics.
- Findings indicate that leveraging cross-project data can improve Recall and F-measure over project-specific data, offering a cost-effective prediction strategy.
An Empirical Study on Software Defect Prediction with a Simplified Metric Set
The paper "An Empirical Study on Software Defect Prediction with a Simplified Metric Set" focuses on the feasibility and effectiveness of using simplified metric sets in software defect prediction, with an emphasis on both within-project defect prediction (WPDP) and cross-project defect prediction (CPDP). The research hinges on several pivotal questions regarding the selection of training data, metric simplification, and classifier choice—each critically examined to optimize predictive accuracy while minimizing computational overhead.
Research Motivation and Methodology
In software engineering, defect prediction models are vital to identify the most defect-prone components, thereby allowing strategic allocation of resources for testing and maintenance. This study aims to validate whether a simplified approach using a reduced metric set can maintain prediction accuracy when historical data is insufficient or resource constraints are present. The paper proposes a simplified metric identification approach that focuses on the Top-k metrics frequently appearing in prediction models, tested across 34 releases of 10 open-source projects sourced from the PROMISE repository.
The researchers constructed predictive models employing six typical classifiers—J48, Logistic Regression, Na\"{\i}ve Bayes, Decision Table, Support Vector Machine (SVM), and Bayesian Network—and compared the performance across three scenarios of training data:
- WPDP using the nearest historical release.
- WPDP using all available historical releases.
- CPDP using the most suitable releases from other projects.
Key Findings
- Training Data Selection:
- CPDP can yield comparable, and often superior, Recall and F-measure relative to WPDP, particularly when project-specific training data is limited. This suggests an advantage in leveraging external data sources in such contexts.
- Simplified Metric Sets:
- The study finds that defect prediction models constructed with a simplified set of Top-5 frequently occurring metrics demonstrate acceptable performance compared to those using the full set of metrics. This reduction significantly lowers data acquisition and processing costs.
- Classifier Performance:
- Simple classifiers, notably Na\"{\i}ve Bayes, were shown to perform exceptionally well using the simplified metric sets, reinforcing the potential efficacy of less complex models in maintaining the balance between precision and recall.
- Minimum Metric Subset:
- A further reduced metric set, such as CBO+LOC+LCOM, presented stable and comparable results across different scenarios and classifier choices, indicating its potential as a universal baseline for defect prediction with general applicability.
Implications and Future Work
The implications of this research are significant, offering a viable pathway to reduce the complexity of defect prediction models without significantly compromising performance. This has practical utility for software projects constrained by limited resources or historical data. Additionally, the findings support the development of cost-effective and generalized defect prediction frameworks. Future work could expand this research by validating the findings across a broader range of languages, software domains, and incorporating more complex dynamic metrics or process measures. Furthermore, exploring the integration of these simplified models into real-time deployment pipelines could present valuable extensions to practical applications of AI in software engineering tasks.