Finding the best learning to rank algorithms for effort-aware defect prediction
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Article number | 107165 |
Journal / Publication | Information and Software Technology |
Volume | 157 |
Online published | 2 Feb 2023 |
Publication status | Published - May 2023 |
Link(s)
Abstract
Context: Effort-Aware Defect Prediction (EADP) ranks software modules or changes based on their predicted number of defects (i.e., considering modules or changes as effort) or defect density (i.e., considering LOC as effort) by using learning to rank algorithms. Ranking instability refers to the inconsistent conclusions produced by existing empirical studies of EADP. The major reason is the poor experimental design, such as comparison of few learning to rank algorithms, the use of small number of datasets or datasets without indicating numbers of defects, and evaluation with inappropriate or few metrics.
Objective: To find a stable ranking of learning to rank algorithms to investigate the best ones for EADP,
Method: We examine the practical effects of 34 algorithms on 49 datasets for EADP. We measure the performance of these algorithms using 7 module-based and 7 LOC-based metrics and run experiments under cross-release and cross-project settings, respectively. Finally, we obtain the ranking of these algorithms by performing the Scott-Knott ESD test.
Results: When module is used as effort, random forest regression performs the best under cross-release setting, and linear regression performs the best under cross-project setting among the learning to rank algorithms; (2) when LOC is used as effort, LTR-linear (Learning-to-Rank with the linear model) performs the best under cross-release setting, and Ranking SVM performs the best under cross-project setting.
Conclusion: This comprehensive experimental procedure allows us to discover a stable ranking of the studied algorithms to select the best ones according to the requirement of software projects. © 2023 Elsevier B.V.
Objective: To find a stable ranking of learning to rank algorithms to investigate the best ones for EADP,
Method: We examine the practical effects of 34 algorithms on 49 datasets for EADP. We measure the performance of these algorithms using 7 module-based and 7 LOC-based metrics and run experiments under cross-release and cross-project settings, respectively. Finally, we obtain the ranking of these algorithms by performing the Scott-Knott ESD test.
Results: When module is used as effort, random forest regression performs the best under cross-release setting, and linear regression performs the best under cross-project setting among the learning to rank algorithms; (2) when LOC is used as effort, LTR-linear (Learning-to-Rank with the linear model) performs the best under cross-release setting, and Ranking SVM performs the best under cross-project setting.
Conclusion: This comprehensive experimental procedure allows us to discover a stable ranking of the studied algorithms to select the best ones according to the requirement of software projects. © 2023 Elsevier B.V.
Research Area(s)
- Empirical study, Learning to rank, Ranking instability, Software defect prediction
Citation Format(s)
Finding the best learning to rank algorithms for effort-aware defect prediction. / Yu, Xiao; Dai, Heng; Li, Li et al.
In: Information and Software Technology, Vol. 157, 107165, 05.2023.
In: Information and Software Technology, Vol. 157, 107165, 05.2023.
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review