TY - JOUR
T1 - On the value of parameter tuning in stacking ensemble model for software regression test effort estimation
AU - Labidi, Taher
AU - Sakhrawi, Zaineb
N1 - Publisher Copyright:
© 2023, The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature.
PY - 2023/10
Y1 - 2023/10
N2 - A type of software testing, regression testing is often costly and labour-intensive. As such, multiple corporations have intensified efforts to estimate the amount of effort required. However, frequent alterations in software projects impact the precision of software regression test effort estimation (SRTEE), which increases the difficulty of managing software projects. Therefore, machine learning (ML) has increasingly been used to develop more accurate SRTEEs. The estimation process of a software project comprises inputs, the model, and outputs. This present study examines the quality of estimation inputs and the model required to deliver accurate estimation outputs. An SRTEE that uses the stacking ensemble model (StackSRTEE) was developed to increase the precision of SRTEE. It consisted of the three most common ML methods, namely neural networks, support vector regression, and decision tree regression. The grid search (GS) technique was then used to tune the hyperparameters of the StackSRTEE before it was trained and tested using a dataset from the International Software Benchmarking Standards Group (ISBSG) repository. The size of the functional change; specifically, enhancement; was used as the primary independent variable to improve the inputs of the StackSRTEE model. With the appropriate features; such as the functional change size of an enhancement; (1) the proposed StackSRTEE model yielded higher accuracy than the three individual ML methods on their own, (2) using GS to tune and set the individual ML methods increased the precision of the SRTEE outputs, and (3) the StackSRTEE-based GS tuning yielded estimations that were more precise.
AB - A type of software testing, regression testing is often costly and labour-intensive. As such, multiple corporations have intensified efforts to estimate the amount of effort required. However, frequent alterations in software projects impact the precision of software regression test effort estimation (SRTEE), which increases the difficulty of managing software projects. Therefore, machine learning (ML) has increasingly been used to develop more accurate SRTEEs. The estimation process of a software project comprises inputs, the model, and outputs. This present study examines the quality of estimation inputs and the model required to deliver accurate estimation outputs. An SRTEE that uses the stacking ensemble model (StackSRTEE) was developed to increase the precision of SRTEE. It consisted of the three most common ML methods, namely neural networks, support vector regression, and decision tree regression. The grid search (GS) technique was then used to tune the hyperparameters of the StackSRTEE before it was trained and tested using a dataset from the International Software Benchmarking Standards Group (ISBSG) repository. The size of the functional change; specifically, enhancement; was used as the primary independent variable to improve the inputs of the StackSRTEE model. With the appropriate features; such as the functional change size of an enhancement; (1) the proposed StackSRTEE model yielded higher accuracy than the three individual ML methods on their own, (2) using GS to tune and set the individual ML methods increased the precision of the SRTEE outputs, and (3) the StackSRTEE-based GS tuning yielded estimations that were more precise.
KW - Grid search tuning technique
KW - Machine learning techniques
KW - Software regression test effort estimation
KW - Software regression testing
KW - Stacking ensemble model
UR - http://www.scopus.com/inward/record.url?scp=85158111910&partnerID=8YFLogxK
U2 - 10.1007/s11227-023-05334-9
DO - 10.1007/s11227-023-05334-9
M3 - Article
AN - SCOPUS:85158111910
SN - 0920-8542
VL - 79
SP - 17123
EP - 17145
JO - Journal of Supercomputing
JF - Journal of Supercomputing
IS - 15
ER -