Evaluation of Prediction Accuracy of Classical and AI-Based Software Reliability Models
Abstract
The growing dependence on software-intensive systems across critical application domains has made accurate software reliability prediction a fundamental requirement for ensuring system quality and safety and user confidence. Traditionally, software reliability assessments have relied on classical mathematical and statistical models that estimate failure behavior during the testing phase. Although these approaches have proven useful in relatively stable and controlled environments, their effectiveness often diminishes when applied to modern software systems characterized by increasing complexity, nonlinear interactions, and dynamic operational conditions. These limitations have motivated the exploration of more adaptive and data-driven prediction techniques.
This study presents a comprehensive evaluation of classical software reliability models and artificial intelligence–based machine learning approaches in terms of their prediction accuracy. Conventional statistical models were compared with a range of supervised and ensemble learning algorithms, including k-nearest neighbors, support vector machines, decision trees, random forests, AdaBoost, and gradient boosting. A comparative analysis was conducted using multiple publicly available software reliability datasets, and the model performance was assessed using standard accuracy and error-based evaluation metrics.
The experimental results indicate that AI-driven models consistently achieve higher prediction accuracy and improved robustness across different datasets compared to traditional reliability models. In particular, ensemble learning techniques demonstrate superior performance by effectively capturing complex failure patterns and reducing the prediction variance. Although classical models offer advantages such as simplicity and interpretability, they lack the flexibility required to model the evolving failure behavior of modern software systems. The findings of this study highlight the potential of data-driven AI techniques as reliable and scalable solutions for software reliability prediction, supporting their increasing adoption in modern software quality engineering.
References
J. Jelinski and P. B. Moranda, “Software reliability research,” in Statistical Computer Performance Evaluation, W. Freiberger, Ed. New York, NY, USA: Academic Press, 1972, pp. 465–484.
A. L. Goel and K. Okumoto, “Time-dependent error-detection rate model for software reliability and other performance measures,” IEEE Trans. Reliab., vol. R-28, no. 3, pp. 206–211, Aug. 1979.
J. D. Musa and K. Okumoto, “A logarithmic Poisson execution time model for software reliability measurement,” in Proc. 7th Int. Conf. Software Engineering, 1984, pp. 230–238.
J. D. Musa, Software Reliability Engineering. New York, NY, USA: McGraw-Hill, 2004.
B. Littlewood and J. L. Verrall, “A Bayesian reliability growth model for computer software,” J. Roy. Stat. Soc. Series C, vol. 22, no. 3, pp. 332–346, 1973.
M. Xie, Software Reliability Modelling. Singapore: World Scientific, 1991.
K. Kanoun and L. Spainhower, Dependability Benchmarking for Computer Systems. Hoboken, NJ, USA: Wiley-IEEE Press, 2008.
T. M. Khoshgoftaar and N. Seliya, “Comparative assessment of software quality classification techniques: An empirical case study,” Empirical Software Engineering, vol. 9, no. 3, pp. 229–257, 2004.
T. M. Khoshgoftaar and E. B. Allen, “Neural network models for software reliability prediction,” IEEE Trans. Reliab., vol. 45, no. 1, pp. 11–18, Mar. 1996.
H. Pham, “System software reliability,” Springer Series in Reliability Engineering, London, U.K.: Springer, 2006.
Refbacks
- There are currently no refbacks.