Context: A software release is the deployment of a version of an evolving software product. Product managers are typically responsible for deciding the release content, time frame, price, and quality of the release. Due to all the dynamic changes in the project and process parameters, the decision is highly complex and of high impact. Objective: This paper has two objectives: i) Comparative analysis of predictive techniques in classifying an ongoing release in terms of its expected release readiness., and ii) Comparative analysis between regular and ensemble classifiers to classify an ongoing release in terms of its expected release readiness. Methodology: We use machine learning classifiers to predict release readiness. We analyzed three OSS projects under Apache Software Foundation from JIRA issue repository. As a retrospective study, we covered a period of 70 months, 85 releases and 1696 issues. We monitored eight established variables to train classifiers in order to predict whether releases will be ready versus non-ready. Predictive performance of different classifiers was compared by measuring precision, recall, F-measure, balanced accuracy, and area under the ROC curve (AUC). Results: Comparative analysis among nine classifiers revealed that ensemble classifiers significantly outperform regular classifiers. Balancing precision and recall, Random Forrest and BaggedADABoost were the two best performers in total, while Naïve Bayes performed best among just the regular classifiers.
%0 Conference Paper
%1 DidarAlAlam:2016:CAP:2896995.2896997
%A Didar Al Alam, S. M.
%A Karim, Muhammad Rezaul
%A Pfahl, Dietmar
%A Ruhe, Günther
%B Proceedings of the 5th International Workshop on Realizing Artificial Intelligence Synergies in Software Engineering
%C New York, NY, USA
%D 2016
%I ACM
%K development
%P 15--21
%R 10.1145/2896995.2896997
%T Comparative Analysis of Predictive Techniques for Release Readiness Classification
%U http://doi.acm.org/10.1145/2896995.2896997
%X Context: A software release is the deployment of a version of an evolving software product. Product managers are typically responsible for deciding the release content, time frame, price, and quality of the release. Due to all the dynamic changes in the project and process parameters, the decision is highly complex and of high impact. Objective: This paper has two objectives: i) Comparative analysis of predictive techniques in classifying an ongoing release in terms of its expected release readiness., and ii) Comparative analysis between regular and ensemble classifiers to classify an ongoing release in terms of its expected release readiness. Methodology: We use machine learning classifiers to predict release readiness. We analyzed three OSS projects under Apache Software Foundation from JIRA issue repository. As a retrospective study, we covered a period of 70 months, 85 releases and 1696 issues. We monitored eight established variables to train classifiers in order to predict whether releases will be ready versus non-ready. Predictive performance of different classifiers was compared by measuring precision, recall, F-measure, balanced accuracy, and area under the ROC curve (AUC). Results: Comparative analysis among nine classifiers revealed that ensemble classifiers significantly outperform regular classifiers. Balancing precision and recall, Random Forrest and BaggedADABoost were the two best performers in total, while Naïve Bayes performed best among just the regular classifiers.
%@ 978-1-4503-4165-3
@inproceedings{DidarAlAlam:2016:CAP:2896995.2896997,
abstract = {Context: A software release is the deployment of a version of an evolving software product. Product managers are typically responsible for deciding the release content, time frame, price, and quality of the release. Due to all the dynamic changes in the project and process parameters, the decision is highly complex and of high impact. Objective: This paper has two objectives: i) Comparative analysis of predictive techniques in classifying an ongoing release in terms of its expected release readiness., and ii) Comparative analysis between regular and ensemble classifiers to classify an ongoing release in terms of its expected release readiness. Methodology: We use machine learning classifiers to predict release readiness. We analyzed three OSS projects under Apache Software Foundation from JIRA issue repository. As a retrospective study, we covered a period of 70 months, 85 releases and 1696 issues. We monitored eight established variables to train classifiers in order to predict whether releases will be ready versus non-ready. Predictive performance of different classifiers was compared by measuring precision, recall, F-measure, balanced accuracy, and area under the ROC curve (AUC). Results: Comparative analysis among nine classifiers revealed that ensemble classifiers significantly outperform regular classifiers. Balancing precision and recall, Random Forrest and BaggedADABoost were the two best performers in total, while Naïve Bayes performed best among just the regular classifiers.},
acmid = {2896997},
added-at = {2016-08-19T16:52:05.000+0200},
address = {New York, NY, USA},
author = {Didar Al Alam, S. M. and Karim, Muhammad Rezaul and Pfahl, Dietmar and Ruhe, G\"{u}nther},
biburl = {https://www.bibsonomy.org/bibtex/2e76d6b7f52cb7dd465c7310ea38c8da2/ispma},
booktitle = {Proceedings of the 5th International Workshop on Realizing Artificial Intelligence Synergies in Software Engineering},
doi = {10.1145/2896995.2896997},
interhash = {78ae1d7cdd346883db4c691559b86864},
intrahash = {e76d6b7f52cb7dd465c7310ea38c8da2},
isbn = {978-1-4503-4165-3},
keywords = {development},
location = {Austin, Texas},
numpages = {7},
pages = {15--21},
publisher = {ACM},
series = {RAISE '16},
timestamp = {2016-08-19T16:55:43.000+0200},
title = {Comparative Analysis of Predictive Techniques for Release Readiness Classification},
url = {http://doi.acm.org/10.1145/2896995.2896997},
year = 2016
}