TY - GEN
T1 - A dataset from change history to support evaluation of software maintenance tasks
AU - Dit, Bogdan
AU - Holtzhauer, Andrew
AU - Poshyvanyk, Denys
AU - Kagdi, Huzefa
PY - 2013
Y1 - 2013
N2 - Approaches that support software maintenance need to be evaluated and compared against existing ones, in order to demonstrate their usefulness in practice. However, oftentimes the lack of well-established sets of benchmarks leads to situations where these approaches are evaluated using different datasets, which results in biased comparisons. In this data paper we describe and make publicly available a set of benchmarks from six Java applications, which can be used in the evaluation of various software engineering (SE) tasks, such as feature location and impact analysis. These datasets consist of textual description of change requests, the locations in the source code where they were implemented, and execution traces. Four of the benchmarks were already used in several SE research papers, and two of them are new. In addition, we describe in detail the methodology used for generating these benchmarks and provide a suite of tools in order to encourage other researchers to validate our datasets and generate new benchmarks for other subject software systems. Our online appendix: http://www.cs.wm.edu/semeru/data/msr13/
AB - Approaches that support software maintenance need to be evaluated and compared against existing ones, in order to demonstrate their usefulness in practice. However, oftentimes the lack of well-established sets of benchmarks leads to situations where these approaches are evaluated using different datasets, which results in biased comparisons. In this data paper we describe and make publicly available a set of benchmarks from six Java applications, which can be used in the evaluation of various software engineering (SE) tasks, such as feature location and impact analysis. These datasets consist of textual description of change requests, the locations in the source code where they were implemented, and execution traces. Four of the benchmarks were already used in several SE research papers, and two of them are new. In addition, we describe in detail the methodology used for generating these benchmarks and provide a suite of tools in order to encourage other researchers to validate our datasets and generate new benchmarks for other subject software systems. Our online appendix: http://www.cs.wm.edu/semeru/data/msr13/
KW - Datasets
KW - Feature location
KW - Generate benchmarks
KW - Impact analysis
UR - http://www.scopus.com/inward/record.url?scp=84889032729&partnerID=8YFLogxK
U2 - 10.1109/MSR.2013.6624019
DO - 10.1109/MSR.2013.6624019
M3 - Conference contribution
AN - SCOPUS:84889032729
SN - 9781467329361
T3 - IEEE International Working Conference on Mining Software Repositories
SP - 131
EP - 134
BT - 2013 10th Working Conference on Mining Software Repositories, MSR 2013 - Proceedings
T2 - 10th International Working Conference on Mining Software Repositories, MSR 2013
Y2 - 18 May 2013 through 19 May 2013
ER -