DocumentCode :
868882
Title :
Making inferences with small numbers of training sets
Author :
Kirsopp, C. ; Shepperd, M.
Author_Institution :
Empirical Software Eng. Res. Group, Bournemouth Univ., UK
Volume :
149
Issue :
5
fYear :
2002
fDate :
10/1/2002 12:00:00 AM
Firstpage :
123
Lastpage :
130
Abstract :
A potential methodological problem with empirical studies that assess project effort prediction system is discussed. Frequently, a hold-out strategy is deployed so that the data set is split into a training and a validation set. Inferences are then made concerning the relative accuracy of the different prediction techniques under examination. This is typically done on very small numbers of sampled training sets. It is shown that such studies can lead to almost random results (particularly where relatively small effects are being studied). To illustrate this problem, two data sets are analysed using a configuration problem for case-based prediction and results generated from 100 training sets. This enables results to be produced with quantified confidence limits. From this it is concluded that in both cases using less than five training sets leads to untrustworthy results, and ideally more than 20 sets should be deployed. Unfortunately, this raises a question over a number of empirical validations of prediction techniques, and so it is suggested that further research is needed as a matter of urgency.
Keywords :
software development management; case-based prediction; configuration problem; empirical validations; hold-out strategy; inferences; methodological problem; prediction techniques; project effort prediction system; sampled training sets; validation set;
fLanguage :
English
Journal_Title :
Software, IEE Proceedings -
Publisher :
iet
ISSN :
1462-5970
Type :
jour
DOI :
10.1049/ip-sen:20020695
Filename :
1049201
Link To Document :
بازگشت