DocumentCode
3060950
Title
Assessing the Reliability of a Human Estimator
Author
Boetticher, Gary D. ; Lokhandwala, Nazim
Author_Institution
Univ. of Houston, Clear Lake City
fYear
2007
fDate
20-26 May 2007
Firstpage
5
Lastpage
5
Abstract
Human-based estimation remains the predominant methodology of choice [1]. Understanding the human estimator is critical for improving the effort estimation process. Every human estimator draws upon their background in terms of domain knowledge, technical knowledge, experience, and education in formulating an estimate. This research uses estimator demographic information to construct over 4000 classifiers which distinguish between the best and worst types of estimators. Various attribute techniques are applied to determine most significant demographics. Best case models produce accuracy rates ranging from 74 to 80 percent. Some of the best case models are presented for gaining insight into how demographics impact effort estimation.
Keywords
knowledge engineering; performance evaluation; demographic information; domain knowledge; human estimator; technical knowledge; Demography; Humans; Lakes; Machine learning; Machine learning algorithms; Maximum likelihood estimation; Neural networks; Parameter estimation; Predictive models; Software algorithms;
fLanguage
English
Publisher
ieee
Conference_Titel
Predictor Models in Software Engineering, 2007. PROMISE'07: ICSE Workshops 2007. International Workshop on
Conference_Location
Minneapolis, MN
Print_ISBN
0-7695-2954-2
Type
conf
DOI
10.1109/PROMISE.2007.2
Filename
4273261
Link To Document