DocumentCode
3658737
Title
How Many Ground Truths Should We Insert? Having Good Quality of Labeling Tasks in Crowdsourcing
Author
Takuya Kubota;Masayoshi Aritsugi
Author_Institution
Grad. Sch. of Sci. &
Volume
2
fYear
2015
fDate
7/1/2015 12:00:00 AM
Firstpage
796
Lastpage
805
Abstract
Having a lot of labels of good quality by crowd sourcing has attracted considerable interest recently. Ground truths can be helpful to this end, but prior work does not adequately address how many ground truths should be used. This paper presents a method for determining the number of ground truths. The number is determined by iteratively calculating the expected quality of labels if a ground truth is inserted into labeling tasks and comparing it with the limit of estimation quality of labels expectedly obtained by crowd sourcing. Our method can be applied to general EM algorithm-based approaches to estimating consensus labels of good quality. We compare our method with an EM algorithm-based approach, which is adopted to our method in the discussions of this paper, in terms of both efficiency of collecting labels from crowd and quality of labels obtained from the collected ones.
Keywords
"Estimation","Crowdsourcing","Labeling","Machine learning algorithms","Computational complexity","Mathematical model","Proposals"
Publisher
ieee
Conference_Titel
Computer Software and Applications Conference (COMPSAC), 2015 IEEE 39th Annual
Electronic_ISBN
0730-3157
Type
conf
DOI
10.1109/COMPSAC.2015.117
Filename
7273702
Link To Document