DocumentCode :
3658737
Title :
How Many Ground Truths Should We Insert? Having Good Quality of Labeling Tasks in Crowdsourcing
Author :
Takuya Kubota;Masayoshi Aritsugi
Author_Institution :
Grad. Sch. of Sci. &
Volume :
2
fYear :
2015
fDate :
7/1/2015 12:00:00 AM
Firstpage :
796
Lastpage :
805
Abstract :
Having a lot of labels of good quality by crowd sourcing has attracted considerable interest recently. Ground truths can be helpful to this end, but prior work does not adequately address how many ground truths should be used. This paper presents a method for determining the number of ground truths. The number is determined by iteratively calculating the expected quality of labels if a ground truth is inserted into labeling tasks and comparing it with the limit of estimation quality of labels expectedly obtained by crowd sourcing. Our method can be applied to general EM algorithm-based approaches to estimating consensus labels of good quality. We compare our method with an EM algorithm-based approach, which is adopted to our method in the discussions of this paper, in terms of both efficiency of collecting labels from crowd and quality of labels obtained from the collected ones.
Keywords :
"Estimation","Crowdsourcing","Labeling","Machine learning algorithms","Computational complexity","Mathematical model","Proposals"
Publisher :
ieee
Conference_Titel :
Computer Software and Applications Conference (COMPSAC), 2015 IEEE 39th Annual
Electronic_ISBN :
0730-3157
Type :
conf
DOI :
10.1109/COMPSAC.2015.117
Filename :
7273702
Link To Document :
بازگشت