• DocumentCode
    3658737
  • Title

    How Many Ground Truths Should We Insert? Having Good Quality of Labeling Tasks in Crowdsourcing

  • Author

    Takuya Kubota;Masayoshi Aritsugi

  • Author_Institution
    Grad. Sch. of Sci. &
  • Volume
    2
  • fYear
    2015
  • fDate
    7/1/2015 12:00:00 AM
  • Firstpage
    796
  • Lastpage
    805
  • Abstract
    Having a lot of labels of good quality by crowd sourcing has attracted considerable interest recently. Ground truths can be helpful to this end, but prior work does not adequately address how many ground truths should be used. This paper presents a method for determining the number of ground truths. The number is determined by iteratively calculating the expected quality of labels if a ground truth is inserted into labeling tasks and comparing it with the limit of estimation quality of labels expectedly obtained by crowd sourcing. Our method can be applied to general EM algorithm-based approaches to estimating consensus labels of good quality. We compare our method with an EM algorithm-based approach, which is adopted to our method in the discussions of this paper, in terms of both efficiency of collecting labels from crowd and quality of labels obtained from the collected ones.
  • Keywords
    "Estimation","Crowdsourcing","Labeling","Machine learning algorithms","Computational complexity","Mathematical model","Proposals"
  • Publisher
    ieee
  • Conference_Titel
    Computer Software and Applications Conference (COMPSAC), 2015 IEEE 39th Annual
  • Electronic_ISBN
    0730-3157
  • Type

    conf

  • DOI
    10.1109/COMPSAC.2015.117
  • Filename
    7273702