Efficient Crowdsourcing via Bayes Quality Control

Article Preview

Abstract:

Crowdsourcing has become an important tool to aggregate the wisdom of the crowd in this Internet age. A central problem in building an online crowdsourcing system is to determine the appropriate number of workers to assign tasks to. We study this problem by formulating a Bayes Net framework and introduce a quantity derived from posterior distribution to measure the convergence of crowd opinions. Using this quantity, our algorithm could stop soliciting opinions from more workers if the distribution of opinions is unlikely to change in future predictions. We empirically demonstrate the effectiveness of the designed strategy by building a challenging fine-grained image annotation task on Amazon Mechanical Turk. Experiment results show that our approach not only saves annotation cost but also guarantees high annotation quality.

You might also be interested in these eBooks

Info:

Periodical:

Advanced Materials Research (Volumes 998-999)

Pages:

1576-1580

Citation:

Online since:

July 2014

Authors:

Export:

Price:

Permissions CCC:

Permissions PLS:

Сopyright:

© 2014 Trans Tech Publications Ltd. All Rights Reserved

Share:

Citation:

* - Corresponding Author

[1] Doan, A; Ramarkrishnan, R; Halevy, A (2011), Crowdsourcing Systems on the World Wide Web, Communications of the ACM 54 (4): 86–96.

DOI: 10.1145/1924421.1924442

Google Scholar

[2] Rulison, Larry (February 14, 2012). A Winning App? They Hope So,. Times Union (Albany).

Google Scholar

[3] Deng, J.; Dong, W.; Socher, R.; Li, L. -J.; Li, K.; and Fei-Fei, L. 2009. ImageNet: A Large-Scale Hierarchical ImageDatabase. In CVPR09.

DOI: 10.1109/cvpr.2009.5206848

Google Scholar

[4] Sorokin, A., and Forsyth, D. 2008. Utility data annotationwith amazon mechanical turk. 2008 IEEE Computer SocietyConference on Computer Vision and Pattern RecognitionWorkshops 51(c): 1–8.

DOI: 10.1109/cvprw.2008.4562953

Google Scholar

[5] P. Welinder, P. P. 2010. Online crowdsourcing: rating annotatorsand obtaining cost-effective labels. In CVPR.

Google Scholar

[6] Welinder, P.; Branson, S.; Belongie, S.; and Perona, P. 2010. The multidimensional wisdom of crowds. In NIPS, 2424–2432.

Google Scholar

[7] Whitehill, J.; Ruvolo, P.; Wu, T.; Bergsma, J.; and Movellan,J. R. 2009. Whose vote should count more: Optimal integrationof labels from labelers of unknown expertise. In NIPS, 2035–(2043).

Google Scholar

[8] Branson, S.; Perona, P.; and Belongie, S. 2011. Strong supervisionfrom weak annotation: Interactive training of deformablepart models. In ICCV, 1832–1839.

DOI: 10.1109/iccv.2011.6126450

Google Scholar

[9] Vijayanarasimhan, S., and Grauman, K. 2011. Large-scalelive active learning: Training object detectors with crawleddata and crowds. In CVPR, 1449–1456.

DOI: 10.1109/cvpr.2011.5995430

Google Scholar