Loading…

Leveraging non-expert crowdsourcing workers for improper task detection in crowdsourcing marketplaces

•We address the problem of improper task detection in crowdsourcing.•We investigate the operational data inside a commercial crowdsourcing marketplace.•We show machine learning is effective in detecting improper tasks in crowdsourcing.•We show the use of both expert and non-expert improves the perfo...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems with applications 2014-05, Vol.41 (6), p.2678-2687
Main Authors: Baba, Yukino, Kashima, Hisashi, Kinoshita, Kei, Yamaguchi, Goushi, Akiyoshi, Yosuke
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c462t-f6c91bd8639db4c2b7f9316291bb44f7a3a263eada3320df5893289b2facbcc13
cites cdi_FETCH-LOGICAL-c462t-f6c91bd8639db4c2b7f9316291bb44f7a3a263eada3320df5893289b2facbcc13
container_end_page 2687
container_issue 6
container_start_page 2678
container_title Expert systems with applications
container_volume 41
creator Baba, Yukino
Kashima, Hisashi
Kinoshita, Kei
Yamaguchi, Goushi
Akiyoshi, Yosuke
description •We address the problem of improper task detection in crowdsourcing.•We investigate the operational data inside a commercial crowdsourcing marketplace.•We show machine learning is effective in detecting improper tasks in crowdsourcing.•We show the use of both expert and non-expert improves the performance of detecting. Controlling the quality of tasks, i.e., propriety of posted jobs, is a major challenge in crowdsourcing marketplaces. Most existing crowdsourcing services prohibit requesters from posting illegal or objectionable tasks. Operators in marketplaces have to monitor tasks continuously to find such improper ones; however, it is very expensive to manually investigate each task. In this paper, we present the results of our trial study on automatic detection of improper tasks to support the monitoring of activities by marketplace operators. We performed experiments using real task data from a commercial crowdsourcing marketplace and showed that the classifier trained by the operators’ judgments achieves a high performance in detecting improper tasks. By analyzing the estimated classifier, we observed several effective features for detecting improper tasks, such as the words appeared in the task information, the amount of money that each worker will receive for the task, and the type of worker qualification option set for a task. In addition, to reduce the annotation costs of the operators and improve classification performance, we considered the use of crowdsourcing for task annotation. We hired a group of crowdsourcing (non-expert) workers to monitor posted tasks and use their judgments to train the classifier. We were able to confirm that applying quality control techniques is beneficial for handling the variability in worker reliability and that it improved the performance of the classifier. Finally, our results showed that the use of non-expert judgments of crowdsourcing workers in combination with expert judgments improves the performance of detecting improper crowdsourcing tasks, and that the use of crowdsourced labels allows a reduction in the required number of expert judgments by 25% while maintaining the level of detection performance.
doi_str_mv 10.1016/j.eswa.2013.11.011
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_1701121840</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0957417413009226</els_id><sourcerecordid>1531032824</sourcerecordid><originalsourceid>FETCH-LOGICAL-c462t-f6c91bd8639db4c2b7f9316291bb44f7a3a263eada3320df5893289b2facbcc13</originalsourceid><addsrcrecordid>eNqFkD9PwzAQxS0EEqXwBZiyILEk-OwkTiQWhPgnVWKB2XKcc-U2jYudtvDtcdSKgQGmk3y_987vEXIJNAMK5c0iw7BTGaPAM4CMAhyRCVSCp6Wo-TGZ0LoQaQ4iPyVnISwoBUGpmBCc4Ra9mtt-nvSuT_FzjX5ItHe7NriN1-Ni5_wSfUiM84ldrb2LTDKosExaHFAP1vWJ7X-JViqKhnWnNIZzcmJUF_DiMKfk_fHh7f45nb0-vdzfzVKdl2xITalraNqq5HXb5Jo1wtQcShYfmzw3QnHFSo6qVZwz2pqiqjmr6oYZpRutgU_J9d43_vFjg2GQKxs0dp3q0W2CjKEBGFQ5_R8tONDozvKIsj0aA4bg0ci1tzHelwQqx_rlQo71y7F-CSDjkSi6OviroFVnvOq1DT_KaFyXVBSRu91zGHvZWvQyaIu9xtb6WK1snf3rzDeku52Z</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1531032824</pqid></control><display><type>article</type><title>Leveraging non-expert crowdsourcing workers for improper task detection in crowdsourcing marketplaces</title><source>Elsevier</source><creator>Baba, Yukino ; Kashima, Hisashi ; Kinoshita, Kei ; Yamaguchi, Goushi ; Akiyoshi, Yosuke</creator><creatorcontrib>Baba, Yukino ; Kashima, Hisashi ; Kinoshita, Kei ; Yamaguchi, Goushi ; Akiyoshi, Yosuke</creatorcontrib><description>•We address the problem of improper task detection in crowdsourcing.•We investigate the operational data inside a commercial crowdsourcing marketplace.•We show machine learning is effective in detecting improper tasks in crowdsourcing.•We show the use of both expert and non-expert improves the performance of detecting. Controlling the quality of tasks, i.e., propriety of posted jobs, is a major challenge in crowdsourcing marketplaces. Most existing crowdsourcing services prohibit requesters from posting illegal or objectionable tasks. Operators in marketplaces have to monitor tasks continuously to find such improper ones; however, it is very expensive to manually investigate each task. In this paper, we present the results of our trial study on automatic detection of improper tasks to support the monitoring of activities by marketplace operators. We performed experiments using real task data from a commercial crowdsourcing marketplace and showed that the classifier trained by the operators’ judgments achieves a high performance in detecting improper tasks. By analyzing the estimated classifier, we observed several effective features for detecting improper tasks, such as the words appeared in the task information, the amount of money that each worker will receive for the task, and the type of worker qualification option set for a task. In addition, to reduce the annotation costs of the operators and improve classification performance, we considered the use of crowdsourcing for task annotation. We hired a group of crowdsourcing (non-expert) workers to monitor posted tasks and use their judgments to train the classifier. We were able to confirm that applying quality control techniques is beneficial for handling the variability in worker reliability and that it improved the performance of the classifier. Finally, our results showed that the use of non-expert judgments of crowdsourcing workers in combination with expert judgments improves the performance of detecting improper crowdsourcing tasks, and that the use of crowdsourced labels allows a reduction in the required number of expert judgments by 25% while maintaining the level of detection performance.</description><identifier>ISSN: 0957-4174</identifier><identifier>EISSN: 1873-6793</identifier><identifier>DOI: 10.1016/j.eswa.2013.11.011</identifier><language>eng</language><publisher>Amsterdam: Elsevier Ltd</publisher><subject>Applications ; Applied sciences ; Classifiers ; Computer science; control theory; systems ; Computer systems and distributed systems. User interface ; Crowdsourcing ; Exact sciences and technology ; Expert systems ; Illegal ; Improper task detection ; Information systems. Data bases ; Judgments ; Machine learning ; Mathematics ; Memory organisation. Data processing ; Monitoring ; Monitors ; Operational research and scientific management ; Operational research. Management science ; Operators ; Probability and statistics ; Quality control ; Reliability theory. Replacement problems ; Reliability, life testing, quality control ; Sciences and techniques of general use ; Software ; Statistics ; Tasks</subject><ispartof>Expert systems with applications, 2014-05, Vol.41 (6), p.2678-2687</ispartof><rights>2013 Elsevier Ltd</rights><rights>2015 INIST-CNRS</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c462t-f6c91bd8639db4c2b7f9316291bb44f7a3a263eada3320df5893289b2facbcc13</citedby><cites>FETCH-LOGICAL-c462t-f6c91bd8639db4c2b7f9316291bb44f7a3a263eada3320df5893289b2facbcc13</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27903,27904</link.rule.ids><backlink>$$Uhttp://pascal-francis.inist.fr/vibad/index.php?action=getRecordDetail&amp;idt=28296075$$DView record in Pascal Francis$$Hfree_for_read</backlink></links><search><creatorcontrib>Baba, Yukino</creatorcontrib><creatorcontrib>Kashima, Hisashi</creatorcontrib><creatorcontrib>Kinoshita, Kei</creatorcontrib><creatorcontrib>Yamaguchi, Goushi</creatorcontrib><creatorcontrib>Akiyoshi, Yosuke</creatorcontrib><title>Leveraging non-expert crowdsourcing workers for improper task detection in crowdsourcing marketplaces</title><title>Expert systems with applications</title><description>•We address the problem of improper task detection in crowdsourcing.•We investigate the operational data inside a commercial crowdsourcing marketplace.•We show machine learning is effective in detecting improper tasks in crowdsourcing.•We show the use of both expert and non-expert improves the performance of detecting. Controlling the quality of tasks, i.e., propriety of posted jobs, is a major challenge in crowdsourcing marketplaces. Most existing crowdsourcing services prohibit requesters from posting illegal or objectionable tasks. Operators in marketplaces have to monitor tasks continuously to find such improper ones; however, it is very expensive to manually investigate each task. In this paper, we present the results of our trial study on automatic detection of improper tasks to support the monitoring of activities by marketplace operators. We performed experiments using real task data from a commercial crowdsourcing marketplace and showed that the classifier trained by the operators’ judgments achieves a high performance in detecting improper tasks. By analyzing the estimated classifier, we observed several effective features for detecting improper tasks, such as the words appeared in the task information, the amount of money that each worker will receive for the task, and the type of worker qualification option set for a task. In addition, to reduce the annotation costs of the operators and improve classification performance, we considered the use of crowdsourcing for task annotation. We hired a group of crowdsourcing (non-expert) workers to monitor posted tasks and use their judgments to train the classifier. We were able to confirm that applying quality control techniques is beneficial for handling the variability in worker reliability and that it improved the performance of the classifier. Finally, our results showed that the use of non-expert judgments of crowdsourcing workers in combination with expert judgments improves the performance of detecting improper crowdsourcing tasks, and that the use of crowdsourced labels allows a reduction in the required number of expert judgments by 25% while maintaining the level of detection performance.</description><subject>Applications</subject><subject>Applied sciences</subject><subject>Classifiers</subject><subject>Computer science; control theory; systems</subject><subject>Computer systems and distributed systems. User interface</subject><subject>Crowdsourcing</subject><subject>Exact sciences and technology</subject><subject>Expert systems</subject><subject>Illegal</subject><subject>Improper task detection</subject><subject>Information systems. Data bases</subject><subject>Judgments</subject><subject>Machine learning</subject><subject>Mathematics</subject><subject>Memory organisation. Data processing</subject><subject>Monitoring</subject><subject>Monitors</subject><subject>Operational research and scientific management</subject><subject>Operational research. Management science</subject><subject>Operators</subject><subject>Probability and statistics</subject><subject>Quality control</subject><subject>Reliability theory. Replacement problems</subject><subject>Reliability, life testing, quality control</subject><subject>Sciences and techniques of general use</subject><subject>Software</subject><subject>Statistics</subject><subject>Tasks</subject><issn>0957-4174</issn><issn>1873-6793</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2014</creationdate><recordtype>article</recordtype><recordid>eNqFkD9PwzAQxS0EEqXwBZiyILEk-OwkTiQWhPgnVWKB2XKcc-U2jYudtvDtcdSKgQGmk3y_987vEXIJNAMK5c0iw7BTGaPAM4CMAhyRCVSCp6Wo-TGZ0LoQaQ4iPyVnISwoBUGpmBCc4Ra9mtt-nvSuT_FzjX5ItHe7NriN1-Ni5_wSfUiM84ldrb2LTDKosExaHFAP1vWJ7X-JViqKhnWnNIZzcmJUF_DiMKfk_fHh7f45nb0-vdzfzVKdl2xITalraNqq5HXb5Jo1wtQcShYfmzw3QnHFSo6qVZwz2pqiqjmr6oYZpRutgU_J9d43_vFjg2GQKxs0dp3q0W2CjKEBGFQ5_R8tONDozvKIsj0aA4bg0ci1tzHelwQqx_rlQo71y7F-CSDjkSi6OviroFVnvOq1DT_KaFyXVBSRu91zGHvZWvQyaIu9xtb6WK1snf3rzDeku52Z</recordid><startdate>20140501</startdate><enddate>20140501</enddate><creator>Baba, Yukino</creator><creator>Kashima, Hisashi</creator><creator>Kinoshita, Kei</creator><creator>Yamaguchi, Goushi</creator><creator>Akiyoshi, Yosuke</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>IQODW</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20140501</creationdate><title>Leveraging non-expert crowdsourcing workers for improper task detection in crowdsourcing marketplaces</title><author>Baba, Yukino ; Kashima, Hisashi ; Kinoshita, Kei ; Yamaguchi, Goushi ; Akiyoshi, Yosuke</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c462t-f6c91bd8639db4c2b7f9316291bb44f7a3a263eada3320df5893289b2facbcc13</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2014</creationdate><topic>Applications</topic><topic>Applied sciences</topic><topic>Classifiers</topic><topic>Computer science; control theory; systems</topic><topic>Computer systems and distributed systems. User interface</topic><topic>Crowdsourcing</topic><topic>Exact sciences and technology</topic><topic>Expert systems</topic><topic>Illegal</topic><topic>Improper task detection</topic><topic>Information systems. Data bases</topic><topic>Judgments</topic><topic>Machine learning</topic><topic>Mathematics</topic><topic>Memory organisation. Data processing</topic><topic>Monitoring</topic><topic>Monitors</topic><topic>Operational research and scientific management</topic><topic>Operational research. Management science</topic><topic>Operators</topic><topic>Probability and statistics</topic><topic>Quality control</topic><topic>Reliability theory. Replacement problems</topic><topic>Reliability, life testing, quality control</topic><topic>Sciences and techniques of general use</topic><topic>Software</topic><topic>Statistics</topic><topic>Tasks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Baba, Yukino</creatorcontrib><creatorcontrib>Kashima, Hisashi</creatorcontrib><creatorcontrib>Kinoshita, Kei</creatorcontrib><creatorcontrib>Yamaguchi, Goushi</creatorcontrib><creatorcontrib>Akiyoshi, Yosuke</creatorcontrib><collection>Pascal-Francis</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Expert systems with applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Baba, Yukino</au><au>Kashima, Hisashi</au><au>Kinoshita, Kei</au><au>Yamaguchi, Goushi</au><au>Akiyoshi, Yosuke</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Leveraging non-expert crowdsourcing workers for improper task detection in crowdsourcing marketplaces</atitle><jtitle>Expert systems with applications</jtitle><date>2014-05-01</date><risdate>2014</risdate><volume>41</volume><issue>6</issue><spage>2678</spage><epage>2687</epage><pages>2678-2687</pages><issn>0957-4174</issn><eissn>1873-6793</eissn><abstract>•We address the problem of improper task detection in crowdsourcing.•We investigate the operational data inside a commercial crowdsourcing marketplace.•We show machine learning is effective in detecting improper tasks in crowdsourcing.•We show the use of both expert and non-expert improves the performance of detecting. Controlling the quality of tasks, i.e., propriety of posted jobs, is a major challenge in crowdsourcing marketplaces. Most existing crowdsourcing services prohibit requesters from posting illegal or objectionable tasks. Operators in marketplaces have to monitor tasks continuously to find such improper ones; however, it is very expensive to manually investigate each task. In this paper, we present the results of our trial study on automatic detection of improper tasks to support the monitoring of activities by marketplace operators. We performed experiments using real task data from a commercial crowdsourcing marketplace and showed that the classifier trained by the operators’ judgments achieves a high performance in detecting improper tasks. By analyzing the estimated classifier, we observed several effective features for detecting improper tasks, such as the words appeared in the task information, the amount of money that each worker will receive for the task, and the type of worker qualification option set for a task. In addition, to reduce the annotation costs of the operators and improve classification performance, we considered the use of crowdsourcing for task annotation. We hired a group of crowdsourcing (non-expert) workers to monitor posted tasks and use their judgments to train the classifier. We were able to confirm that applying quality control techniques is beneficial for handling the variability in worker reliability and that it improved the performance of the classifier. Finally, our results showed that the use of non-expert judgments of crowdsourcing workers in combination with expert judgments improves the performance of detecting improper crowdsourcing tasks, and that the use of crowdsourced labels allows a reduction in the required number of expert judgments by 25% while maintaining the level of detection performance.</abstract><cop>Amsterdam</cop><pub>Elsevier Ltd</pub><doi>10.1016/j.eswa.2013.11.011</doi><tpages>10</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0957-4174
ispartof Expert systems with applications, 2014-05, Vol.41 (6), p.2678-2687
issn 0957-4174
1873-6793
language eng
recordid cdi_proquest_miscellaneous_1701121840
source Elsevier
subjects Applications
Applied sciences
Classifiers
Computer science
control theory
systems
Computer systems and distributed systems. User interface
Crowdsourcing
Exact sciences and technology
Expert systems
Illegal
Improper task detection
Information systems. Data bases
Judgments
Machine learning
Mathematics
Memory organisation. Data processing
Monitoring
Monitors
Operational research and scientific management
Operational research. Management science
Operators
Probability and statistics
Quality control
Reliability theory. Replacement problems
Reliability, life testing, quality control
Sciences and techniques of general use
Software
Statistics
Tasks
title Leveraging non-expert crowdsourcing workers for improper task detection in crowdsourcing marketplaces
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-23T02%3A43%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Leveraging%20non-expert%20crowdsourcing%20workers%20for%20improper%20task%20detection%20in%20crowdsourcing%20marketplaces&rft.jtitle=Expert%20systems%20with%20applications&rft.au=Baba,%20Yukino&rft.date=2014-05-01&rft.volume=41&rft.issue=6&rft.spage=2678&rft.epage=2687&rft.pages=2678-2687&rft.issn=0957-4174&rft.eissn=1873-6793&rft_id=info:doi/10.1016/j.eswa.2013.11.011&rft_dat=%3Cproquest_cross%3E1531032824%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c462t-f6c91bd8639db4c2b7f9316291bb44f7a3a263eada3320df5893289b2facbcc13%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1531032824&rft_id=info:pmid/&rfr_iscdi=true