The Impact of Task Abandonment in Crowdsourcing

Research output: Contribution to journalArticleResearchpeer review

Authors

  • Lei Han
  • Kevin Roitero
  • Ujwal Gadiraju
  • Cristina Sarasua
  • Alessandro Checco
  • Eddy Maddalena
  • Gianluca Demartini

Research Organisations

External Research Organisations

  • University of Queensland
  • University of Udine
  • Universität Zürich (UZH)
  • The University of Sheffield
  • University of Southampton
View graph of relations

Details

Original languageEnglish
Article number8873609
Pages (from-to)2266-2279
Number of pages14
JournalIEEE Transactions on Knowledge and Data Engineering
Volume33
Issue number5
Publication statusPublished - 17 Oct 2019

Abstract

Crowdsourcing has become a standard methodology to collect manually annotated data such as relevance judgments at scale. On crowdsourcing platforms like Amazon MTurk or FigureEight, crowd workers select tasks to work on based on different dimensions such as task reward and requester reputation. Requesters then receive the judgments of workers who self-selected into the tasks and completed them successfully. Several crowd workers, however, preview tasks, begin working on them, reaching varying stages of task completion without finally submitting their work. Such behavior results in unrewarded effort which remains invisible to requesters. In this paper, we conduct an investigation of the phenomenon of task abandonment, the act of workers previewing or beginning a task and deciding not to complete it. We follow a three-fold methodology which includes 1) investigating the prevalence and causes of task abandonment by means of a survey over different crowdsourcing platforms, 2) data-driven analysis of logs collected during a large-scale relevance judgment experiment, and 3) controlled experiments measuring the effect of different dimensions on abandonment. Our results show that task abandonment is a widely spread phenomenon. Apart from accounting for a considerable amount of wasted human effort, this bears important implications on the hourly wages of workers as they are not rewarded for tasks that they do not complete. We also show how task abandonment may have strong implications on the use of collected data (for example, on the evaluation of Information Retrieval systems).

Keywords

    crowdsourcing, relevance judgments, Task abandonment

ASJC Scopus subject areas

Cite this

The Impact of Task Abandonment in Crowdsourcing. / Han, Lei; Roitero, Kevin; Gadiraju, Ujwal et al.
In: IEEE Transactions on Knowledge and Data Engineering, Vol. 33, No. 5, 8873609, 17.10.2019, p. 2266-2279.

Research output: Contribution to journalArticleResearchpeer review

Han, L, Roitero, K, Gadiraju, U, Sarasua, C, Checco, A, Maddalena, E & Demartini, G 2019, 'The Impact of Task Abandonment in Crowdsourcing', IEEE Transactions on Knowledge and Data Engineering, vol. 33, no. 5, 8873609, pp. 2266-2279. https://doi.org/10.1109/TKDE.2019.2948168
Han, L., Roitero, K., Gadiraju, U., Sarasua, C., Checco, A., Maddalena, E., & Demartini, G. (2019). The Impact of Task Abandonment in Crowdsourcing. IEEE Transactions on Knowledge and Data Engineering, 33(5), 2266-2279. Article 8873609. https://doi.org/10.1109/TKDE.2019.2948168
Han L, Roitero K, Gadiraju U, Sarasua C, Checco A, Maddalena E et al. The Impact of Task Abandonment in Crowdsourcing. IEEE Transactions on Knowledge and Data Engineering. 2019 Oct 17;33(5):2266-2279. 8873609. doi: 10.1109/TKDE.2019.2948168
Han, Lei ; Roitero, Kevin ; Gadiraju, Ujwal et al. / The Impact of Task Abandonment in Crowdsourcing. In: IEEE Transactions on Knowledge and Data Engineering. 2019 ; Vol. 33, No. 5. pp. 2266-2279.
Download
@article{5026a909b6404d15a49793adfc509c8e,
title = "The Impact of Task Abandonment in Crowdsourcing",
abstract = "Crowdsourcing has become a standard methodology to collect manually annotated data such as relevance judgments at scale. On crowdsourcing platforms like Amazon MTurk or FigureEight, crowd workers select tasks to work on based on different dimensions such as task reward and requester reputation. Requesters then receive the judgments of workers who self-selected into the tasks and completed them successfully. Several crowd workers, however, preview tasks, begin working on them, reaching varying stages of task completion without finally submitting their work. Such behavior results in unrewarded effort which remains invisible to requesters. In this paper, we conduct an investigation of the phenomenon of task abandonment, the act of workers previewing or beginning a task and deciding not to complete it. We follow a three-fold methodology which includes 1) investigating the prevalence and causes of task abandonment by means of a survey over different crowdsourcing platforms, 2) data-driven analysis of logs collected during a large-scale relevance judgment experiment, and 3) controlled experiments measuring the effect of different dimensions on abandonment. Our results show that task abandonment is a widely spread phenomenon. Apart from accounting for a considerable amount of wasted human effort, this bears important implications on the hourly wages of workers as they are not rewarded for tasks that they do not complete. We also show how task abandonment may have strong implications on the use of collected data (for example, on the evaluation of Information Retrieval systems). ",
keywords = "crowdsourcing, relevance judgments, Task abandonment",
author = "Lei Han and Kevin Roitero and Ujwal Gadiraju and Cristina Sarasua and Alessandro Checco and Eddy Maddalena and Gianluca Demartini",
note = "Funding information: This work is supported in part by the EU{\textquoteright}s H2020 research and innovation programme under Grant Agreement No. 732328, the Erasmus+ project DISKOW (Project No. 60171990), and by the ARC Discovery Project under Grant No. DP190102141.",
year = "2019",
month = oct,
day = "17",
doi = "10.1109/TKDE.2019.2948168",
language = "English",
volume = "33",
pages = "2266--2279",
journal = "IEEE Transactions on Knowledge and Data Engineering",
issn = "1041-4347",
publisher = "IEEE Computer Society",
number = "5",

}

Download

TY - JOUR

T1 - The Impact of Task Abandonment in Crowdsourcing

AU - Han, Lei

AU - Roitero, Kevin

AU - Gadiraju, Ujwal

AU - Sarasua, Cristina

AU - Checco, Alessandro

AU - Maddalena, Eddy

AU - Demartini, Gianluca

N1 - Funding information: This work is supported in part by the EU’s H2020 research and innovation programme under Grant Agreement No. 732328, the Erasmus+ project DISKOW (Project No. 60171990), and by the ARC Discovery Project under Grant No. DP190102141.

PY - 2019/10/17

Y1 - 2019/10/17

N2 - Crowdsourcing has become a standard methodology to collect manually annotated data such as relevance judgments at scale. On crowdsourcing platforms like Amazon MTurk or FigureEight, crowd workers select tasks to work on based on different dimensions such as task reward and requester reputation. Requesters then receive the judgments of workers who self-selected into the tasks and completed them successfully. Several crowd workers, however, preview tasks, begin working on them, reaching varying stages of task completion without finally submitting their work. Such behavior results in unrewarded effort which remains invisible to requesters. In this paper, we conduct an investigation of the phenomenon of task abandonment, the act of workers previewing or beginning a task and deciding not to complete it. We follow a three-fold methodology which includes 1) investigating the prevalence and causes of task abandonment by means of a survey over different crowdsourcing platforms, 2) data-driven analysis of logs collected during a large-scale relevance judgment experiment, and 3) controlled experiments measuring the effect of different dimensions on abandonment. Our results show that task abandonment is a widely spread phenomenon. Apart from accounting for a considerable amount of wasted human effort, this bears important implications on the hourly wages of workers as they are not rewarded for tasks that they do not complete. We also show how task abandonment may have strong implications on the use of collected data (for example, on the evaluation of Information Retrieval systems).

AB - Crowdsourcing has become a standard methodology to collect manually annotated data such as relevance judgments at scale. On crowdsourcing platforms like Amazon MTurk or FigureEight, crowd workers select tasks to work on based on different dimensions such as task reward and requester reputation. Requesters then receive the judgments of workers who self-selected into the tasks and completed them successfully. Several crowd workers, however, preview tasks, begin working on them, reaching varying stages of task completion without finally submitting their work. Such behavior results in unrewarded effort which remains invisible to requesters. In this paper, we conduct an investigation of the phenomenon of task abandonment, the act of workers previewing or beginning a task and deciding not to complete it. We follow a three-fold methodology which includes 1) investigating the prevalence and causes of task abandonment by means of a survey over different crowdsourcing platforms, 2) data-driven analysis of logs collected during a large-scale relevance judgment experiment, and 3) controlled experiments measuring the effect of different dimensions on abandonment. Our results show that task abandonment is a widely spread phenomenon. Apart from accounting for a considerable amount of wasted human effort, this bears important implications on the hourly wages of workers as they are not rewarded for tasks that they do not complete. We also show how task abandonment may have strong implications on the use of collected data (for example, on the evaluation of Information Retrieval systems).

KW - crowdsourcing

KW - relevance judgments

KW - Task abandonment

UR - http://www.scopus.com/inward/record.url?scp=85077968612&partnerID=8YFLogxK

U2 - 10.1109/TKDE.2019.2948168

DO - 10.1109/TKDE.2019.2948168

M3 - Article

AN - SCOPUS:85077968612

VL - 33

SP - 2266

EP - 2279

JO - IEEE Transactions on Knowledge and Data Engineering

JF - IEEE Transactions on Knowledge and Data Engineering

SN - 1041-4347

IS - 5

M1 - 8873609

ER -