Quality assessment in crowdsourced classification tasks
Purpose – Ensuring quality is one of the most significant challenges in microtask crowdsourcing tasks. Aggregation of the collected data from the crowd is one of the important steps to infer the correct answer, but the existing study seems to be limited to the single-step task. This study aims to lo...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Emerald Publishing
2019-12-01
|
Series: | International Journal of Crowd Science |
Subjects: | |
Online Access: | https://www.emerald.com/insight/content/doi/10.1108/IJCS-06-2019-0017/full/pdf?title=quality-assessment-in-crowdsourced-classification-tasks |
id |
doaj-8f8564fd1971401f9beb50132c0f8ddf |
---|---|
record_format |
Article |
spelling |
doaj-8f8564fd1971401f9beb50132c0f8ddf2021-07-30T16:00:24ZengEmerald PublishingInternational Journal of Crowd Science2398-72942019-12-013322224810.1108/IJCS-06-2019-0017634485Quality assessment in crowdsourced classification tasksQiong Bu0Elena Simperl1Adriane Chapman2Eddy Maddalena3School of Electronics and Computer Science, University of Southampton, Southampton, UKSchool of Electronics and Computer Science, University of Southampton, Southampton, UKSchool of Electronics and Computer Science, University of Southampton, Southampton, UKSchool of Electronics and Computer Science, University of Southampton, Southampton, UKPurpose – Ensuring quality is one of the most significant challenges in microtask crowdsourcing tasks. Aggregation of the collected data from the crowd is one of the important steps to infer the correct answer, but the existing study seems to be limited to the single-step task. This study aims to look at multiple-step classification tasks and understand aggregation in such cases; hence, it is useful for assessing the classification quality. Design/methodology/approach – The authors present a model to capture the information of the workflow, questions and answers for both single- and multiple-question classification tasks. They propose an adapted approach on top of the classic approach so that the model can handle tasks with several multiple-choice questions in general instead of a specific domain or any specific hierarchical classifications. They evaluate their approach with three representative tasks from existing citizen science projects in which they have the gold standard created by experts. Findings – The results show that the approach can provide significant improvements to the overall classification accuracy. The authors’ analysis also demonstrates that all algorithms can achieve higher accuracy for the volunteer- versus paid-generated data sets for the same task. Furthermore, the authors observed interesting patterns in the relationship between the performance of different algorithms and workflow-specific factors including the number of steps and the number of available options in each step. Originality/value – Due to the nature of crowdsourcing, aggregating the collected data is an important process to understand the quality of crowdsourcing results. Different inference algorithms have been studied for simple microtasks consisting of single questions with two or more answers. However, as classification tasks typically contain many questions, the proposed method can be applied to a wide range of tasks including both single- and multiple-question classification tasks.https://www.emerald.com/insight/content/doi/10.1108/IJCS-06-2019-0017/full/pdf?title=quality-assessment-in-crowdsourced-classification-tasksaggregationclassificationtask-orientated crowdsourcingquality assessment |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Qiong Bu Elena Simperl Adriane Chapman Eddy Maddalena |
spellingShingle |
Qiong Bu Elena Simperl Adriane Chapman Eddy Maddalena Quality assessment in crowdsourced classification tasks International Journal of Crowd Science aggregation classification task-orientated crowdsourcing quality assessment |
author_facet |
Qiong Bu Elena Simperl Adriane Chapman Eddy Maddalena |
author_sort |
Qiong Bu |
title |
Quality assessment in crowdsourced classification tasks |
title_short |
Quality assessment in crowdsourced classification tasks |
title_full |
Quality assessment in crowdsourced classification tasks |
title_fullStr |
Quality assessment in crowdsourced classification tasks |
title_full_unstemmed |
Quality assessment in crowdsourced classification tasks |
title_sort |
quality assessment in crowdsourced classification tasks |
publisher |
Emerald Publishing |
series |
International Journal of Crowd Science |
issn |
2398-7294 |
publishDate |
2019-12-01 |
description |
Purpose – Ensuring quality is one of the most significant challenges in microtask crowdsourcing tasks. Aggregation of the collected data from the crowd is one of the important steps to infer the correct answer, but the existing study seems to be limited to the single-step task. This study aims to look at multiple-step classification tasks and understand aggregation in such cases; hence, it is useful for assessing the classification quality. Design/methodology/approach – The authors present a model to capture the information of the workflow, questions and answers for both single- and multiple-question classification tasks. They propose an
adapted approach on top of the classic approach so that the model can handle tasks with several multiple-choice questions in general instead of a specific domain or any specific hierarchical classifications. They evaluate their approach with three representative tasks from existing citizen science projects in which they have the gold standard created by experts. Findings – The results show that the approach can provide significant improvements to the overall classification accuracy. The authors’ analysis also demonstrates that all algorithms can achieve higher accuracy for the volunteer- versus paid-generated data sets for the same task. Furthermore, the authors
observed interesting patterns in the relationship between the performance of different algorithms and workflow-specific factors including the number of steps and the number of available options in each step. Originality/value – Due to the nature of crowdsourcing, aggregating the collected data is an important process to understand the quality of crowdsourcing results. Different inference algorithms have been studied for simple microtasks consisting of single questions with two or more answers. However, as classification tasks typically contain many questions, the proposed method can be applied to a wide range of tasks including both single- and multiple-question classification tasks. |
topic |
aggregation classification task-orientated crowdsourcing quality assessment |
url |
https://www.emerald.com/insight/content/doi/10.1108/IJCS-06-2019-0017/full/pdf?title=quality-assessment-in-crowdsourced-classification-tasks |
work_keys_str_mv |
AT qiongbu qualityassessmentincrowdsourcedclassificationtasks AT elenasimperl qualityassessmentincrowdsourcedclassificationtasks AT adrianechapman qualityassessmentincrowdsourcedclassificationtasks AT eddymaddalena qualityassessmentincrowdsourcedclassificationtasks |
_version_ |
1721247465636626432 |