6-17 - Microtask Crowdsourcing for Music Score Transcriptions: an Experiment with Error Detection
Ioannis Petros Samiotis, Sihang Qiu, Andrea Mauri, Cynthia C. S. Liem, Christoph Lofi, Alessandro Bozzon
Keywords: Domain knowledge, Representations of music, Evaluation, datasets, and reproducibility, Annotation protocols, Human-centered MIR, Human-computer interaction and interfaces, MIR fundamentals and methodology, Multimodality, MIR tasks, Music transcription and annotation, Optical Music Recognition (OMR)
Abstract:
Human annotation is still an essential part of modern transcription workflows for digitizing music scores, either as a standalone approach where a single expert annotator transcribes a complete score, or for supporting an automated Optical Music Recognition (OMR) system. Research on human computation has shown the effectiveness of crowdsourcing for scaling out human work by defining a large number of microtasks which can easily be distributed and executed. However, microtask design for music transcription is a research area that remains unaddressed. This paper focuses on the design of a crowdsourcing task to detect errors in a score transcription which can be deployed in either automated or human-driven transcription workflows. We conduct an experiment where we study two design parameters: 1) the size of the score to be annotated and 2) the modality in which it is presented in the user interface. We analyze the performance and reliability of non-specialised crowdworkers on Amazon Mechanical Turk with respect to these design parameters, differentiated by worker experience and types of transcription errors. Results are encouraging, and pave the way for scalable and efficient crowd-assisted music transcription systems.