Please use this identifier to cite or link to this item:
Volltext verfügbar? / Dokumentlieferung
doi:10.22028/D291-38851
Title: | Design Choices in Crowdsourcing Discourse Relation Annotations: The Effect of Worker Selection and Training |
Author(s): | Scholman, Merel Cleo Johanna Pyatkin, Valentina Yung, Frances Pikyu Dagan, Ido Tsarfaty, Reut Demberg, Vera |
Editor(s): | Calzolari, Nicoletta |
Language: | English |
Title: | Language Resources and Evaluation Conference, LREC 2022, 20-25 June 2022 : Palais du Pharo, Marseille, France : conference proceedings |
Pages: | 2148-2156 |
Publisher/Platform: | European Language Resources Association |
Year of Publication: | 2022 |
Place of publication: | Paris |
Place of the conference: | Marseille, France |
Free key words: | discourse annotations crowdsourcing training participant selection |
DDC notations: | 004 Computer science, internet 400 Language, linguistics |
Publikation type: | Conference Paper |
Abstract: | Obtaining linguistic annotation from novice crowdworkers is far from trivial. A case in point is the annotation of discourse relations, which is a complicated task. Recent methods have obtained promising results by extracting relation labels from either discourse connectives (DCs) or question-answer (QA) pairs that participants provide. The current contribution studies the effect of worker selection and training on the agreement on implicit relation labels between workers and gold labels, for both the DC and the QA method. In Study 1, workers were not specifically selected or trained, and the results show that there is much room for improvement. Study 2 shows that a combination of selection and training does lead to improved results, but the method is cost- and time-intensive. Study 3 shows that a selection-only approach is a viable alternative; it results in annotations of comparable quality compared to annotations from trained participants. The results generalized over both the DC and QA method and therefore indicate that a selection-only approach could also be effective for other crowdsourced discourse annotation tasks. |
URL of the first publication: | https://aclanthology.org/2022.lrec-1.231/ |
Link to this record: | urn:nbn:de:bsz:291--ds-388514 hdl:20.500.11880/35058 http://dx.doi.org/10.22028/D291-38851 |
ISBN: | 979-10-95546-72-6 |
Date of registration: | 31-Jan-2023 |
Faculty: | MI - Fakultät für Mathematik und Informatik |
Department: | MI - Informatik |
Professorship: | MI - Prof. Dr. Vera Demberg |
Collections: | SciDok - Der Wissenschaftsserver der Universität des Saarlandes |
Files for this record:
There are no files associated with this item.
Items in SciDok are protected by copyright, with all rights reserved, unless otherwise indicated.