Show simple item record

dc.contributor.authorPAUN, Sen_US
dc.contributor.authorCarpenter, Ben_US
dc.contributor.authorChamberlain, Jen_US
dc.contributor.authorHovy, Den_US
dc.contributor.authorKruschwitz, Uen_US
dc.contributor.authorPOESIO, Men_US
dc.date.accessioned2019-02-04T15:31:46Z
dc.date.available2018-11-30en_US
dc.date.submitted2018-12-07T17:15:42.808Z
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/55140
dc.description.abstractThe analysis of crowdsourced annotations in NLP is concerned with identifying 1) gold standard labels, 2) annotator accuracies and biases, and 3) item difficulties and error patterns. Traditionally, majority voting was used for 1), and coefficients of agreement for 2) and 3). Lately, model-based analy- sis of corpus annotations have proven better at all three tasks. But there has been rel- atively little work comparing them on the same datasets. This paper aims to fill this gap by analyzing six models of annotation, covering different approaches to annotator ability, item difficulty, and parameter pool- ing (tying) across annotators and items. We evaluate these models along four aspects: comparison to gold labels, predictive accu- racy for new annotations, annotator char- acterization, and item difficulty, using four datasets with varying degrees of noise in the form of random (spammy) annotators. We conclude with guidelines for model selec- tion, application, and implementation.en_US
dc.languageEnglishen_US
dc.publisherAssociation for Computational Linguistics and MIT Pressen_US
dc.relation.ispartofTransactions of the Association for Computational Linguisticsen_US
dc.rightsThis is a pre-copyedited, author-produced version of an article accepted for publication in Transactions of the Association for Computational Linguistics following peer review.
dc.subjectprobabilistic annotation modelsen_US
dc.subjectcomputational linguisticsen_US
dc.subjectcrowdsourcingen_US
dc.titleComparing Bayesian Models of Annotationen_US
dc.typeArticle
dc.rights.holder© 2019 Association for Computational Linguistics and MIT Press
pubs.notesNo embargoen_US
pubs.publication-statusAccepteden_US
pubs.publisher-urlhttps://www.mitpressjournals.org/loi/taclen_US
dcterms.dateAccepted2018-11-30en_US
rioxxterms.funderDefault funderen_US
rioxxterms.identifier.projectDefault projecten_US
qmul.funderDisagreements in Language Interpretation (DALI)::European Research Councilen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record