Show simple item record

dc.contributor.authorUlčar, M
dc.contributor.authorŽagar, A
dc.contributor.authorArmendariz, CS
dc.contributor.authorRepar, A
dc.contributor.authorPollak, S
dc.contributor.authorPurver, M
dc.contributor.authorRobnik-Šikonja, M
dc.date.accessioned2021-08-20T13:12:30Z
dc.date.available2021-08-20T13:12:30Z
dc.date.issued2021
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/73681
dc.description.abstractThe current dominance of deep neural networks in natural language processing is based on contextual embeddings such as ELMo, BERT, and BERT derivatives. Most existing work focuses on English; in contrast, we present here the first multilingual empirical comparison of two ELMo and several monolingual and multilingual BERT models using 14 tasks in nine languages. In monolingual settings, our analysis shows that monolingual BERT models generally dominate, with a few exceptions such as the dependency parsing task, where they are not competitive with ELMo models trained on large corpora. In cross-lingual settings, BERT models trained on only a few languages mostly do best, closely followed by massively multilingual BERT models.en_US
dc.rightsThis article is distributed under the terms of the CC-BY-SA Licence. Attribution — You must give appropriate credit, provide a link to the license, and indicate if changes were made. You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use. ShareAlike — If you remix, transform, or build upon the material, you must distribute your contributions under the same license as the original. No additional restrictions — You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits.
dc.subjectcs.CLen_US
dc.subjectcs.CLen_US
dc.titleEvaluation of contextual embeddings on less-resourced languagesen_US
dc.typeArticleen_US
dc.rights.holder© 2021, The Author(s)
pubs.author-urlhttp://arxiv.org/abs/2107.10614v1en_US
pubs.notesNot knownen_US
rioxxterms.funderDefault funderen_US
rioxxterms.identifier.projectDefault projecten_US
qmul.funderEMBEDDIA: Cross-Lingual Embeddings for Less-Represented Languages in European News Media::European Commissionen_US
qmul.funderEMBEDDIA: Cross-Lingual Embeddings for Less-Represented Languages in European News Media::European Commissionen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record