An exploratory analysis of multilingual word-level quality estimation with cross-lingual transformers
Abstract
Most studies on word-level Quality Estimation (QE) of machine translation focus on language-specific models. The obvious disadvantages of these approaches are the need for labelled data for each language pair and the high cost required to maintain several language-specific models. To overcome these problems, we explore different approaches to multilingual, word-level QE. We show that these QE models perform on par with the current language-specific models. In the cases of zero-shot and few-shot QE, we demonstrate that it is possible to accurately predict word-level quality for any given new language pair from models trained on other language pairs. Our findings suggest that the word-level QE models based on powerful pre-trained transformers that we propose in this paper generalise well across languages, making them more useful in real-world scenarios.Citation
Ranasinghe, T., Orasan, C. and Mitkov, R. (2021) An exploratory analysis of multilingual word-level quality estimation with cross-lingual transformers. Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Short Papers), pages 434–440 August 1–6, 2021.Additional Links
https://aclanthology.org/2021.acl-short.55/Type
Conference contributionLanguage
enDescription
© 2021 The Authors. Published by Association for Computational Linguistics. This is an open access article available under a Creative Commons licence. The published version can be accessed at the following link on the publisher’s website: https://aclanthology.org/2021.acl-short.55ISBN
9781954085534ae974a485f413a2113503eed53cd6c53
10.18653/v1/2021.acl-short.55
Scopus Count
Except where otherwise noted, this item's license is described as https://creativecommons.org/licenses/by/4.0/