TY - GEN
T1 - TransQuest: Translation Quality Estimation with Cross-lingual Transformers
AU - Ranasinghe, Tharindu
AU - Orasan, Constantin
AU - Mitkov, Ruslan
N1 - ACL materials are Copyright © 1963–2023 ACL; Materials published in or after 2016 are licensed on a Creative Commons Attribution 4.0 International License.
PY - 2020/12
Y1 - 2020/12
N2 - Recent years have seen big advances in the field of sentence-level quality estimation (QE), largely as a result of using neural-based architectures. However, the majority of these methods work only on the language pair they are trained on and need retraining for new language pairs. This process can prove difficult from a technical point of view and is usually computationally expensive. In this paper we propose a simple QE framework based on cross-lingual transformers, and we use it to implement and evaluate two different neural architectures. Our evaluation shows that the proposed methods achieve state-of-the-art results outperforming current open-source quality estimation frameworks when trained on datasets from WMT. In addition, the framework proves very useful in transfer learning settings, especially when dealing with low-resourced languages, allowing us to obtain very competitive results.
AB - Recent years have seen big advances in the field of sentence-level quality estimation (QE), largely as a result of using neural-based architectures. However, the majority of these methods work only on the language pair they are trained on and need retraining for new language pairs. This process can prove difficult from a technical point of view and is usually computationally expensive. In this paper we propose a simple QE framework based on cross-lingual transformers, and we use it to implement and evaluate two different neural architectures. Our evaluation shows that the proposed methods achieve state-of-the-art results outperforming current open-source quality estimation frameworks when trained on datasets from WMT. In addition, the framework proves very useful in transfer learning settings, especially when dealing with low-resourced languages, allowing us to obtain very competitive results.
UR - https://www.aclweb.org/anthology/2020.coling-main.445
UR - https://aclanthology.org/2020.coling-main.445/
U2 - 10.18653/v1/2020.coling-main.445
DO - 10.18653/v1/2020.coling-main.445
M3 - Conference publication
SP - 5070
EP - 5081
BT - COLING 2020: The 28th International Conference on Computational Linguistics
ER -