Téléchargement | - Voir la version finale : Machine translation reference-less evaluation using YiSi-2 with bilingual mappings of massive multilingual language model (PDF, 689 Kio)
|
---|
Auteur | Rechercher : Lo, Chi-Kiu1; Rechercher : Larkin, Samuel1 |
---|
Affiliation | - Conseil national de recherches du Canada. Technologies numériques
|
---|
Format | Texte, Article |
---|
Conférence | 5th Conference on Machine Translation (WMT 2020), November 19-20, 2020, (Held Online) |
---|
Résumé | We present a study on using YiSi-2 with massive multilingual pretrained language models for machine translation (MT) reference-less evaluation. Aiming at finding better semantic representation for semantic MT evaluation, we first test YiSi-2 with contextual embeddings extracted from different layers of two different pretrained models, multilingual BERT and XLM-RoBERTa. We also experiment with learning bilingual mappings that transform the vector subspace of the source language to be closer to that of the target language in the pretrained model to obtain more accurate cross-lingual semantic similarity representations. Our results show that YiSi-2's correlation with human direct assessment on translation quality is greatly improved by replacing multilingual BERT with XLM-RoBERTa and projecting the source embeddings into the target embedding space using a cross-lingual linear projection (CLP) matrix learnt from a small development set. |
---|
Date de publication | 2020-11-20 |
---|
Maison d’édition | Association for Computational Linguistics (ACL) |
---|
Licence | |
---|
Dans | |
---|
Langue | anglais |
---|
Publications évaluées par des pairs | Oui |
---|
Identificateur | 2020.wmt-1.100 |
---|
Exporter la notice | Exporter en format RIS |
---|
Signaler une correction | Signaler une correction (s'ouvre dans un nouvel onglet) |
---|
Identificateur de l’enregistrement | 635c72b8-0570-44bf-862f-cb8265ea2a0a |
---|
Enregistrement créé | 2022-07-14 |
---|
Enregistrement modifié | 2022-07-15 |
---|