Incorporating Pre-ordering Representations for Low-resource Neural Machine Translation

Yuan Gao, Feng Hou, Ruili Wang

Research output: Chapter in Book/Conference proceedingConference contributionpeer-review

Abstract

Pre-ordering, a data pre-processing technique that rearranges the order of source language words to align with the target language's natural word order, has been proven to be effective in improving machine translation performance, particularly for low-resource language pairs. Existing research, however, primarily exploits the position embeddings of tokens in the pre-ordered sentences and never explored directly learning contextualized representations that encapsulate richer semantic information and also reflect the position of words in the pre-ordered sentences. In this work, we propose a novel method that leverages the representations of pre-ordered sentences during the training process. We propose a Cross-Encoder Consistency (CEC) block to guide the encoder to output hidden states that reflect the word order of the target language by closing representations of the original and pre-ordered sentence. Additionally, we incorporate a Sentence Encoding Consistency (SEC) block to help the model preserve the semantic integrity of the source sentence. Experimental results on various low-resource NMT benchmarks demonstrate the effectiveness of our approach, resulting in substantial improvements in translation quality (e.g., up to +1.46 BLEU points in Tr → En translation).

Original languageEnglish
Title of host publicationProceedings of the 6th ACM International Conference on Multimedia in Asia, MMAsia 2024
PublisherAssociation for Computing Machinery, Inc
ISBN (Electronic)9798400712739
DOIs
Publication statusPublished - 28 Dec 2024
Externally publishedYes
Event6th ACM International Conference on Multimedia in Asia, MMAsia 2024 - Auckland, New Zealand
Duration: 3 Dec 20246 Dec 2024

Publication series

NameProceedings of the 6th ACM International Conference on Multimedia in Asia, MMAsia 2024

Conference

Conference6th ACM International Conference on Multimedia in Asia, MMAsia 2024
Country/TerritoryNew Zealand
CityAuckland
Period3/12/246/12/24

Keywords

  • Consistency Learning
  • Low-resource languages
  • Neural Machine Translation
  • Pre-ordering
  • Pre-trained Language Model

ASJC Scopus subject areas

  • Computer Graphics and Computer-Aided Design
  • Human-Computer Interaction

Fingerprint

Dive into the research topics of 'Incorporating Pre-ordering Representations for Low-resource Neural Machine Translation'. Together they form a unique fingerprint.

Cite this