TY - GEN
T1 - Incorporating Pre-ordering Representations for Low-resource Neural Machine Translation
AU - Gao, Yuan
AU - Hou, Feng
AU - Wang, Ruili
N1 - Publisher Copyright:
© 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM.
PY - 2024/12/28
Y1 - 2024/12/28
N2 - Pre-ordering, a data pre-processing technique that rearranges the order of source language words to align with the target language's natural word order, has been proven to be effective in improving machine translation performance, particularly for low-resource language pairs. Existing research, however, primarily exploits the position embeddings of tokens in the pre-ordered sentences and never explored directly learning contextualized representations that encapsulate richer semantic information and also reflect the position of words in the pre-ordered sentences. In this work, we propose a novel method that leverages the representations of pre-ordered sentences during the training process. We propose a Cross-Encoder Consistency (CEC) block to guide the encoder to output hidden states that reflect the word order of the target language by closing representations of the original and pre-ordered sentence. Additionally, we incorporate a Sentence Encoding Consistency (SEC) block to help the model preserve the semantic integrity of the source sentence. Experimental results on various low-resource NMT benchmarks demonstrate the effectiveness of our approach, resulting in substantial improvements in translation quality (e.g., up to +1.46 BLEU points in Tr → En translation).
AB - Pre-ordering, a data pre-processing technique that rearranges the order of source language words to align with the target language's natural word order, has been proven to be effective in improving machine translation performance, particularly for low-resource language pairs. Existing research, however, primarily exploits the position embeddings of tokens in the pre-ordered sentences and never explored directly learning contextualized representations that encapsulate richer semantic information and also reflect the position of words in the pre-ordered sentences. In this work, we propose a novel method that leverages the representations of pre-ordered sentences during the training process. We propose a Cross-Encoder Consistency (CEC) block to guide the encoder to output hidden states that reflect the word order of the target language by closing representations of the original and pre-ordered sentence. Additionally, we incorporate a Sentence Encoding Consistency (SEC) block to help the model preserve the semantic integrity of the source sentence. Experimental results on various low-resource NMT benchmarks demonstrate the effectiveness of our approach, resulting in substantial improvements in translation quality (e.g., up to +1.46 BLEU points in Tr → En translation).
KW - Consistency Learning
KW - Low-resource languages
KW - Neural Machine Translation
KW - Pre-ordering
KW - Pre-trained Language Model
UR - http://www.scopus.com/inward/record.url?scp=85216223854&partnerID=8YFLogxK
U2 - 10.1145/3696409.3700267
DO - 10.1145/3696409.3700267
M3 - Conference contribution
AN - SCOPUS:85216223854
T3 - Proceedings of the 6th ACM International Conference on Multimedia in Asia, MMAsia 2024
BT - Proceedings of the 6th ACM International Conference on Multimedia in Asia, MMAsia 2024
PB - Association for Computing Machinery, Inc
T2 - 6th ACM International Conference on Multimedia in Asia, MMAsia 2024
Y2 - 3 December 2024 through 6 December 2024
ER -