Abstract
Data augmentation is an effective way to enhance the performance of neural machine translation models, especially for low-resource languages. Existing data augmentation methods are either at a token level or a sentence level. The data augmented using token level methods lack syntactic diversity and may alter original meanings. Sentence level methods usually generate low-quality source sentences that are not semantically paired with the original target sentences. In this paper, we propose a novel data augmentation method to generate diverse, high-quality and meaning-preserved new instances. Our method leverages high-quality translation models trained with high-resource languages to rephrase an original sentence by translating it into an intermediate language and then back to the original language. Through this process, the high-performing translation models guarantee the quality of the rephrased sentences, and the syntactic knowledge from the intermediate language can bring syntactic diversity to the rephrased sentences. Experimental results show our method can enhance the performance in various low-resource machine translation tasks. Moreover, by combining our method with other techniques that facilitate NMT, we can yield even better results.
Original language | English |
---|---|
Pages | 35-47 |
Number of pages | 13 |
Publication status | Published - 2023 |
Externally published | Yes |
Event | 19th Machine Translation Summit, MT Summit 2023 - Macau, China Duration: 4 Sept 2023 → 8 Sept 2023 |
Conference
Conference | 19th Machine Translation Summit, MT Summit 2023 |
---|---|
Country/Territory | China |
City | Macau |
Period | 4/09/23 → 8/09/23 |
ASJC Scopus subject areas
- Language and Linguistics
- Human-Computer Interaction
- Software