@inproceedings{7e84960b96f84faabe0dd14cfcc83891,
title = "A Novel Two-step Fine-tuning Framework for Transfer Learning in Low-Resource Neural Machine Translation",
abstract = "Existing transfer learning methods for neural machine translation typically use a well-trained translation model (i.e., a parent model) of a high-resource language pair to directly initialize a translation model (i.e., a child model) of a low-resource language pair, and the child model is then fine-tuned with corresponding datasets. In this paper, we propose a novel two-step fine-tuning (TSFT) framework for transfer learning in low-resource neural machine translation. In the first step, we adjust the parameters of the parent model to fit the child language by using the child source data. In the second step, we transfer the adjusted parameters to the child model and fine-tune it with a proposed distillation loss for efficient optimization. Our experimental results on five low-resource translations demonstrate that our framework yields significant improvements over various strong transfer learning baselines. Further analysis demonstrated the effectiveness of different components in our framework.",
author = "Yuan Gao and Feng Hou and Ruili Wang",
note = "Publisher Copyright: {\textcopyright} 2024 Association for Computational Linguistics.; 2024 Findings of the Association for Computational Linguistics: NAACL 2024 ; Conference date: 16-06-2024 Through 21-06-2024",
year = "2024",
language = "English",
series = "Findings of the Association for Computational Linguistics: NAACL 2024 - Findings",
publisher = "Association for Computational Linguistics (ACL)",
pages = "3214--3224",
editor = "Kevin Duh and Helena Gomez and Steven Bethard",
booktitle = "Findings of the Association for Computational Linguistics",
address = "United States",
}