Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
Date
Availability
1-1 of 1
Jianhui Pang
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
Rethinking the Exploitation of Monolingual Data for Low-Resource Neural Machine Translation
Open AccessPublisher: Journals Gateway
Computational Linguistics (2024) 50 (1): 25–47.
Published: 01 March 2024
FIGURES
| View All (4)
Abstract
View articletitled, Rethinking the Exploitation of Monolingual Data for Low-Resource
Neural Machine Translation
View
PDF
for article titled, Rethinking the Exploitation of Monolingual Data for Low-Resource
Neural Machine Translation
The utilization of monolingual data has been shown to be a promising strategy for addressing low-resource machine translation problems. Previous studies have demonstrated the effectiveness of techniques such as back-translation and self-supervised objectives, including masked language modeling, causal language modeling, and denoise autoencoding, in improving the performance of machine translation models. However, the manner in which these methods contribute to the success of machine translation tasks and how they can be effectively combined remains an under-researched area. In this study, we carry out a systematic investigation of the effects of these techniques on linguistic properties through the use of probing tasks, including source language comprehension, bilingual word alignment, and translation fluency. We further evaluate the impact of pre-training, back-translation, and multi-task learning on bitexts of varying sizes. Our findings inform the design of more effective pipelines for leveraging monolingual data in extremely low-resource and low-resource machine translation tasks. Experiment results show consistent performance gains in seven translation directions, which provide further support for our conclusions and understanding of the role of monolingual data in machine translation.