Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
Date
Availability
1-1 of 1
Jianhui Pang
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
Salute the Classic: Revisiting Challenges of Machine Translation in the Age of Large Language Models
Publisher: Journals Gateway
Transactions of the Association for Computational Linguistics (2025) 13: 73–95.
Published: 07 January 2025
FIGURES
| View All (11)
Abstract
View articletitled, Salute the Classic: Revisiting Challenges of Machine Translation in
the Age of Large Language Models
View
PDF
for article titled, Salute the Classic: Revisiting Challenges of Machine Translation in
the Age of Large Language Models
The evolution of Neural Machine Translation (NMT) has been significantly influenced by six core challenges (Koehn and Knowles, 2017 ) that have acted as benchmarks for progress in this field. This study revisits these challenges, offering insights into their ongoing relevance in the context of advanced Large Language Models (LLMs): domain mismatch , amount of parallel data , rare word prediction , translation of long sentences , attention model as word alignment , and sub-optimal beam search . Our empirical findings show that LLMs effectively reduce reliance on parallel data for major languages during pretraining and significantly improve translation of long sentences containing approximately 80 words, even translating documents up to 512 words. Despite these improvements, challenges in domain mismatch and rare word prediction persist. While NMT-specific challenges like word alignment and beam search may not apply to LLMs, we identify three new challenges in LLM-based translation: inference efficiency, translation of low-resource languages during pretraining, and human-aligned evaluation.