Assessing and Improving Punctuation Robustness in English-Marathi Machine Translation
Abstract
Researchers developed Virām, a diagnostic benchmark for evaluating punctuation robustness in English-to-Marathi machine translation, showing that specialized fine-tuned models outperform standard baselines and current LLMs in handling punctuation-ambiguous text.
Punctuation plays a critical role in resolving semantic and structural ambiguity in written language. Machine Translation (MT) systems are now widely applied across diverse domains and languages, including many low-resource settings. In this work, we focus on Marathi, a low- to middle-resource language. We introduce Virām, the first diagnostic benchmark for assessing punctuation robustness in English-to-Marathi machine translation, consisting of 54 manually curated, punctuation-ambiguous instances. We evaluate two primary strategies for enhancing reliability: a pipeline-based restore-then-translate approach and direct fine-tuned on punctuation-varied data. Our results demonstrate that specialized fine-tuned models and pipeline systems significantly improve translation quality over standard baselines on the Virām benchmark. Qualitative analysis reveals that the original model may result in wrong translations leading to wrong interpretations, while fine-tuned models significantly improve overall reliability. Furthermore, we find that current Large Language Models (LLMs) lag behind these task-specific approaches in preserving meaning for punctuation-ambiguous text, thus necessitating further research in this area. The code and dataset is available at https://github.com/KaustubhShejole/Viram_Marathi.
Models citing this paper 5
Browse 5 models citing this paperDatasets citing this paper 1
Spaces citing this paper 1
Collections including this paper 0
No Collection including this paper