TY - GEN
T1 - Irish-based Large Language Model with Extreme Low-Resource Settings in Machine Translation
AU - Tran, Khanh Tung
AU - O'Sullivan, Barry
AU - Nguyen, Hoang D.
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - Large Language Models (LLMs) have demonstrated exceptional performances in a wide range of natural language processing tasks. However, their success does not always extend to machine translation, particularly in challenging scenarios such as translating low-resource languages. This study investigates the multilingual capability of LLMs, with a case study on Irish, an extremely low-resource language, focusing on translation tasks between English and Irish. We propose a dynamic, efficient language adaptation framework for English-centric LLMs, which involves layer-specific adjustments and subsequent fine-tuning for machine translation. Our findings highlight several key insights: (1) different layers in the LLM serve distinct functions such as language understanding and task reasoning, (2) effective translation requires extensive pre-training on both source and target languages, and (3) targeted fine-tuning for machine translation leads to significant improvements of 36.7% for English to Irish and 133.4% for Irish to English compared to the previous state-of-the-art.
AB - Large Language Models (LLMs) have demonstrated exceptional performances in a wide range of natural language processing tasks. However, their success does not always extend to machine translation, particularly in challenging scenarios such as translating low-resource languages. This study investigates the multilingual capability of LLMs, with a case study on Irish, an extremely low-resource language, focusing on translation tasks between English and Irish. We propose a dynamic, efficient language adaptation framework for English-centric LLMs, which involves layer-specific adjustments and subsequent fine-tuning for machine translation. Our findings highlight several key insights: (1) different layers in the LLM serve distinct functions such as language understanding and task reasoning, (2) effective translation requires extensive pre-training on both source and target languages, and (3) targeted fine-tuning for machine translation leads to significant improvements of 36.7% for English to Irish and 133.4% for Irish to English compared to the previous state-of-the-art.
UR - https://www.scopus.com/pages/publications/85204908624
M3 - Conference proceeding
AN - SCOPUS:85204908624
T3 - LoResMT 2024 - 7th Workshop on Technologies for Machine Translation of Low-Resource Languages, Proceedings of the Workshop
SP - 85
EP - 93
BT - LoResMT 2024 - 7th Workshop on Technologies for Machine Translation of Low-Resource Languages, Proceedings of the Workshop
A2 - Ojha, Atul Kr.
A2 - Ojha, Atul Kr.
A2 - Liu, Chao-hong
A2 - Vylomova, Ekaterina
A2 - Pirinen, Flammie
A2 - Abbott, Jade
A2 - Washington, Jonathan
A2 - Oco, Nathaniel
A2 - Malykh, Valentin
A2 - Logacheva, Varvara Skolkovo
A2 - Zhao, Xiaobing
PB - Association for Computational Linguistics (ACL)
T2 - 7th Workshop on Technologies for Machine Translation of Low-Resource Languages, LoResMT 2024 at ACL 2024
Y2 - 15 August 2024
ER -