Affiliation(s):
College of Control Science and Engineering, China University of Petroleum (East China), Shandong 266580, China;
moreAffiliation(s): College of Control Science and Engineering, China University of Petroleum (East China), Shandong 266580, China; School of Computer Science, University of Sydney, NSW 2006, Australia; JD Explore Academy, JD.com Inc., Beijing 100101, China; School of Cyber Science and Technology, Shenzhen Campus of Sun Yat-sen University, Guangdong 518107, China;
less
Changtong Zan, Liang Ding, Li Shen3, Yibing Zhan, Xinghao Yang, Weifeng Liu. Building accurate translation-tailored LLMswith language aware instruction tuning[J]. Frontiers of Information Technology & Electronic Engineering,in press.https://doi.org/10.1631/FITEE.2400458
@article{title="Building accurate translation-tailored LLMswith language aware instruction tuning", author="Changtong Zan, Liang Ding, Li Shen3, Yibing Zhan, Xinghao Yang, Weifeng Liu", journal="Frontiers of Information Technology & Electronic Engineering", year="in press", publisher="Zhejiang University Press & Springer", doi="https://doi.org/10.1631/FITEE.2400458" }
%0 Journal Article %T Building accurate translation-tailored LLMswith language aware instruction tuning %A Changtong Zan %A Liang Ding %A Li Shen3 %A Yibing Zhan %A Xinghao Yang %A Weifeng Liu %J Frontiers of Information Technology & Electronic Engineering %P %@ 2095-9184 %D in press %I Zhejiang University Press & Springer doi="https://doi.org/10.1631/FITEE.2400458"
TY - JOUR T1 - Building accurate translation-tailored LLMswith language aware instruction tuning A1 - Changtong Zan A1 - Liang Ding A1 - Li Shen3 A1 - Yibing Zhan A1 - Xinghao Yang A1 - Weifeng Liu J0 - Frontiers of Information Technology & Electronic Engineering SP - EP - %@ 2095-9184 Y1 - in press PB - Zhejiang University Press & Springer ER - doi="https://doi.org/10.1631/FITEE.2400458"
Abstract: Large language models (LLMs) exhibit remarkable capabilities in various natural language processing tasks, such as machine translation. However, the large number of LLM parameters incurs significant costs during inference. Previous studies have attempted to train translation-tailored LLMs with moderately sized models by fine-tuning them on translation data. Nevertheless, when applying zero-shot translation directions not included in the fine-tuning data, the issue of ignoring instructions and thus translating into the wrong language, i.e., the off-target translation issue, remains unsolved. In this work, we design a two-stage fine-tuning algorithm to improve the instruction-following ability of translation-tailored LLMs, particularly for maintaining accurate translation directions. We first fine-tune LLMs on the translation dataset to elicit basic translation capabilities. In the second stage, we construct instruction-conflicting samples by randomly replacing the instructions with incorrect ones. Then, we introduce an extra unlikelihood loss to reduce the probability assigned to those samples. Experiments on IWSLT and WMT benchmarks using the LLaMA2 and LLaMA3 models, spanning 16 zero-shot directions, demonstrate that, compared to the competitive baseline-translation-finetuned LLaMA, our method could effectively reduce the off-target translation ratio (up to ?62.4%), thus improving translation quality (up to +9.7 BLEU). Analysis shows that our method can preserve the model's performance on other tasks, such as supervised translation and general tasks. Code is released at: https://github.com/alphadl/LanguageAware_Tuning.
Darkslateblue:Affiliate; Royal Blue:Author; Turquoise:Article
Reference
Open peer comments: Debate/Discuss/Question/Opinion
Open peer comments: Debate/Discuss/Question/Opinion
<1>