CLC number:
On-line Access: 2024-12-30
Received: 2024-05-30
Revision Accepted: 2024-11-27
Crosschecked: 0000-00-00
Cited: 0
Clicked: 24
Changtong Zan, Liang Ding, Li Shen3, Yibing Zhan, Xinghao Yang, Weifeng Liu. Building accurate translation-tailored LLMswith language aware instruction tuning[J]. Frontiers of Information Technology & Electronic Engineering, 1998, -1(-1): .
@article{title="Building accurate translation-tailored LLMswith language aware instruction tuning",
author="Changtong Zan, Liang Ding, Li Shen3, Yibing Zhan, Xinghao Yang, Weifeng Liu",
journal="Frontiers of Information Technology & Electronic Engineering",
volume="-1",
number="-1",
pages="",
year="1998",
publisher="Zhejiang University Press & Springer",
doi="10.1631/FITEE.2400458"
}
%0 Journal Article
%T Building accurate translation-tailored LLMswith language aware instruction tuning
%A Changtong Zan
%A Liang Ding
%A Li Shen3
%A Yibing Zhan
%A Xinghao Yang
%A Weifeng Liu
%J Journal of Zhejiang University SCIENCE C
%V -1
%N -1
%P
%@ 2095-9184
%D 1998
%I Zhejiang University Press & Springer
%DOI 10.1631/FITEE.2400458
TY - JOUR
T1 - Building accurate translation-tailored LLMswith language aware instruction tuning
A1 - Changtong Zan
A1 - Liang Ding
A1 - Li Shen3
A1 - Yibing Zhan
A1 - Xinghao Yang
A1 - Weifeng Liu
J0 - Journal of Zhejiang University Science C
VL - -1
IS - -1
SP -
EP -
%@ 2095-9184
Y1 - 1998
PB - Zhejiang University Press & Springer
ER -
DOI - 10.1631/FITEE.2400458
Abstract: large language models (LLMs) exhibit remarkable capabilities in various natural language processing tasks, such as machine translation. However, the large number of LLM parameters incurs significant costs during inference. Previous studies have attempted to train translation-tailored LLMs with moderately sized models by fine-tuning them on translation data. Nevertheless, when applying zero-shot translation directions not included in the fine-tuning data, the issue of ignoring instructions and thus translating into the wrong language, i.e., the off-target translation issue, remains unsolved. In this work, we design a two-stage fine-tuning algorithm to improve the instruction-following ability of translation-tailored LLMs, particularly for maintaining accurate translation directions. We first fine-tune LLMs on the translation dataset to elicit basic translation capabilities. In the second stage, we construct instruction-conflicting samples by randomly replacing the instructions with incorrect ones. Then, we introduce an extra unlikelihood loss to reduce the probability assigned to those samples. Experiments on IWSLT and WMT benchmarks using the LLaMA2 and LLaMA3 models, spanning 16 zero-shot directions, demonstrate that, compared to the competitive baseline-translation-finetuned LLaMA, our method could effectively reduce the off-target translation ratio (up to ?62.4%), thus improving translation quality (up to +9.7 BLEU). Analysis shows that our method can preserve the model's performance on other tasks, such as supervised translation and general tasks. Code is released at: https://github.com/alphadl/LanguageAware_Tuning.
Open peer comments: Debate/Discuss/Question/Opinion
<1>