LKMT: Linguistics Knowledge-DrivenMulti-Task NeuralMachine Translation for Urdu and English
The result's identifiers
Result code in IS VaVaI
<a href="https://www.isvavai.cz/riv?ss=detail&h=RIV%2F00216208%3A11320%2F25%3AMHAJJG69" target="_blank" >RIV/00216208:11320/25:MHAJJG69 - isvavai.cz</a>
Result on the web
<a href="https://www.scopus.com/inward/record.uri?eid=2-s2.0-85206828110&doi=10.32604%2fcmc.2024.054673&partnerID=40&md5=11cc8e5a13114b45ab4a4331dd861ce1" target="_blank" >https://www.scopus.com/inward/record.uri?eid=2-s2.0-85206828110&doi=10.32604%2fcmc.2024.054673&partnerID=40&md5=11cc8e5a13114b45ab4a4331dd861ce1</a>
DOI - Digital Object Identifier
<a href="http://dx.doi.org/10.32604/cmc.2024.054673" target="_blank" >10.32604/cmc.2024.054673</a>
Alternative languages
Result language
angličtina
Original language name
LKMT: Linguistics Knowledge-DrivenMulti-Task NeuralMachine Translation for Urdu and English
Original language description
Thanks to the strong representation capability of pre-trained language models, supervised machine translation models have achieved outstanding performance. However, the performances of these models drop sharply when the scale of the parallel training corpus is limited. Considering the pre-trained language model has a strong ability formonolingual representation, it is the key challenge formachine translation to construct the in-depth relationship between the source and target language by injecting the lexical and syntactic information into pre-trained language models. To alleviate the dependence on the parallel corpus, we propose a Linguistics Knowledge-Driven Multi- Task (LKMT) approach to inject part-of-speech and syntactic knowledge into pre-trained models, thus enhancing the machine translation performance. On the one hand, we integrate part-of-speech and dependency labels into the embedding layer and exploit large-scale monolingual corpus to update all parameters of pre-trained language models, thus ensuring the updated language model contains potential lexical and syntactic information. On the other hand, we leverage an extra self-attention layer to explicitly inject linguistic knowledge into the pre-trained language model-enhanced machine translation model. Experiments on the benchmark dataset show that our proposed LKMT approach improves the Urdu-English translation accuracy by 1.97 points and the English-Urdu translation accuracy by 2.42 points, highlighting the effectiveness of our LKMT framework. Detailed ablation experiments confirm the positive impact of part-of-speech and dependency parsing on machine translation. © 2024 The Authors.
Czech name
—
Czech description
—
Classification
Type
J<sub>SC</sub> - Article in a specialist periodical, which is included in the SCOPUS database
CEP classification
—
OECD FORD branch
10201 - Computer sciences, information science, bioinformathics (hardware development to be 2.2, social aspect to be 5.8)
Result continuities
Project
—
Continuities
—
Others
Publication year
2024
Confidentiality
S - Úplné a pravdivé údaje o projektu nepodléhají ochraně podle zvláštních právních předpisů
Data specific for result type
Name of the periodical
Computers, Materials and Continua
ISSN
1546-2218
e-ISSN
—
Volume of the periodical
81
Issue of the periodical within the volume
1
Country of publishing house
US - UNITED STATES
Number of pages
19
Pages from-to
951-969
UT code for WoS article
—
EID of the result in the Scopus database
2-s2.0-85206828110