All

What are you looking for?

All
Projects
Results
Organizations

Quick search

  • Projects supported by TA ČR
  • Excellent projects
  • Projects with the highest public support
  • Current projects

Smart search

  • That is how I find a specific +word
  • That is how I leave the -word out of the results
  • “That is how I can find the whole phrase”

LKMT: Linguistics Knowledge-DrivenMulti-Task NeuralMachine Translation for Urdu and English

The result's identifiers

  • Result code in IS VaVaI

    <a href="https://www.isvavai.cz/riv?ss=detail&h=RIV%2F00216208%3A11320%2F25%3AMHAJJG69" target="_blank" >RIV/00216208:11320/25:MHAJJG69 - isvavai.cz</a>

  • Result on the web

    <a href="https://www.scopus.com/inward/record.uri?eid=2-s2.0-85206828110&doi=10.32604%2fcmc.2024.054673&partnerID=40&md5=11cc8e5a13114b45ab4a4331dd861ce1" target="_blank" >https://www.scopus.com/inward/record.uri?eid=2-s2.0-85206828110&doi=10.32604%2fcmc.2024.054673&partnerID=40&md5=11cc8e5a13114b45ab4a4331dd861ce1</a>

  • DOI - Digital Object Identifier

    <a href="http://dx.doi.org/10.32604/cmc.2024.054673" target="_blank" >10.32604/cmc.2024.054673</a>

Alternative languages

  • Result language

    angličtina

  • Original language name

    LKMT: Linguistics Knowledge-DrivenMulti-Task NeuralMachine Translation for Urdu and English

  • Original language description

    Thanks to the strong representation capability of pre-trained language models, supervised machine translation models have achieved outstanding performance. However, the performances of these models drop sharply when the scale of the parallel training corpus is limited. Considering the pre-trained language model has a strong ability formonolingual representation, it is the key challenge formachine translation to construct the in-depth relationship between the source and target language by injecting the lexical and syntactic information into pre-trained language models. To alleviate the dependence on the parallel corpus, we propose a Linguistics Knowledge-Driven Multi- Task (LKMT) approach to inject part-of-speech and syntactic knowledge into pre-trained models, thus enhancing the machine translation performance. On the one hand, we integrate part-of-speech and dependency labels into the embedding layer and exploit large-scale monolingual corpus to update all parameters of pre-trained language models, thus ensuring the updated language model contains potential lexical and syntactic information. On the other hand, we leverage an extra self-attention layer to explicitly inject linguistic knowledge into the pre-trained language model-enhanced machine translation model. Experiments on the benchmark dataset show that our proposed LKMT approach improves the Urdu-English translation accuracy by 1.97 points and the English-Urdu translation accuracy by 2.42 points, highlighting the effectiveness of our LKMT framework. Detailed ablation experiments confirm the positive impact of part-of-speech and dependency parsing on machine translation. © 2024 The Authors.

  • Czech name

  • Czech description

Classification

  • Type

    J<sub>SC</sub> - Article in a specialist periodical, which is included in the SCOPUS database

  • CEP classification

  • OECD FORD branch

    10201 - Computer sciences, information science, bioinformathics (hardware development to be 2.2, social aspect to be 5.8)

Result continuities

  • Project

  • Continuities

Others

  • Publication year

    2024

  • Confidentiality

    S - Úplné a pravdivé údaje o projektu nepodléhají ochraně podle zvláštních právních předpisů

Data specific for result type

  • Name of the periodical

    Computers, Materials and Continua

  • ISSN

    1546-2218

  • e-ISSN

  • Volume of the periodical

    81

  • Issue of the periodical within the volume

    1

  • Country of publishing house

    US - UNITED STATES

  • Number of pages

    19

  • Pages from-to

    951-969

  • UT code for WoS article

  • EID of the result in the Scopus database

    2-s2.0-85206828110