Return to search

Multilingual Neural Machine Translation for Low Resource Languages

Machine Translation (MT) is the task of mapping a source language to a target language. The recent introduction of neural MT (NMT) has shown promising results for high-resource language, however, poorly performing for low-resource language (LRL) settings. Furthermore, the vast majority of the 7, 000+ languages around the world do not
have parallel data, creating a zero-resource language (ZRL) scenario. In this thesis, we
present our approach to improving NMT for LRL and ZRL, leveraging a multilingual NMT
modeling (M-NMT), an approach that allows building a single NMT to translate across
multiple source and target languages. This thesis begins by i) analyzing the effectiveness
of M-NMT for LRL and ZRL translation tasks, spanning two NMT modeling architectures (Recurrent and Transformer), ii) presents a self-learning approach for improving the zero-shot translation directions of ZRLs, iii) proposes a dynamic transfer-learning approach from a pre-trained (parent) model to a LRL (child) model by tailoring to the
vocabulary entries of the latter, iv) extends M-NMT to translate from a source language
to specific language varieties (e.g. dialects), and finally, v) proposes an approach that
can control the verbosity of an NMT model output. Our experimental findings show the
effectiveness of the proposed approaches in improving NMT of LRLs and ZRLs.

Identiferoai:union.ndltd.org:unitn.it/oai:iris.unitn.it:11572/257906
Date20 April 2020
CreatorsLakew, Surafel Melaku
ContributorsLakew, Surafel Melaku
PublisherUniversità degli studi di Trento, place:Trento
Source SetsUniversità di Trento
LanguageEnglish
Detected LanguageEnglish
Typeinfo:eu-repo/semantics/doctoralThesis
Rightsinfo:eu-repo/semantics/closedAccess
Relationfirstpage:1, lastpage:178, numberofpages:178

Page generated in 0.0039 seconds