Author
Listed:
- Yirong Pan
(Xinjiang Technical Institute of Physics & Chemistry, Chinese Academy of Sciences, Urumqi 830011, China
Department of Computer and Control, University of Chinese Academy of Sciences, Beijing 100049, China
Xinjiang Laboratory of Minority Speech and Language Information Processing, Urumqi 830011, China)
- Xiao Li
(Xinjiang Technical Institute of Physics & Chemistry, Chinese Academy of Sciences, Urumqi 830011, China
Department of Computer and Control, University of Chinese Academy of Sciences, Beijing 100049, China
Xinjiang Laboratory of Minority Speech and Language Information Processing, Urumqi 830011, China)
- Yating Yang
(Xinjiang Technical Institute of Physics & Chemistry, Chinese Academy of Sciences, Urumqi 830011, China
Department of Computer and Control, University of Chinese Academy of Sciences, Beijing 100049, China
Xinjiang Laboratory of Minority Speech and Language Information Processing, Urumqi 830011, China)
- Rui Dong
(Xinjiang Technical Institute of Physics & Chemistry, Chinese Academy of Sciences, Urumqi 830011, China
Department of Computer and Control, University of Chinese Academy of Sciences, Beijing 100049, China
Xinjiang Laboratory of Minority Speech and Language Information Processing, Urumqi 830011, China)
Abstract
Benefitting from the rapid development of artificial intelligence (AI) and deep learning, the machine translation task based on neural networks has achieved impressive performance in many high-resource language pairs. However, the neural machine translation (NMT) models still struggle in the translation task on agglutinative languages with complex morphology and limited resources. Inspired by the finding that utilizing the source-side linguistic knowledge can further improve the NMT performance, we propose a multi-source neural model that employs two separate encoders to encode the source word sequence and the linguistic feature sequences. Compared with the standard NMT model, we utilize an additional encoder to incorporate the linguistic features of lemma, part-of-speech (POS) tag, and morphological tag by extending the input embedding layer of the encoder. Moreover, we use a serial combination method to integrate the conditional information from the encoders with the outputs of the decoder, which aims to enhance the neural model to learn a high-quality context representation of the source sentence. Experimental results show that our approach is effective for the agglutinative language translation, which achieves the highest improvements of +2.4 BLEU points on Turkish–English translation task and +0.6 BLEU points on Uyghur–Chinese translation task.
Suggested Citation
Yirong Pan & Xiao Li & Yating Yang & Rui Dong, 2020.
"Multi-Source Neural Model for Machine Translation of Agglutinative Language,"
Future Internet, MDPI, vol. 12(6), pages 1-14, June.
Handle:
RePEc:gam:jftint:v:12:y:2020:i:6:p:96-:d:366622
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jftint:v:12:y:2020:i:6:p:96-:d:366622. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.