Tailoring neural architectures for translating from
morphologically rich languages
Passban, Peyman, Way, AndyORCID: 0000-0001-5736-5930 and Liu, QunORCID: 0000-0002-7000-1792
(2018)
Tailoring neural architectures for translating from
morphologically rich languages.
In: 27th International Conference on Computational Linguistics, 20-26 Aug 2018, Santa Fe, New Mexico, USA.
A morphologically complex word (MCW) is a hierarchical constituent with meaning-preserving
subunits, so word-based models which rely on surface forms might not be powerful enough
to translate such structures. When translating from morphologically rich languages (MRLs), a
source word could be mapped to several words or even a full sentence on the target side, which
means an MCW should not be treated as an atomic unit. In order to provide better translations
for MRLs, we boost the existing neural machine translation (NMT) architecture with a doublechannel encoder and a double-attentive decoder. The main goal targeted in this research is to
provide richer information on the encoder side and redesign the decoder accordingly to benefit
from such information. Our experimental results demonstrate that we could achieve our goal as
the proposed model outperforms existing subword- and character-based architectures and showed
significant improvements on translating from German, Russian, and Turkish into English.
This item is licensed under a Creative Commons Attribution-NonCommercial-Share Alike 3.0 License. View License
Funders:
ADAPT Centre for Digital Content Technology which is funded under the SFI Research Centres Programme (Grant 13/RC/2106) and is co-funded under the European Regional Development Fund.
ID Code:
23195
Deposited On:
17 Apr 2019 12:07 by
Thomas Murtagh
. Last Modified 17 Apr 2019 12:07