1 |
Toward Multilingual Neural Machine Translation with Universal Encoder and Decoder
|
|
|
|
BASE
|
|
Show details
|
|
2 |
The Karlsruhe Institute of Technology Systems for the News Translation Task in WMT 2017
|
|
|
|
BASE
|
|
Show details
|
|
4 |
KIT Lecture Translator: Multilingual Speech Translation with One-Shot Learning
|
|
|
|
BASE
|
|
Show details
|
|
5 |
Lecture Translator Speech translation framework for simultaneous lecture translation
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Improving Zero-shot Translation with Language-Independent Constraints
|
|
|
|
BASE
|
|
Show details
|
|
7 |
Lexical Translation Model Using A Deep Neural Network Architecture
|
|
|
|
BASE
|
|
Show details
|
|
10 |
The Karlsruhe Institute of Technology Systems for the News Translation Task in WMT 2016
|
|
|
|
BASE
|
|
Show details
|
|
12 |
Improving Multilingual Neural Machine Translation For Low-Resource Languages: French,English - Vietnamese ...
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Combining Advanced Methods in Japanese-Vietnamese Neural Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
15 |
Effective Strategies in Zero-Shot Neural Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
16 |
Toward Multilingual Neural Machine Translation with Universal Encoder and Decoder ...
|
|
|
|
Abstract:
In this paper, we present our first attempts in building a multilingual Neural Machine Translation framework under a unified approach. We are then able to employ attention-based NMT for many-to-many multilingual translation tasks. Our approach does not require any special treatment on the network architecture and it allows us to learn minimal number of free parameters in a standard way of training. Our approach has shown its effectiveness in an under-resourced translation scenario with considerable improvements up to 2.6 BLEU points. In addition, the approach has achieved interesting and promising results when applied in the translation task that there is no direct parallel corpus between source and target languages. ...
|
|
Keyword:
Computation and Language cs.CL; FOS Computer and information sciences
|
|
URL: https://dx.doi.org/10.48550/arxiv.1611.04798 https://arxiv.org/abs/1611.04798
|
|
BASE
|
|
Hide details
|
|
17 |
Using Factored Word Representation in Neural Network Language Models
|
|
|
|
BASE
|
|
Show details
|
|
18 |
Lexical Translation Model Using a Deep Neural Network Architecture ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|