1 |
When do Contrastive Word Alignments Improve Many-to-many Neural Machine Translation? ...
|
|
|
|
BASE
|
|
Show details
|
|
2 |
VISA: An Ambiguous Subtitles Dataset for Visual Scene-Aware Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
3 |
Linguistically-driven Multi-task Pre-training for Low-resource Neural Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
4 |
Attending Self-Attention: A Case Study ofVisually Grounded Supervision in Vision-and-Language Transformers ...
|
|
|
|
BASE
|
|
Show details
|
|
5 |
Lightweight Cross-Lingual Sentence Representation Learning ...
|
|
|
|
BASE
|
|
Show details
|
|
6 |
A Corpus for English-Japanese Multimodal Neural Machine Translation with Comparable Sentences ...
|
|
|
|
BASE
|
|
Show details
|
|
7 |
A Comprehensive Survey of Multilingual Neural Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
8 |
Lexically Cohesive Neural Machine Translation with Copy Mechanism ...
|
|
|
|
Abstract:
Lexically cohesive translations preserve consistency in word choices in document-level translation. We employ a copy mechanism into a context-aware neural machine translation model to allow copying words from previous translation outputs. Different from previous context-aware neural machine translation models that handle all the discourse phenomena implicitly, our model explicitly addresses the lexical cohesion problem by boosting the probabilities to output words consistently. We conduct experiments on Japanese to English translation using an evaluation dataset for discourse translation. The results showed that the proposed model significantly improved lexical cohesion compared to previous context-aware models. ...
|
|
Keyword:
Artificial Intelligence cs.AI; Computation and Language cs.CL; FOS Computer and information sciences
|
|
URL: https://dx.doi.org/10.48550/arxiv.2010.05193 https://arxiv.org/abs/2010.05193
|
|
BASE
|
|
Hide details
|
|
9 |
iParaphrasing: Extracting Visually Grounded Paraphrases via an Image ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|