1 |
Quality Assurance of Generative Dialog Models in an Evolving Conversational Agent Used for Swedish Language Practice ...
|
|
|
|
BASE
|
|
Show details
|
|
2 |
Slangvolution: A Causal Analysis of Semantic Change and Frequency Dynamics in Slang ...
|
|
|
|
BASE
|
|
Show details
|
|
3 |
Similarity between person roles in a card sorting experiment ...
|
|
|
|
BASE
|
|
Show details
|
|
4 |
SPT-Code: Sequence-to-Sequence Pre-Training for Learning Source Code Representations ...
|
|
|
|
BASE
|
|
Show details
|
|
6 |
Ensemble of Opinion Dynamics Models to Understand the Role of the Undecided in the Vaccination Debate ...
|
|
|
|
BASE
|
|
Show details
|
|
9 |
Generating Authentic Adversarial Examples beyond Meaning-preserving with Doubly Round-trip Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
Pirá: A Bilingual Portuguese-English Dataset for Question-Answering about the Ocean ...
|
|
|
|
BASE
|
|
Show details
|
|
11 |
A comparative study of several parameterizations for speaker recognition ...
|
|
|
|
BASE
|
|
Show details
|
|
12 |
A Neural Pairwise Ranking Model for Readability Assessment ...
|
|
|
|
BASE
|
|
Show details
|
|
13 |
A bilingual approach to specialised adjectives through word embeddings in the karstology domain ...
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Speaker verification in mismatch training and testing conditions ...
|
|
|
|
BASE
|
|
Show details
|
|
15 |
Universal Conditional Masked Language Pre-training for Neural Machine Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
16 |
SMDT: Selective Memory-Augmented Neural Document Translation ...
|
|
|
|
BASE
|
|
Show details
|
|
17 |
Learning How to Translate North Korean through South Korean ...
|
|
|
|
BASE
|
|
Show details
|
|
18 |
When do Contrastive Word Alignments Improve Many-to-many Neural Machine Translation? ...
|
|
|
|
BASE
|
|
Show details
|
|
19 |
Conditional Bilingual Mutual Information Based Adaptive Training for Neural Machine Translation ...
|
|
|
|
Abstract:
Token-level adaptive training approaches can alleviate the token imbalance problem and thus improve neural machine translation, through re-weighting the losses of different target tokens based on specific statistical metrics (e.g., token frequency or mutual information). Given that standard translation models make predictions on the condition of previous target contexts, we argue that the above statistical metrics ignore target context information and may assign inappropriate weights to target tokens. While one possible solution is to directly take target contexts into these statistical metrics, the target-context-aware statistical computing is extremely expensive, and the corresponding storage overhead is unrealistic. To solve the above issues, we propose a target-context-aware metric, named conditional bilingual mutual information (CBMI), which makes it feasible to supplement target context information for statistical metrics. Particularly, our CBMI can be formalized as the log quotient of the translation ... : Accepted at ACL 2022 as a long paper of main conference. The code is available at: https://github.com/songmzhang/CBMI ...
|
|
Keyword:
Computation and Language cs.CL; FOS Computer and information sciences
|
|
URL: https://arxiv.org/abs/2203.02951 https://dx.doi.org/10.48550/arxiv.2203.02951
|
|
BASE
|
|
Hide details
|
|
|
|