2 |
On the Representation Collapse of Sparse Mixture of Experts ...
|
|
|
|
BASE
|
|
Show details
|
|
4 |
Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment ...
|
|
|
|
BASE
|
|
Show details
|
|
5 |
MT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs ...
|
|
|
|
BASE
|
|
Show details
|
|
6 |
XLM-E: Cross-lingual Language Model Pre-training via ELECTRA ...
|
|
|
|
BASE
|
|
Show details
|
|
7 |
InfoXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training ...
|
|
|
|
BASE
|
|
Show details
|
|
8 |
Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word Alignment ...
|
|
|
|
BASE
|
|
Show details
|
|
9 |
mT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
A Robust and Domain-Adaptive Approach for Low-Resource Named Entity Recognition ...
|
|
|
|
BASE
|
|
Show details
|
|
11 |
InfoXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training ...
|
|
|
|
BASE
|
|
Show details
|
|
12 |
XLM-T: Scaling up Multilingual Machine Translation with Pretrained Cross-lingual Transformer Encoders ...
|
|
|
|
BASE
|
|
Show details
|
|
13 |
Can Monolingual Pretrained Models Help Cross-Lingual Classification? ...
|
|
|
|
BASE
|
|
Show details
|
|
14 |
Cross-Lingual Natural Language Generation via Pre-Training ...
|
|
|
|
BASE
|
|
Show details
|
|
|
|