DE eng

Search in the Catalogues and Directories

Hits 1 – 7 of 7

1
huggingface/datasets: 1.18.1 ...
BASE
Show details
2
huggingface/transformers: v4.4.0: S2T, M2M100, I-BERT, mBART-50, DeBERTa-v2, XLSR-Wav2Vec2 ...
BASE
Show details
3
huggingface/datasets: 1.16.0 ...
BASE
Show details
4
Transformers: State-of-the-Art Natural Language Processing ...
BASE
Show details
5
Transformers: State-of-the-Art Natural Language Processing ...
BASE
Show details
6
huggingface/transformers: ProphetNet, Blenderbot, SqueezeBERT, DeBERTa ...
BASE
Show details
7
MultiFiT: Efficient Multi-lingual Language Model Fine-tuning ...
Abstract: Pretrained language models are promising particularly for low-resource languages as they only require unlabelled data. However, training existing models requires huge amounts of compute, while pretrained cross-lingual models often underperform on low-resource languages. We propose Multi-lingual language model Fine-Tuning (MultiFiT) to enable practitioners to train and fine-tune language models efficiently in their own language. In addition, we propose a zero-shot method using an existing pretrained cross-lingual model. We evaluate our methods on two widely used cross-lingual classification datasets where they outperform models pretrained on orders of magnitude more data and compute. We release all models and code. ... : Proceedings of EMNLP-IJCNLP 2019 ...
Keyword: Computation and Language cs.CL; FOS Computer and information sciences; Machine Learning cs.LG
URL: https://arxiv.org/abs/1909.04761
https://dx.doi.org/10.48550/arxiv.1909.04761
BASE
Hide details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
7
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern