DE eng

Search in the Catalogues and Directories

Hits 1 – 3 of 3

1
gaBERT -- an Irish Language Model ...
Abstract: The BERT family of neural language models have become highly popular due to their ability to provide sequences of text with rich context-sensitive token encodings which are able to generalise well to many Natural Language Processing tasks. Over 120 monolingual BERT models covering over 50 languages have been released, as well as a multilingual model trained on 104 languages. We introduce, gaBERT, a monolingual BERT model for the Irish language. We compare our gaBERT model to multilingual BERT and show that gaBERT provides better representations for a downstream parsing task. We also show how different filtering criteria, vocabulary size and the choice of subword tokenisation model affect downstream performance. We release gaBERT and related code to the community. ...
Keyword: Computation and Language cs.CL; FOS Computer and information sciences
URL: https://dx.doi.org/10.48550/arxiv.2107.12930
https://arxiv.org/abs/2107.12930
BASE
Hide details
2
Revisiting Tri-training of Dependency Parsers ...
Wagner, Joachim; Foster, Jennifer. - : arXiv, 2021
BASE
Show details
3
Cross-lingual Parsing with Polyglot Training and Multi-treebank Learning: A Faroese Case Study ...
BASE
Show details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
3
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern