1 |
WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models ...
|
|
|
|
BASE
|
|
Show details
|
|
2 |
Regularization Advantages of Multilingual Neural Language Models for Low Resource Domains ...
|
|
|
|
Abstract:
Neural language modeling (LM) has led to significant improvements in several applications, including Automatic Speech Recognition. However, they typically require large amounts of training data, which is not available for many domains and languages. In this study, we propose a multilingual neural language model architecture, trained jointly on the domain-specific data of several low-resource languages. The proposed multilingual LM consists of language specific word embeddings in the encoder and decoder, and one language specific LSTM layer, plus two LSTM layers with shared parameters across the languages. This multilingual LM model facilitates transfer learning across the languages, acting as an extra regularizer in very low-resource scenarios. We integrate our proposed multilingual approach with a state-of-the-art highly-regularized neural LM, and evaluate on the conversational data domain for four languages over a range of training data sizes. Compared to monolingual LMs, the results show significant ...
|
|
Keyword:
Computation and Language cs.CL; FOS Computer and information sciences; Machine Learning cs.LG; Machine Learning stat.ML
|
|
URL: https://dx.doi.org/10.48550/arxiv.1906.01496 https://arxiv.org/abs/1906.01496
|
|
BASE
|
|
Hide details
|
|
3 |
STACKED NEURAL NETWORKS WITH PARAMETER SHARING FOR MULTILINGUAL LANGUAGE MODELING
|
|
|
|
In: http://infoscience.epfl.ch/record/272000 (2019)
|
|
BASE
|
|
Show details
|
|
4 |
Generalizing translation models in the probabilistic relevance framework
|
|
|
|
BASE
|
|
Show details
|
|
|
|