DE eng

Search in the Catalogues and Directories

Hits 1 – 8 of 8

1
Examining the Inductive Bias of Neural Language Models with Artificial Languages ...
BASE
Show details
2
Modeling the Unigram Distribution ...
Abstract: Read paper: https://www.aclanthology.org/2021.findings-acl.326 Abstract: The unigram distribution is the non-contextual probability of finding a specific word form in a corpus. While of central importance to the study of language, it is commonly approximated by each word's sample frequency in the corpus. This approach, being highly dependent on sample size, assigns zero probability to any out-of-vocabulary (oov) word form. As a result, it produces negatively biased probabilities for any oov word form, while positively biased probabilities to in-corpus words. In this work, we argue in favor of properly modeling the unigram distribution---claiming it should be a central task in natural language processing. With this in mind, we present a novel model for estimating it in a language (a neuralization of Goldwater et al.'s (2011) model) and show it produces much better estimates across a diverse set of 7 languages than the naïive use of neural character-level language models. ...
URL: https://dx.doi.org/10.48448/fx5z-4a29
https://underline.io/lecture/26417-modeling-the-unigram-distribution
BASE
Hide details
3
Language Model Evaluation Beyond Perplexity ...
BASE
Show details
4
Higher-order Derivatives of Weighted Finite-state Machines ...
BASE
Show details
5
On Finding the K-best Non-projective Dependency Trees ...
BASE
Show details
6
Determinantal Beam Search ...
BASE
Show details
7
Is Sparse Attention more Interpretable? ...
BASE
Show details
8
A Cognitive Regularizer for Language Modeling ...
BASE
Show details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
8
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern