DE eng

Search in the Catalogues and Directories

Hits 1 – 18 of 18

1
Universal Dependencies 2.9
Zeman, Daniel; Nivre, Joakim; Abrams, Mitchell. - : Universal Dependencies Consortium, 2021
BASE
Show details
2
Universal Dependencies 2.8.1
Zeman, Daniel; Nivre, Joakim; Abrams, Mitchell. - : Universal Dependencies Consortium, 2021
BASE
Show details
3
Universal Dependencies 2.8
Zeman, Daniel; Nivre, Joakim; Abrams, Mitchell. - : Universal Dependencies Consortium, 2021
BASE
Show details
4
A Digital Corpus of St. Lawrence Island Yupik ...
BASE
Show details
5
Akuzipik/Yupik (St. Lawrence Island, Alaska, USA; Chukotka, Russia) - Language Snapshot ...
Koonooka, Christopher Petuwaq; Schreiner, Sylvia L.R.; Soldati, Giulia Masella. - : Language Documentation and Description, 2021
BASE
Show details
6
A digital corpus of St. Lawrence Island Yupik for the Yupik Community
BASE
Show details
7
Morphology Matters: A Multilingual Language Modeling Analysis ...
BASE
Show details
8
Multidirectional leveraging for computational morphology and language documentation and revitalization
Schreiner, Sylvia L. R.; Schwartz, Lane; Hunt, Benjamin. - : University of Hawaii Press, 2020
BASE
Show details
9
Multidirectional leveraging for computational morphology and language documentation and revitalization
Schreiner, Sylvia L. R.; Schwartz, Lane; Hunt, Benjamin. - : University of Hawaii Press, 2020
BASE
Show details
10
Community-Focused Language Documentation in Support of Language Education and Revitalization for St. Lawrence Island Yupik
Schwartz, Lane; Schreiner, Sylvia L.R.; Chen, Emily. - : Centre interuniversitaire d’études et de recherches autochtones (CIÉRA), 2019. : Érudit, 2019
BASE
Show details
11
Unsupervised Grammar Induction with Depth-bounded PCFG ...
BASE
Show details
12
Liinnaqumalghiit: A web-based tool for addressing orthographic transparency in St. Lawrence Island/Central Siberian Yupik
Schwartz, Lane; Chen, Emily. - : University of Hawaii Press, 2017
BASE
Show details
13
Liinnaqumalghiit: A web-based tool for addressing orthographic transparency in St. Lawrence Island/Central Siberian Yupik
Schwartz, Lane; Chen, Emily. - : University of Hawaii Press, 2017
BASE
Show details
14
Compiling contextualized lists of frequent vocabulary from user- supplied corpora using natural language processing techniques
Abdar, Omid. - 2016
BASE
Show details
15
Better splitting algorithms for parallel corpus processing
In: The Prague bulletin of mathematical linguistics. - Praha : Univ. (2012) 98, 109-119
BLLDB
OLC Linguistik
Show details
16
An incremental syntactic language model for statistical phrase-based translation.
Abstract: University of Minnesota Ph.D. dissertation. February 2012. Major: Computer science. Advisor: William Schuler. 1 computer file (PDF); xvii, 238 pages, appendices A-B. ; Modern machine translation techniques typically incorporate both a translation model, which guides how individual words and phrases can be translated, and a language model (LM), which promotes fluency as translated words and phrases are combined into a translated sentence. Most attempts to inform the translation process with linguistic knowledge have focused on infusing syntax into translation models. We present a novel technique for incorporating syntactic knowledge as a language model in the context of statistical phrase-based machine translation (Koehn et al., 2003), one of the most widely used modern translation paradigms. The major contributions of this work are as follows: #15; We present a formal definition of an incremental syntactic language model as a Hierarchical Hidden Markov Model (HHMM), and detail how this model is estimated from a treebank corpus of labelled data. #15; The HHMM syntactic language model has been used in prior work involving parsing, speech recognition, and semantic role labelling. We present the first complete algorithmic definition of the HHMM as a language model. #15; We develop a novel and general method for incorporating any generative incremental language model into phrase-based machine translation. We integrate our HHMM incremental syntactic language model into Moses, the prevailing phrase-based decoder. #15; We present empirical results that demonstrate substantial improvements in perplexity for our syntactic language model over traditional n-gram language models; we also present empirical results on a constrained Urdu-English translation task that demonstrate the use of our syntactic LM.A standard measure of language model quality is average per-word perplexity. We present empirical results evaluating perplexity of various n-gram language models and our syntactic language model on both in-domain and out-of-domain test sets. On an in-domain test set, a traditional 5-gram language model trained on the same data as our syntactic language model outperforms the syntactic language model in terms of perplexity. We find that interpolating the 5-gram LM with the syntactic LM results in improved perplexity results, a 10% absolute reduction in perplexity compared to the 5-gram LM alone. On an out-of-domain test set, we find that our syntactic LM substantially outperforms all other LMs trained on the same training data. The syntactic LM demonstrates a 58% absolute reduction in perplexity over a 5-gram language model trained on the same training data. On this same out-of-domain test set, we further show that interpolating our syntactic language model with a large Gigaword-scale 5-gram language model results in the best overall perplexity results — a 61% absolute reduction in perplexity compared to the Gigaword-scale 5-gram language model alone, a 76% absolute reduction in perplexity compared to the syntactic LM alone, and a 90% absolute reduction in perplexity compared to the original smaller 5-gram language model. A language model with low perplexity is a theoretically good model of the language; it is expected that using an LM with low perplexity as a component of a machine translation system should result in more fluent translations. We present empirical results on a constrained Urdu-English translation task and perform an informal manual evaluation of translation results which suggests that the use of our incremental syntactic language model is indeed serving to guide the translation algorithm towards more fluent target language translations.
Keyword: Computer Science; Hierarchical hidden Markov model; Machine translation; Phrase-based translation; Syntactic language model; Syntax
URL: http://purl.umn.edu/121791
BASE
Hide details
17
Incremental Syntactic Language Models for Phrase-Based Translation
In: DTIC (2011)
BASE
Show details
18
Hierarchical phrase-based grammar extraction in Joshua : suffix arrays and prefix trees
In: The Prague bulletin of mathematical linguistics. - Praha : Univ. (2010) 93, 157-166
BLLDB
OLC Linguistik
Show details

Catalogues
0
0
2
0
0
0
0
Bibliographies
2
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
16
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern