DE eng

Search in the Catalogues and Directories

Hits 1 – 6 of 6

1
Automatic acquisition of LFG resources for German - as good as it gets
In: Rehbein, Ines and van Genabith, Josef orcid:0000-0003-1322-7944 (2009) Automatic acquisition of LFG resources for German - as good as it gets. In: Lexical Functional Grammar 2009, 13-16 July 2009, Cambridge, UK. (2009)
BASE
Show details
2
TePaCoC - a testsuite for testing parser performance on complex German grammatical constructions
In: Kübler, Sandra and Rehbein, Ines (2009) TePaCoC - a testsuite for testing parser performance on complex German grammatical constructions. In: TLT 7 - 7th International Workshop on Treebanks and Linguistic Theories, 23-24 January 2009, Groningen, The Netherlands. (2009)
Abstract: Traditionally, parsers are evaluated against gold standard test data. This can cause problems if there is a mismatch between the data structures and representations used by the parser and the gold standard. A particular case in point is German, for which two treebanks (TiGer and TüBa-D/Z) are available with highly different annotation schemes for the acquisition of (e.g.) PCFG parsers. The differences between the TiGer and TüBa-D/Z annotation schemes make fair and unbiased parser evaluation difficult [7, 9, 12]. The resource (TEPACOC) presented in this paper takes a different approach to parser evaluation: instead of providing evaluation data in a single annotation scheme, TEPACOC uses comparable sentences and their annotations for 5 selected key grammatical phenomena (with 20 sentences each per phenomena) from both TiGer and TüBa-D/Z resources. This provides a 2 times 100 sentence comparable testsuite which allows us to evaluate TiGer-trained parsers against the TiGer part of TEPACOC, and TüBa-D/Z-trained parsers against the TüBa-D/Z part of TEPACOC for key phenomena, instead of comparing them against a single (and potentially biased) gold standard. To overcome the problem of inconsistency in human evaluation and to bridge the gap between the two different annotation schemes, we provide an extensive error classification, which enables us to compare parser output across the two different treebanks. In the remaining part of the paper we present the testsuite and describe the grammatical phenomena covered in the data. We discuss the different annotation strategies used in the two treebanks to encode these phenomena and present our error classification of potential parser errors.
Keyword: Machine translating; parsing
URL: http://doras.dcu.ie/15179/
BASE
Hide details
3
Treebank annotation schemes and parser evaluation for German
In: Rehbein, Ines and van Genabith, Josef (2007) Treebank annotation schemes and parser evaluation for German. In: EMNLP-CoNLL 2007 - Joint Meeting of the Conference on Empirical Methods in Natural Language Processing and the Conference on Computational Natural Language Learning, 28-30 June 2007, Prague, Czech Republic. (2007)
BASE
Show details
4
Evaluating evaluation measures
In: Rehbein, Ines and van Genabith, Josef (2007) Evaluating evaluation measures. In: NODALIDA 2007 - 16th Nordic Conference on Computational Linguistic, 25-26 May 2007, Tartu, Estonia. (2007)
BASE
Show details
5
Why is it so difficult to compare treebanks? TIGER and TüBa-D/Z revisited
In: Rehbein, Ines and van Genabith, Josef (2007) Why is it so difficult to compare treebanks? TIGER and TüBa-D/Z revisited. In: TLT 2007 - The 6th International Workshop on Treebanks and Linguistic Theories, 7-8 December, 2007, Bergen, Norway. (2007)
BASE
Show details
6
German particle verbs and pleonastic prepositions
In: Rehbein, Ines and van Genabith, Josef (2006) German particle verbs and pleonastic prepositions. In: Third ACL-SIGSEM Workshop on Prepositions, 3 April 2006, Trento, Italy. (2006)
BASE
Show details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
6
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern