DE eng

Search in the Catalogues and Directories

Page: 1 2 3 4 5 6 7 8 9 10
Hits 101 – 120 of 191

101
Enhancing biomedical word embeddings by retrofitting to verb clusters
Chiu, B; Baker, Simon; Palmer, M. - : Association for Computational Linguistics, 2019. : https://www.aclweb.org/anthology/W19-50, 2019. : BioNLP 2019 - SIGBioMed Workshop on Biomedical Natural Language Processing, Proceedings of the 18th BioNLP Workshop and Shared Task, 2019
BASE
Show details
102
Bayesian learning for neural dependency parsing
Shareghi, E; Li, Y; Zhu, Y. - : NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference, 2019
BASE
Show details
103
Modeling Language Variation and Universals: A Survey on Typological Linguistics for Natural Language Processing
Reichart, Roi; Shutova, Ekaterina; Korhonen, Anna-Leena. - : MIT Press - Journals, 2019. : COMPUTATIONAL LINGUISTICS, 2019
BASE
Show details
104
A systematic study of leveraging subword information for learning word representations
Zhu, Y; Korhonen, Anna-Leena; Vulić, I. - : NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies - Proceedings of the Conference, 2019
Abstract: The use of subword-level information (e.g., characters, character n-grams, morphemes) has become ubiquitous in modern word representation learning. Its importance is attested especially for morphologically rich languages which generate a large number of rare words. Despite a steadily increasing interest in such subword-informed word representations, their systematic comparative analysis across typologically diverse languages and different tasks is still missing. In this work, we deliver such a study focusing on the variation of two crucial components required for subword-level integration into word representation models: 1) segmentation of words into subword units, and 2) subword composition functions to obtain final word representations. We propose a general framework for learning subword-informed word representations that allows for easy experimentation with different segmentation and composition components, also including more advanced techniques based on position embeddings and self-attention. Using the unified framework, we run experiments over a large number of subword-informed word representation configurations (60 in total) on 3 tasks (general and rare word similarity, dependency parsing, fine-grained entity typing) for 5 languages representing 3 language types. Our main results clearly indicate that there is no ``one-size-fits-all'' configuration, as performance is both language- and task-dependent. We also show that configurations based on unsupervised segmentation (e.g., BPE, Morfessor) are sometimes comparable to or even outperform the ones based on supervised word segmentation.
URL: https://www.repository.cam.ac.uk/handle/1810/292619
https://doi.org/10.17863/CAM.39780
BASE
Hide details
105
Multilingual and cross-lingual graded lexical entailment
Glavaš, Goran; Vulić, Ivan; Ponzetto, Simone Paolo. - : Association for Computational Linguistics, 2019
BASE
Show details
106
How to (properly) evaluate cross-lingual word embeddings: On strong baselines, comparative analyses, and some misconceptions
Glavaš, Goran; Litschko, Robert; Ruder, Sebastian. - : Association for Computational Linguistics, 2019
BASE
Show details
107
A spreading activation framework for tracking conceptual complexity of texts
Hulpus, Ioana; Štajner, Sanja; Stuckenschmidt, Heiner. - : Association for Computational Linguistics, ACL, 2019
BASE
Show details
108
Cross-lingual semantic specialization via lexical relation induction
Glavaš, Goran; Vulić, Ivan; Korhonen, Anna. - : Association for Computational Linguistics, 2019
BASE
Show details
109
Generalized tuning of distributional word vectors for monolingual and cross-lingual lexical entailment
Vulić, Ivan; Glavaš, Goran. - : Association for Computational Linguistics, 2019
BASE
Show details
110
Informing unsupervised pretraining with external linguistic knowledge
Lauscher, Anne; Vulić, Ivan; Ponti, Edoardo Maria. - : Cornell University, 2019
BASE
Show details
111
Do we really need fully unsupervised cross-lingual embeddings?
Vulić, Ivan; Glavaš, Goran; Reichart, Roi. - : Association for Computational Linguistics, 2019
BASE
Show details
112
Modeling Language Variation and Universals: A Survey on Typological Linguistics for Natural Language Processing
In: Computational Linguistics, Vol 45, Iss 3, Pp 559-601 (2019) (2019)
BASE
Show details
113
Modeling Language Variation and Universals: A Survey on Typological Linguistics for Natural Language Processing
In: https://hal.archives-ouvertes.fr/hal-01856176 ; 2018 (2018)
BASE
Show details
114
Modeling Language Variation and Universals: A Survey on Typological Linguistics for Natural Language Processing ...
BASE
Show details
115
Bio-SimVerb ...
Chiu, Hon Wing; Pyysalo, Sampo; Vulic, Ivan. - : Apollo - University of Cambridge Repository, 2018
BASE
Show details
116
Adversarial Propagation and Zero-Shot Cross-Lingual Transfer of Word Vector Specialization ...
BASE
Show details
117
Post-Specialisation: Retrofitting Vectors of Words Unseen in Lexical Resources ...
BASE
Show details
118
Isomorphic Transfer of Syntactic Structures in Cross-Lingual NLP ...
Ponti, Edoardo; Reichart, Roi; Korhonen, Anna-Leena. - : Apollo - University of Cambridge Repository, 2018
BASE
Show details
119
Acquiring verb classes through bottom-up semantic verb clustering ...
Majewska, Olga; McCarthy, D; Vulić, I. - : Apollo - University of Cambridge Repository, 2018
BASE
Show details
120
Language Modeling for Morphologically Rich Languages: Character-Aware Modeling for Word-Level Prediction ...
Gerz, Daniela; Vulić, Ivan; Ponti, Edoardo. - : Apollo - University of Cambridge Repository, 2018
BASE
Show details

Page: 1 2 3 4 5 6 7 8 9 10

Catalogues
1
0
2
0
5
0
0
Bibliographies
3
0
0
0
0
0
1
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
180
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern