DE eng

Search in the Catalogues and Directories

Hits 1 – 4 of 4

1
Automatic Selection of Context Configurations for Improved Class-Specific Word Representations ...
Vulić, Ivan; Schwartz, Roy; Rappoport, Ari. - : Apollo - University of Cambridge Repository, 2017
BASE
Show details
2
Morph-fitting: Fine-tuning word vector spaces with simple language-specific rules ...
Vulic, Ivan; Mrkšic, N; Reichart, R. - : Apollo - University of Cambridge Repository, 2017
BASE
Show details
3
Morph-fitting: Fine-tuning word vector spaces with simple language-specific rules
Vulic, Ivan; Mrkšic, N; Reichart, R. - : Association for Computational Linguistics, 2017. : ACL 2017 - 55th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers), 2017
BASE
Show details
4
Automatic Selection of Context Configurations for Improved Class-Specific Word Representations
Rappoport, Ari; Reichart, Roi; Korhonen, Anna-Leena; Schwartz, Roy; Vulić, Ivan. - : Association for Computational Linguistics, 2017. : https://arxiv.org/pdf/1608.05528.pdf, 2017. : Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL 2017), 2017
Abstract: This paper is concerned with identifying contexts useful for training word representation models for different word classes such as adjectives (A), verbs (V), and nouns (N). We introduce a simple yet effective framework for an automatic selection of {\em class-specific context configurations}. We construct a context configuration space based on universal dependency relations between words, and efficiently search this space with an adapted beam search algorithm. In word similarity tasks for each word class, we show that our framework is both effective and efficient. Particularly, it improves the Spearman's rho correlation with human scores on SimLex-999 over the best previously proposed class-specific contexts by 6 (A), 6 (V) and 5 (N) rho points. With our selected context configurations, we train on only 14% (A), 26.2% (V), and 33.6% (N) of all dependency-based contexts, resulting in a reduced training time. Our results generalise: we show that the configurations our algorithm learns for one English training setup outperform previously proposed context types in another training setup for English. Moreover, basing the configuration space on universal dependencies, it is possible to transfer the learned configurations to German and Italian. We also demonstrate improved per-class results over other context types in these two languages.
Keyword: Context configurations; Context selection; Multilinguality; Transfer learning; Word representations
URL: https://doi.org/10.17863/CAM.10795
https://www.repository.cam.ac.uk/handle/1810/269692
BASE
Hide details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
4
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern