DE eng

Search in the Catalogues and Directories

Hits 1 – 5 of 5

1
Multi-Head Self-Attention with Role-Guided Masks ...
Abstract: The state of the art in learning meaningful semantic representations of words is the Transformer model and its attention mechanisms. Simply put, the attention mechanisms learn to attend to specific parts of the input dispensing recurrence and convolutions. While some of the learned attention heads have been found to play linguistically interpretable roles, they can be redundant or prone to errors. We propose a method to guide the attention heads towards roles identified in prior work as important. We do this by defining role-specific masks to constrain the heads to attend to specific parts of the input, such that different heads are designed to play different roles. Experiments on text classification and machine translation using 7 different datasets show that our method outperforms competitive attention-based, CNN, and RNN baselines. ... : Accepted at ECIR@2021 ...
Keyword: Computation and Language cs.CL; FOS Computer and information sciences
URL: https://arxiv.org/abs/2012.12366
https://dx.doi.org/10.48550/arxiv.2012.12366
BASE
Hide details
2
Generating Fact Checking Explanations ...
BASE
Show details
3
Part of Speech Based Term Weighting for Information Retrieval ...
Lioma, Christina; Blanco, Roi. - : arXiv, 2017
BASE
Show details
4
Non-Compositional Term Dependence for Information Retrieval ...
BASE
Show details
5
Part of speech N-grams for information retrieval
BASE
Show details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
5
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern