DE eng

Search in the Catalogues and Directories

Hits 1 – 7 of 7

1
Value-aware Approximate Attention ...
Abstract: Anthology paper link: https://aclanthology.org/2021.emnlp-main.753/ Abstract: Following the success of dot-product attention in Transformers, numerous approximations have been recently proposed to address its quadratic complexity with respect to the input length. However, all approximations thus far have ignored the contribution of the value vectors to the quality of approximation. In this work, we argue that research efforts should be directed towards approximating the true output of the attention sub-layer, which includes the value vectors. We propose a valueaware objective, and show theoretically and empirically that an optimal approximation of a value-aware objective substantially outperforms an optimal approximation that ignores values, in the context of language modeling. Moreover, we show that the choice of kernel function for computing attention similarity can substantially affect the quality of sparse approximations, where kernel functions that are less skewed are more affected by the value vectors. ...
Keyword: Computational Linguistics; Machine Learning; Machine Learning and Data Mining; Natural Language Processing
URL: https://underline.io/lecture/37465-value-aware-approximate-attention
https://dx.doi.org/10.48448/tfks-ze53
BASE
Hide details
2
Memory-efficient Transformers via Top-k Attention ...
BASE
Show details
3
Achieving Model Robustness through Discrete Adversarial Training ...
BASE
Show details
4
COVR: A Test-Bed for Visually Grounded Compositional Generalization with Real Images ...
BASE
Show details
5
Transformer Feed-Forward Layers Are Key-Value Memories ...
BASE
Show details
6
What's in Your Head? Emergent Behaviour in Multi-Task Transformer Models ...
BASE
Show details
7
Finding needles in a haystack: Sampling Structurally-diverse Training Sets from Synthetic Data for Compositional Generalization ...
BASE
Show details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
7
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern