DE eng

Search in the Catalogues and Directories

Hits 1 – 6 of 6

1
Summarize-then-Answer: Generating Concise Explanations for Multi-hop Reading Comprehension ...
BASE
Show details
2
SHAPE: Shifted Absolute Position Embedding for Transformers ...
BASE
Show details
3
Incorporating Residual and Normalization Layers into Analysis of Masked Language Models ...
Abstract: Anthology paper link: https://aclanthology.org/2021.emnlp-main.373/ Abstract: Transformer architecture has become ubiquitous in the natural language processing field. To interpret the Transformer-based models, their attention patterns have been extensively analyzed. However, the Transformer architecture is not only composed of the multi-head attention; other components can also contribute to Transformers’ progressive performance. In this study, we extended the scope of the analysis of Transformers from solely the attention patterns to the whole attention block, i.e., multi-head attention, residual connection, and layer normalization. Our analysis of Transformer-based masked language models shows that the token-to-token interaction performed via attention has less impact on the intermediate representations than previously assumed. These results provide new intuitive explanations of existing reports; for example, discarding the learned attention patterns tends not to adversely affect the performance. The codes ...
Keyword: Computational Linguistics; Language Models; Machine Learning; Machine Learning and Data Mining; Natural Language Processing
URL: https://dx.doi.org/10.48448/jh7m-qw81
https://underline.io/lecture/38024-incorporating-residual-and-normalization-layers-into-analysis-of-masked-language-models
BASE
Hide details
4
Pseudo Zero Pronoun Resolution Improves Zero Anaphora Resolution ...
BASE
Show details
5
Exploring Methods for Generating Feedback Comments for Writing Learning ...
BASE
Show details
6
Transformer-based Lexically Constrained Headline Generation ...
BASE
Show details

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
6
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern