DE eng

Search in the Catalogues and Directories

Page: 1 2 3 4 5 6...52
Hits 21 – 40 of 1.029

21
A Prototype Free/Open-Source Morphological Analyser and Generator for Sakha ...
BASE
Show details
22
Automatic Error Type Annotation for Arabic ...
BASE
Show details
23
Developing Conversational Data and Detection of Conversational Humor in Telugu ...
BASE
Show details
24
An Information-Theoretic Characterization of Morphological Fusion ...
BASE
Show details
25
Cross-document Event Identity via Dense Annotation ...
BASE
Show details
26
Navigating the Kaleidoscope of COVID-19 Misinformation Using Deep Learning ...
BASE
Show details
27
(Mis)alignment Between Stance Expressed in Social Media Data and Public Opinion Surveys ...
BASE
Show details
28
Adversarial Regularization as Stackelberg Game: An Unrolled Optimization Approach ...
BASE
Show details
29
Rewards with Negative Examples for Reinforced Topic-Focused Abstractive Summarization ...
BASE
Show details
30
Distantly-Supervised Named Entity Recognition with Noise-Robust Learning and Language Model Augmented Self-Training ...
BASE
Show details
31
Low-Resource Dialogue Summarization with Domain-Agnostic Multi-Source Pretraining ...
BASE
Show details
32
HittER: Hierarchical Transformers for Knowledge Graph Embeddings ...
BASE
Show details
33
Ara-Women-Hate: The first Arabic Hate Speech corpus regarding Women ...
BASE
Show details
34
Detecting Gender Bias using Explainability ...
BASE
Show details
35
HETFORMER: Heterogeneous Transformer with Sparse Attention for Long-Text Extractive Summarization ...
Abstract: Anthology paper link: https://aclanthology.org/2021.emnlp-main.13/ Abstract: To capture the semantic graph structure from raw text, most existing summarization approaches are built on GNNs with a pre-trained model. However, these methods suffer from cumbersome procedures and inefficient computations for long-text documents. To mitigate these issues, this paper proposes HETFORMER, a Transformer-based pre-trained model with multi-granularity sparse attentions for long-text extractive summarization. Specifically, we model different types of semantic nodes in raw text as a potential heterogeneous graph and directly learn heterogeneous relationships (edges) among nodes by Transformer. Extensive experiments on both single- and multi-document summarization tasks show that HETFORMER achieves state-of-the-art performance in Rouge F1 while using less memory and fewer parameters. ...
Keyword: Computational Linguistics; Language Models; Machine Learning; Machine Learning and Data Mining; Natural Language Processing; Text Summarization
URL: https://underline.io/lecture/37574-hetformer-heterogeneous-transformer-with-sparse-attention-for-long-text-extractive-summarization
https://dx.doi.org/10.48448/za5a-2696
BASE
Hide details
36
Blindness to Modality Helps Entailment Graph Mining ...
BASE
Show details
37
Efficient-FedRec: Efficient Federated Learning Framework for Privacy-Preserving News Recommendation ...
BASE
Show details
38
Characterizing Test Anxiety on Social Media ...
BASE
Show details
39
Not All Negatives are Equal: Label-Aware Contrastive Loss for Fine-grained Text Classification ...
BASE
Show details
40
Improving Graph-based Sentence Ordering with Iteratively Predicted Pairwise Orderings ...
BASE
Show details

Page: 1 2 3 4 5 6...52

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
1.029
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern