DE eng

Search in the Catalogues and Directories

Page: 1 2 3 4 5...13
Hits 1 – 20 of 246

1
Automatic Error Type Annotation for Arabic ...
BASE
Show details
2
Navigating the Kaleidoscope of COVID-19 Misinformation Using Deep Learning ...
BASE
Show details
3
HittER: Hierarchical Transformers for Knowledge Graph Embeddings ...
BASE
Show details
4
Detecting Gender Bias using Explainability ...
BASE
Show details
5
HETFORMER: Heterogeneous Transformer with Sparse Attention for Long-Text Extractive Summarization ...
BASE
Show details
6
Not All Negatives are Equal: Label-Aware Contrastive Loss for Fine-grained Text Classification ...
BASE
Show details
7
Contrastive Code Representation Learning ...
BASE
Show details
8
Unsupervised Multi-View Post-OCR Error Correction With Language Models ...
BASE
Show details
9
AttentionRank: Unsupervised Keyphrase Extraction using Self and Cross Attentions ...
BASE
Show details
10
Automatic Fact-Checking with Document-level Annotations using BERT and Multiple Instance Learning ...
BASE
Show details
11
Towards the Early Detection of Child Predators in Chat Rooms: A BERT-based Approach ...
BASE
Show details
12
Semantic Categorization of Social Knowledge for Commonsense Question Answering ...
BASE
Show details
13
Pre-train or Annotate? Domain Adaptation with a Constrained Budget ...
BASE
Show details
14
Stepmothers are mean and academics are pretentious: What do pretrained language models learn about you? ...
BASE
Show details
15
CLIFF: Contrastive Learning for Improving Faithfulness and Factuality in Abstractive Summarization ...
BASE
Show details
16
Automatic Text Evaluation through the Lens of Wasserstein Barycenters ...
BASE
Show details
17
Combining sentence and table evidence to predict veracity of factual claims using TaPaS and RoBERTa ...
BASE
Show details
18
Meta Distant Transfer Learning for Pre-trained Language Models ...
Abstract: Anthology paper link: https://aclanthology.org/2021.emnlp-main.768/ Abstract: With the wide availability of Pre-trained Language Models (PLMs), multi-task fine-tuning across domains has been extensively applied. For tasks related to distant domains with different class label sets, PLMs may memorize non- transferable knowledge for the target domain and suffer from negative transfer. Inspired by meta-learning, we propose the Meta Distant Transfer Learning (Meta-DTL) framework to learn the cross-task knowledge for PLM-based methods. Meta-DTL first employs task representation learning to mine implicit relations among multiple tasks and classes. Based on the results, it trains a PLM-based meta-learner to capture the transferable knowledge across tasks. The weighted maximum entropy regularizers are proposed to make meta-learner more task-agnostic and unbiased. Finally, the meta-learner can be fine-tuned to fit each task with better parameter initialization. We evaluate Meta-DTL using both BERT and ALBERT on seven ...
Keyword: Computational Linguistics; Language Models; Machine Learning; Machine Learning and Data Mining; Natural Language Processing
URL: https://dx.doi.org/10.48448/xp16-y443
https://underline.io/lecture/37379-meta-distant-transfer-learning-for-pre-trained-language-models
BASE
Hide details
19
How to Train BERT with an Academic Budget ...
BASE
Show details
20
Temporal Adaptation of BERT and Performance on Downstream Document Classification: Insights from Social Media ...
BASE
Show details

Page: 1 2 3 4 5...13

Catalogues
0
0
0
0
0
0
0
Bibliographies
0
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
246
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern