DE eng

Search in the Catalogues and Directories

Page: 1 2
Hits 1 – 20 of 35

1
"Laughing at you or with you": The Role of Sarcasm in Shaping the Disagreement Space ...
BASE
Show details
2
Minimally-Supervised Morphological Segmentation using Adaptor Grammars with Linguistic Priors ...
BASE
Show details
3
Don't Go Far Off: An Empirical Study on Neural Poetry Translation ...
BASE
Show details
4
Figurative Language in Recognizing Textual Entailment ...
BASE
Show details
5
Multi-Task Learning and Adapted Knowledge Models for Emotion-Cause Extraction ...
BASE
Show details
6
Weakly-Supervised Methods for Suicide Risk Assessment: Role of Related Domains ...
BASE
Show details
7
Don't Go Far Off: An Empirical Study on Neural Poetry Translation ...
BASE
Show details
8
ENTRUST: Argument Reframing with Language Models and Entailment ...
BASE
Show details
9
Metaphor Generation with Conceptual Mappings ...
BASE
Show details
10
Implicit Premise Generation with Discourse-aware Commonsense Knowledge Models ...
Abstract: Anthology paper link: https://aclanthology.org/2021.emnlp-main.504/ Abstract: Enthymemes are defined as arguments where a premise or conclusion is left implicit. We tackle the task of generating the implicit premise in an enthymeme, which requires not only an understanding of the stated conclusion and premise but also additional inferences that could depend on commonsense knowledge. The largest available dataset for enthymemes (Habernal et al., 2018) consists of 1.7k samples, which is not large enough to train a neural text generation model. To address this issue, we take advantage of a similar task and dataset: Abductive reasoning in narrative text (Bhagavatula et al., 2020). However, we show that simply using a state-of-the-art seq2seq model fine-tuned on this data might not generate meaningful implicit premises associated with the given enthymemes. We demonstrate that encoding discourse-aware commonsense during fine-tuning improves the quality of the generated implicit premises and outperforms all other ...
Keyword: Language Models; Natural Language Processing; Semantic Evaluation; Sociolinguistics
URL: https://underline.io/lecture/37589-implicit-premise-generation-with-discourse-aware-commonsense-knowledge-models
https://dx.doi.org/10.48448/ggg5-ec35
BASE
Hide details
11
ENTRUST: Argument Reframing with Language Models and Entailment ...
NAACL 2021 2021; Chakrabarty, Tuhin; Hidey, Christopher. - : Underline Science Inc., 2021
BASE
Show details
12
Metaphor Generation with Conceptual Mappings ...
BASE
Show details
13
$R^3$: Reverse, Retrieve, and Rank for Sarcasm Generation with Commonsense Knowledge ...
BASE
Show details
14
Fact vs. Opinion: the Role of Argumentation Features in News Classification ...
BASE
Show details
15
DeSePtion: Dual Sequence Prediction and Adversarial Examples for Improved Fact-Checking ...
BASE
Show details
16
Interpreting Verbal Irony: Linguistic Strategies and the Connection to the Type of Semantic Incongruity ...
Ghosh, Debanjan; Musi, Elena; Upasani, Kartikeya. - : University of Mass Amherst, 2020
BASE
Show details
17
Interpreting Verbal Irony: Linguistic Strategies and the Connection to the Type of Semantic Incongruity
BASE
Show details
18
Interpreting Verbal Irony: Linguistic Strategies and the Connection to the Type of Semantic Incongruity
In: Proceedings of the Society for Computation in Linguistics (2020)
BASE
Show details
19
Interpreting Verbal Irony: Linguistic Strategies and the Connection to the Type of Semantic Incongruity ...
BASE
Show details
20
Sarcasm Analysis using Conversation Context ...
BASE
Show details

Page: 1 2

Catalogues
0
0
1
0
0
0
0
Bibliographies
1
0
0
0
0
0
0
0
0
Linked Open Data catalogues
0
Online resources
0
0
0
0
Open access documents
34
0
0
0
0
© 2013 - 2024 Lin|gu|is|tik | Imprint | Privacy Policy | Datenschutzeinstellungen ändern