1 |
Analyzing and Learning the Language for Different Types of Harassment
|
|
|
|
In: Publications (2020)
|
|
BASE
|
|
Show details
|
|
2 |
Analyzing and Learning the Language for Different Types of Harassment
|
|
|
|
In: Amit P. Sheth (2020)
|
|
BASE
|
|
Show details
|
|
3 |
A Quality Type-Aware Annotated Corpus and Lexicon for Harassment Research
|
|
|
|
In: Valerie Shalin (2020)
|
|
BASE
|
|
Show details
|
|
4 |
Analyzing and Learning the Language for Different Types of Harassment
|
|
|
|
In: Computer Science and Engineering Faculty Publications (2020)
|
|
BASE
|
|
Show details
|
|
5 |
Analyzing and Learning the Language for Different Types of Harassment
|
|
|
|
In: Amit P. Sheth (2020)
|
|
Abstract:
THIS ARTICLE USES WORDS OR LANGUAGE THAT IS CONSIDERED PROFANE, VULGAR, OR OFFENSIVE BY SOME READERS. The presence of a significant amount of harassment in user-generated content and its negative impact calls for robust automatic detection approaches. This requires the identification of different types of harassment. Earlier work has classified harassing language in terms of hurtfulness, abusiveness, sentiment, and profanity. However, to identify and understand harassment more accurately, it is essential to determine the contextual type that captures the interrelated conditions in which harassing language occurs. In this paper we introduce the notion of contextual type in harassment by distinguishing between five contextual types: (i) sexual, (ii) racial, (iii) appearance-related, (iv) intellectual and (v) political. We utilize an annotated corpus from Twitter distinguishing these types of harassment. We study the context of each kind to shed light on the linguistic meaning, interpretation, and distribution, with results from two lines of investigation: an extensive linguistic analysis, and the statistical distribution of uni-grams. We then build type- aware classifiers to automate the identification of type-specific harassment. Our experiments demonstrate that these classifiers provide competitive accuracy for identifying and analyzing harassment on social media. We present extensive discussion and significant observations about the effectiveness of type-aware classifiers using a detailed comparison setup, providing insight into the role of type-dependent features.
|
|
Keyword:
Computer Engineering; Electrical and Computer Engineering
|
|
URL: https://works.bepress.com/amit_sheth/635
|
|
BASE
|
|
Hide details
|
|
6 |
Analyzing and Learning the Language for Different Types of Harassment
|
|
|
|
In: Valerie Shalin (2020)
|
|
BASE
|
|
Show details
|
|
7 |
Analyzing and Learning the Language for Different Types of Harassment
|
|
|
|
In: Krishnaprasad Thirunarayan (2020)
|
|
BASE
|
|
Show details
|
|
8 |
A Quality Type-Aware Annotated Corpus and Lexicon for Harassment Research
|
|
|
|
In: Amit P. Sheth (2019)
|
|
BASE
|
|
Show details
|
|
9 |
A Quality Type-aware Annotated Corpus and Lexicon for Harassment Research ...
|
|
|
|
BASE
|
|
Show details
|
|
10 |
A Quality Type-Aware Annotated Corpus and Lexicon for Harassment Research
|
|
|
|
In: Faculty Publications (2018)
|
|
BASE
|
|
Show details
|
|
11 |
RQUERY: Rewriting Natural Language Queries on Knowledge Graphs to Alleviate the Vocabulary Mismatch Problem
|
|
|
|
In: Publications (2017)
|
|
BASE
|
|
Show details
|
|
12 |
RQUERY: Rewriting Natural Language Queries on Knowledge Graphs to Alleviate the Vocabulary Mismatch Problem
|
|
|
|
In: Kno.e.sis Publications (2017)
|
|
BASE
|
|
Show details
|
|
13 |
RQUERY: Rewriting Natural Language Queries on Knowledge Graphs to Alleviate the Vocabulary Mismatch Problem
|
|
|
|
In: Amit P. Sheth (2017)
|
|
BASE
|
|
Show details
|
|
|
|