On the Impact of Word Representation in Hate Speech and Offensive Language Detection and Explanation

Author(s):  
Ruijia (Roger) Hu ◽  
Wyatt Dorris ◽  
Nishant Vishwamitra ◽  
Feng Luo ◽  
Matthew Costello
Author(s):  
Vildan Mercan ◽  
Akhtar Jamil ◽  
Alaa Ali Hameed ◽  
Irfan Ahmed Magsi ◽  
Sibghatullah Bazai ◽  
...  

2020 ◽  
Author(s):  
Hammad Rizwan ◽  
Muhammad Haroon Shakeel ◽  
Asim Karim

Author(s):  
Dr. Sweeta Bansal

As we know that the social crowd is increasing day by day, so is the hatred among them online. This hatred gives rise to hate speech/comments that are passed to one another online. Recently, the hate speech has increased so much that we need a way to stop them or at least contain it to minimum. Keeping this problem in mind, we have introduced a way in which we can detect the class of comments that are posted online and stop its spread if it belongs to hateful category. We have used Natural Language Processing methods and Logistic Regression algorithm to achieve our goal.


Author(s):  
Gretel Liz De la Peña Sarracén ◽  
Paolo Rosso

AbstractThe proliferation of harmful content on social media affects a large part of the user community. Therefore, several approaches have emerged to control this phenomenon automatically. However, this is still a quite challenging task. In this paper, we explore the offensive language as a particular case of harmful content and focus our study in the analysis of keywords in available datasets composed of offensive tweets. Thus, we aim to identify relevant words in those datasets and analyze how they can affect model learning. For keyword extraction, we propose an unsupervised hybrid approach which combines the multi-head self-attention of BERT and a reasoning on a word graph. The attention mechanism allows to capture relationships among words in a context, while a language model is learned. Then, the relationships are used to generate a graph from what we identify the most relevant words by using the eigenvector centrality. Experiments were performed by means of two mechanisms. On the one hand, we used an information retrieval system to evaluate the impact of the keywords in recovering offensive tweets from a dataset. On the other hand, we evaluated a keyword-based model for offensive language detection. Results highlight some points to consider when training models with available datasets.


Author(s):  
Patricia Chiril ◽  
Endang Wahyu Pamungkas ◽  
Farah Benamara ◽  
Véronique Moriceau ◽  
Viviana Patti

AbstractHate Speech and harassment are widespread in online communication, due to users' freedom and anonymity and the lack of regulation provided by social media platforms. Hate speech is topically focused (misogyny, sexism, racism, xenophobia, homophobia, etc.), and each specific manifestation of hate speech targets different vulnerable groups based on characteristics such as gender (misogyny, sexism), ethnicity, race, religion (xenophobia, racism, Islamophobia), sexual orientation (homophobia), and so on. Most automatic hate speech detection approaches cast the problem into a binary classification task without addressing either the topical focus or the target-oriented nature of hate speech. In this paper, we propose to tackle, for the first time, hate speech detection from a multi-target perspective. We leverage manually annotated datasets, to investigate the problem of transferring knowledge from different datasets with different topical focuses and targets. Our contribution is threefold: (1) we explore the ability of hate speech detection models to capture common properties from topic-generic datasets and transfer this knowledge to recognize specific manifestations of hate speech; (2) we experiment with the development of models to detect both topics (racism, xenophobia, sexism, misogyny) and hate speech targets, going beyond standard binary classification, to investigate how to detect hate speech at a finer level of granularity and how to transfer knowledge across different topics and targets; and (3) we study the impact of affective knowledge encoded in sentic computing resources (SenticNet, EmoSenticNet) and in semantically structured hate lexicons (HurtLex) in determining specific manifestations of hate speech. We experimented with different neural models including multitask approaches. Our study shows that: (1) training a model on a combination of several (training sets from several) topic-specific datasets is more effective than training a model on a topic-generic dataset; (2) the multi-task approach outperforms a single-task model when detecting both the hatefulness of a tweet and its topical focus in the context of a multi-label classification approach; and (3) the models incorporating EmoSenticNet emotions, the first level emotions of SenticNet, a blend of SenticNet and EmoSenticNet emotions or affective features based on Hurtlex, obtained the best results. Our results demonstrate that multi-target hate speech detection from existing datasets is feasible, which is a first step towards hate speech detection for a specific topic/target when dedicated annotated data are missing. Moreover, we prove that domain-independent affective knowledge, injected into our models, helps finer-grained hate speech detection.


2020 ◽  
Vol 117 (37) ◽  
pp. 22800-22804
Author(s):  
Amalia Álvarez-Benjumea ◽  
Fabian Winter

Terrorist attacks often fuel online hate and increase the expression of xenophobic and antiminority messages. Previous research has focused on the impact of terrorist attacks on prejudiced attitudes toward groups linked to the perpetrators as the cause of this increase. We argue that social norms can contain the expression of prejudice after the attacks. We report the results of a combination of a natural and a laboratory-in-the-field (lab-in-the-field) experiment in which we exploit data collected about the occurrence of two consecutive Islamist terrorist attacks in Germany, the Würzburg and Ansbach attacks, in July 2016. The experiment compares the effect of the terrorist attacks in hate speech toward refugees in contexts where a descriptive norm against the use of hate speech is evidently in place to contexts in which the norm is ambiguous because participants observe antiminority comments. Hate toward refugees, but not toward other minority groups, increased as a result of the attacks only in the absence of a strong norm. These results imply that attitudinal changes due to terrorist attacks are more likely to be voiced if norms erode.


The vocabulary of a language is a variable quantity, it is constantly changing, responding to the needs of life and reflecting its new realities. The events taking place in the South-East of Ukraine since March 2014 have significantly changed the usual picture of the world of the parties involved in this conflict, led to a new interpretation of reality, the emergence of new mental constructs, objectified in the language using a number of lexical innovations, most of which fall under the definition of „hate speech”. The purpose of this article is to try to examine the impact of the armed conflict in the South-East of Ukraine on the emergence of lexical innovations in the Russian language, to identify ways of forming new units and their main thematic clusters. The material for the work was neoplasms recorded in electronic Russian and Russian-speaking Ukrainian mass media, as well as selected from social networks and videos. The analysis showed that in the context of the armed conflict in the South-East of Ukraine, the characteristic manifestations of „hate speech” are mainly numerous new categories-labels with a pronounced conflict potential. The priority in this regard is offensive and derogatory nominations of representatives of the opposite camp, taking into account their worldview / ideological, national / ethnic, territorial / regional characteristics. The military jargon has also undergone a significant update, incorporating not only the reactualized slangisms of the era of the Afghan campaign of 1979-89, but also lexical innovations caused by the military and political realities of the current armed conflict in the Donbas. Neologisms are formed in accordance with the existing methods in the Russian language (word formation, semantic derivation, borrowing). At the same time, non-standard word-forming techniques are also used (language play, homophony, etc.).


Author(s):  
Zhiwei Gao ◽  
Shuntaro Yada ◽  
Shoko Wakamiya ◽  
Eiji Aramaki

Sign in / Sign up

Export Citation Format

Share Document