Abstract
Some users of social media are spreading racist, sexist, and otherwise hateful content. For the purpose of training a hate speech detection system, the reliability of the annotations is crucial, but there is no universally agreed-upon definition. We collected potentially hateful messages and asked two groups of internet users to determine whether they were hate speech or not, whether they should be banned or not and to rate their degree of offensiveness. One of the groups was shown a definition prior to completing the survey. We aimed to assess whether hate speech can be annotated reliably, and the extent to which existing definitions are in accordance with subjective ratings. Our results indicate that showing users a definition caused them to partially align their own opinion with the definition but did not improve reliability, which was very low overall. We conclude that the presence of hate speech should perhaps not be considered a binary yes-or-no decision, and raters need more detailed instructions for the annotation.
Original language | English |
---|---|
Title of host publication | NLP4CMC III: 3rd Workshop on Natural Language Processing for Computer-Mediated Communication |
Editors | Stefanie Dipper |
Publisher | Ruhr-Universitat Bochum |
Pages | 6-9 |
Number of pages | 4 |
Publication status | Published - 9 Nov 2016 |
Event | 3rd Workshop on Natural Language Processing for Computer-Mediated Communication / Social Media - Bochum, Germany Duration: 22 Sep 2016 → 22 Sep 2016 https://sites.google.com/site/nlp4cmc2016/ |
Publication series
Name | Bochumer Linguistische Arbeitsberichte |
---|---|
Volume | 17 |
ISSN (Electronic) | 2190-0949 |
Workshop
Workshop | 3rd Workshop on Natural Language Processing for Computer-Mediated Communication / Social Media |
---|---|
Abbreviated title | NLP4CMC 2016 |
Country/Territory | Germany |
City | Bochum |
Period | 22/09/16 → 22/09/16 |
Internet address |
Keywords
- hate speech
- hate speech detection
- social media
- social media analytics
- natural language processing
Fingerprint
Dive into the research topics of 'Measuring the Reliability of Hate Speech Annotations: The Case of the European Refugee Crisis'. Together they form a unique fingerprint.Profiles
-
Bjorn Ross
- School of Informatics - Lecturer in Computational Social Science
- Institute of Language, Cognition and Computation
- Language, Interaction and Robotics
Person: Academic: Research Active