Measuring the Reliability of Hate Speech Annotations : The Case of the European Refugee Crisis

Roß, Björn LSF; Rist, Michael LSF; Carbonell, Guillermo LSF; Cabrera, Benjamin LSF; Kurowsky, Nils; Wojatzki, Michael Maximilian LSF

Some users of social media are spreading racist, sexist, and otherwise hateful content. For the purpose of training a hate speech detection system, the reliability of the annotations is crucial, but there is no universally agreed-upon definition. We collected potentially hateful messages and asked two groups of internet users to determine whether they were hate speech or not, whether they should be banned or not and to rate their degree of offensiveness. One of the groups was shown a definition prior to completing the survey. We aimed to assess whether hate speech can be annotated reliably, and the extent to which existing definitions are in accordance with subjective ratings. Our results indicate that showing users a definition caused them to partially align their own opinion with the definition but did not improve reliability, which was very low overall. We conclude that the presence of hate speech should perhaps not be considered a binary yes-or-no decision, and raters need more detailed instructions for the annotation.

Share and cite

Citation style:

Roß, Björn / Rist, Michael / Carbonell, Guillermo / et al: Measuring the Reliability of Hate Speech Annotations. The Case of the European Refugee Crisis. 2016.

Could not load citation form. Default citation form is displayed.

Rights

Use and reproduction:
This work may be used under a
CC BY-ND 4.0 LogoCreative Commons Attribution - NoDerivatives 4.0 License (CC BY-ND 4.0)
.

Export