The risk of racial bias in hate speech detection

Abstract

We investigate how annotators insensitivity to differences in dialect can lead to racial bias in automatic hate speech detection models, potentially amplifying harm against minority populations. We first uncover unexpected correlations between surface markers of African

Publication
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics

Main contributions

Coming soon

Limitations

Coming soon

Future Directions

Coming soon

Related