PapersFlow Research Brief

Physical Sciences · Computer Science

Hate Speech and Cyberbullying Detection
Research Guide

What is Hate Speech and Cyberbullying Detection?

Hate Speech and Cyberbullying Detection is the application of machine learning, natural language processing, and deep learning techniques to automatically identify and categorize hate speech, offensive language, and cyberbullying in social media platforms such as Twitter.

This field encompasses 40,950 works focused on automated detection of abusive content to mitigate online harassment. Techniques include lexical methods, machine learning classifiers, and deep learning models trained on social media data. Davidson et al. (2017) highlighted challenges in distinguishing hate speech from other offensive language, noting low precision in lexical approaches.

Topic Hierarchy

100%
graph TD D["Physical Sciences"] F["Computer Science"] S["Artificial Intelligence"] T["Hate Speech and Cyberbullying Detection"] D --> F F --> S S --> T style T fill:#DC5238,stroke:#c4452e,stroke-width:2px
Scroll to zoom • Drag to pan
41.0K
Papers
N/A
5yr Growth
273.4K
Total Citations

Research Sub-Topics

Multilingual Hate Speech Detection

This sub-topic covers the development of language-agnostic and multilingual models for identifying hate speech across diverse languages and scripts on social media. Researchers study transfer learning techniques, cross-lingual embeddings, and evaluation benchmarks for non-English languages.

14 papers

Contextual Offensive Language Detection

This sub-topic focuses on incorporating conversational context, sarcasm, and pragmatics to distinguish offensive intent from benign usage in social media text. Researchers investigate transformer-based models and discourse analysis for improved precision.

15 papers

Cyberbullying Detection in Conversations

This sub-topic examines sequence modeling and graph-based methods to detect bullying patterns across threaded conversations and user interactions on platforms like Twitter. Researchers analyze temporal dynamics and victim-perpetrator relationships.

11 papers

Targeted Hate Speech Classification

This sub-topic involves categorizing hate speech by targeted groups such as race, gender, or religion using fine-grained taxonomies and entity recognition. Researchers develop datasets and hierarchical classifiers for nuanced moderation.

15 papers

Explainable Hate Speech Detection

This sub-topic explores interpretable AI techniques like attention mechanisms and counterfactuals to provide rationales for hate speech predictions. Researchers focus on human-AI alignment and regulatory compliance in content moderation.

10 papers

Why It Matters

Automated detection supports moderation on platforms like Twitter by flagging hate speech and cyberbullying, reducing online harassment exposure for users. Davidson et al. (2017) in "Automated Hate Speech Detection and the Problem of Offensive Language" analyzed Twitter data, showing lexical methods classify all messages with specific terms as hate speech, achieving low precision and requiring advanced classifiers for better separation of hate from mere offense. Tokunaga (2010) in "Following you home from school: A critical review and synthesis of research on cyberbullying victimization" synthesized studies revealing cyberbullying's persistent effects, underscoring detection's role in promoting online safety. These methods enable scalable content filtering, as seen in efforts to categorize abusive posts amid rising social media use.

Reading Guide

Where to Start

"Automated Hate Speech Detection and the Problem of Offensive Language" by Davidson et al. (2017), as it directly addresses core technical challenges in distinguishing hate speech from offense using Twitter data, providing an accessible entry to methods and limitations.

Key Papers Explained

Davidson et al. (2017) in "Automated Hate Speech Detection and the Problem of Offensive Language" establishes detection challenges with empirical Twitter analysis, cited 2347 times. Tokunaga (2010) in "Following you home from school: A critical review and synthesis of research on cyberbullying victimization" (2404 citations) builds context by reviewing victimization effects, informing why detection matters. Butler (1997) in "Excitable Speech: A Politics of the Performative" (6341 citations) offers theoretical grounding on speech as conduct, connecting to automated regulation debates in later works.

Paper Timeline

100%
graph LR P0["Methods of coping with social de...
1985 · 2.2K cites"] P1["Excitable Speech: A Politics of ...
1997 · 6.3K cites"] P2["E-Moderating
2004 · 2.2K cites"] P3["Following you home from school: ...
2010 · 2.4K cites"] P4["Misinformation and Its Correction
2012 · 2.7K cites"] P5["Automated Hate Speech Detection ...
2017 · 2.3K cites"] P6["Defining “Fake News”
2017 · 2.0K cites"] P0 --> P1 P1 --> P2 P2 --> P3 P3 --> P4 P4 --> P5 P5 --> P6 style P1 fill:#DC5238,stroke:#c4452e,stroke-width:2px
Scroll to zoom • Drag to pan

Most-cited paper highlighted in red. Papers ordered chronologically.

Advanced Directions

Current frontiers emphasize supervised learning refinements for contextual accuracy, as per Davidson et al. (2017), amid 40,950 works. No recent preprints or news in the last 12 months indicate steady maturation without major shifts. Focus persists on deep learning for Twitter-scale data.

Papers at a Glance

# Paper Year Venue Citations Open Access
1 Excitable Speech: A Politics of the Performative 1997 6.3K
2 Misinformation and Its Correction 2012 Gothic.net 2.7K
3 Following you home from school: A critical review and synthesi... 2010 Computers in Human Beh... 2.4K
4 Automated Hate Speech Detection and the Problem of Offensive L... 2017 Proceedings of the Int... 2.3K
5 E-Moderating 2004 2.2K
6 Methods of coping with social desirability bias: A review 1985 European Journal of So... 2.2K
7 Defining “Fake News” 2017 Digital Journalism 2.0K
8 Cognitive consequences of forced compliance. 1959 PubMed 1.9K
9 Tracking Epistemic Violence, Tracking Practices of Silencing 2011 Hypatia 1.8K
10 Fake news on Twitter during the 2016 U.S. presidential election 2019 Science 1.7K

Frequently Asked Questions

What is the main challenge in automated hate speech detection?

A key challenge is separating hate speech from other offensive language, as lexical methods have low precision by classifying all messages with certain terms as hate speech. Davidson et al. (2017) in "Automated Hate Speech Detection and the Problem of Offensive Language" demonstrated this issue using Twitter data and previous supervised learning approaches. Advanced classifiers improve accuracy by learning contextual distinctions.

How does cyberbullying victimization persist according to research?

Cyberbullying victimization follows individuals beyond immediate encounters, with effects lingering in daily life. Tokunaga (2010) in "Following you home from school: A critical review and synthesis of research on cyberbullying victimization" reviewed studies showing its critical synthesis across contexts. Detection systems aim to interrupt these patterns on platforms like Twitter.

What techniques are used for hate speech detection on social media?

Techniques include machine learning, natural language processing, and deep learning to identify abusive content on Twitter. The field totals 40,950 works emphasizing these methods for categorization. Davidson et al. (2017) showed supervised learning outperforms lexical detection in precision.

Why do lexical methods fail in offensive language detection?

Lexical methods fail due to overgeneralization, labeling all offensive terms as hate speech without context. Davidson et al. (2017) reported low precision in such approaches on social media data. Contextual models using NLP address this limitation.

What is the scale of research in hate speech and cyberbullying detection?

Research comprises 40,950 works centered on automated detection in social media. Growth data over five years is unavailable, but citation leaders like Davidson et al. (2017) with 2347 citations highlight active focus. Topics include machine learning and NLP applications.

Open Research Questions

  • ? How can detection models better distinguish hate speech from benign offensive language in diverse social media contexts?
  • ? What are the long-term psychological impacts of undetected cyberbullying and how do they inform detection thresholds?
  • ? Which contextual features from NLP improve precision in low-resource languages for hate speech detection?
  • ? How do performative aspects of hate speech, as theorized theoretically, translate to automated classification challenges?

Research Hate Speech and Cyberbullying Detection with AI

PapersFlow provides specialized AI tools for Computer Science researchers. Here are the most relevant for this topic:

See how researchers in Computer Science & AI use PapersFlow

Field-specific workflows, example queries, and use cases.

Computer Science & AI Guide

Start Researching Hate Speech and Cyberbullying Detection with AI

Search 474M+ papers, run AI-powered literature reviews, and write with integrated citations — all in one workspace.

See how PapersFlow works for Computer Science researchers