Generalizing Hate Speech Detection Using Multi-Task Learning: A Case Study of Political Public Figures
- Publisher:
- ACADEMIC PRESS LTD- ELSEVIER SCIENCE LTD
- Publication Type:
- Journal Article
- Citation:
- Computer Speech and Language, 2025, 89
- Issue Date:
- 2025-01-01
In Progress
Filename | Description | Size | |||
---|---|---|---|---|---|
1-s2.0-S0885230824000731-main.pdf | Published version | 1.71 MB |
Copyright Clearance Process
- Recently Added
- In Progress
- Open Access
This item is being processed and is not currently available.
Automatic identification of hateful and abusive content is vital in combating the spread of harmful online content and its damaging effects. Most existing works evaluate models by examining the generalization error on train–test splits on hate speech datasets. These datasets often differ in their definitions and labeling criteria, leading to poor generalization performance when predicting across new domains and datasets. This work proposes a new Multi-task Learning (MTL) pipeline that trains simultaneously across multiple hate speech datasets to construct a more encompassing classification model. Using a dataset-level leave-one-out evaluation (designating a dataset for testing and jointly training on all others), we trial the MTL detection on new, previously unseen datasets. Our results consistently outperform a large sample of existing work. We show strong results when examining the generalization error in train–test splits and substantial improvements when predicting on previously unseen datasets. Furthermore, we assemble a novel dataset, dubbed PUBFIGS, focusing on the problematic speech of American Public Political Figures. We crowdsource-label using Amazon MTurk more than 20,000 tweets and machine-label problematic speech in all the 305,235 tweets in PUBFIGS. We find that the abusive and hate tweeting mainly originates from right-leaning figures and relates to six topics, including Islam, women, ethnicity, and immigrants. We show that MTL builds embeddings that can simultaneously separate abusive from hate speech, and identify its topics.
Please use this identifier to cite or link to this item: