Context: Final Project for CS7643/CS7650 courses at Georgia Tech, Fall 2021
Abstract: Hate Speech remains one of the most pressing issues that come with the rise of social media. While social media has been a blessing in allowing users to connect with same minded people, it can be a breeding ground for hateful ideas and extremism, and when allowed to fester, can lead to physical violence, as shown in many recent events. In our study, we aim to compare different novel algorithms, not traditionally used for hate speech detection but rather used for other natural language tasks, to see if any of them are more effective are recognizing hate speech. By comparing the LSTM, BiLSTM, HybridLSTMCNN, HybridCNNLSTM, PyramidCNN, TinyBert and DistillBert algorithms and comparing their accuracy scores and model sizes, we will see if any of these novel algorithms offer some benefits to the traditional model of hate speech recognition.
Authors:
- Aymane Abdali (CS7643/CS7650)
- Maya Boumerdassi (CS7643)
- Mohamed Amine Dassouli (CS7643/CS7650)
- Jonathan Hou (CS7643)
- Richard Huang (CS7643/CS7650)
Please refer to install.md.
Dataset Paper | Dataset Link1 | Dataset Link2
For this study, we use the sub-task A of the OLID Dataset. This dataset contains English tweets annotated using a three-level annotation hierarchy and was used in the OffensEval challenge in 2019.
Preprocessing functions can be found in preprocess_utils.py.
Please refer to training.md.
All of our best trained models can be found here.
This is the qualitative part of our study. We use Captum to interpret our trained models. Here are some examples from our best trained model DistillBert_2021-12-08_16-39-08_trained_testAcc=0.7960.pth
on the test set:
Please refer to interpret.md for more details.