In this project, we compare the performance of machine learning and deep learning mod- els, including Naive Bayes, Logistic Regres- sion, Decision Tree, LSTM, CNN, and BERT on a toxic comment detection challenge from Kaggle. The main goal of this project is to help mitigate content moderation. Therefore, After training the models, we analyze popular online and social media platforms, including Reddit and Twitter, by detecting toxic content using the trained model. We further show the vulnerability of the models using adversarial examples.
NOTE: if you have trouble loading the scripts, click reload after the first fail.
- script: non_nn_models.ipynb for trainning non neural networks including Naive Bayes, Logistic Regression and Decision Tree
- script: BERT_Toxic.ipynb for trainning BERT
- script: kaggle_benchmark.py for trainning CNN
- script: BERT_CNN.ipynb for trainning BERT-CNN
- script: LSTM.ipynb for traininig LSTM
- script: BERT_tSNE.ipynb for t-SNE visualization
- covid19-tweet-eda.ipynb covid19 related tweets EDA
- script of adversarial attack with Perspective API.
Team Member:
- Danni Chen
- Fuzail Khan
- Don Le