TOXIC COMMENT CLASSIFICATION COMPARISON BETWEEN LSTM, BILSTM, GRU, AND BIGRU

Jonathan Kevin Giustino, Yonathan Purbo Santosa

Abstract


One of the biggest problems in modern internet age is toxicity. In this study, our aim is to draw an effective method to classify toxicity in text comment in this case specifically on Wikipedia comment. Old Regular Rule based model attacks, Machine Learning Methods suffer from rules based approaches and thus are incapable of accurate detection of toxicity while maintaining precision at the same time. To overcome this limitation, recurrent neural networks (RNNs), wherein, long short-term memory (LSTM) networks, gated recurrent unit(GRU) have been proposed. In this paper, author compares the LSTM, BiLSTM, GRU, and BiGRU for multi label classification for which is the best model to use towards jigsaw toxicity challenge dataset to classify toxicity. This study will be finding out which of the model is the best for classification and the difference between different type of pre-processing. We’ll go ahead to use League of Legends tribunal datasets from kaggle as our base. The results obtained were that, the highest accuracy on the test was attained by BiLSTM model without cleaning with on 87.208% accuracy, 55.205% Precision, 68.540% Recall, and 60.623% F1-Score the result also shows while preprocessing on cleaning improve the resultant metrics by a marginal amount for regular LSTM, GRU, and BiGRU it doesn’t always improve the result.


Keywords


Toxicity; LSTM; GRU; BiLSTM; BiGRU

Full Text:

PDF

References


S. Murnion, W. J. Buchanan, A. Smales, and G. Russell, “Machine learning and semantic analysis of in-game chat for cyberbullying,” Computers & Security, vol. 76, pp. 197–213, Jul. 2018, doi: 10.1016/j.cose.2018.02.016.

M. Martens, S. Shen, A. Iosup, and F. Kuipers, “Toxicity detection in multiplayer online games,” in 2015 International Workshop on Network and Systems Support for Games (NetGames), Zagreb: IEEE, Dec. 2015, pp. 1–6. doi: 10.1109/NetGames.2015.7382991.

K. Dubey, R. Nair, Mohd. U. Khan, and Prof. S. Shaikh, “Toxic Comment Detection using LSTM,” in 2020 Third International Conference on Advances in Electronics, Computers and Communications (ICAECC), Bengaluru, India: IEEE, Dec. 2020, pp. 1–8. doi: 10.1109/ICAECC50550.2020.9339521.

M. Anand and R. Eswari, “Classification of Abusive Comments in Social Media using Deep Learning,” in 2019 3rd International Conference on Computing Methodologies and Communication (ICCMC), Erode, India: IEEE, Mar. 2019, pp. 974–977. doi: 10.1109/ICCMC.2019.8819734.

C. Esposito, G. A. Landrum, N. Schneider, N. Stiefl, and S. Riniker, “GHOST: Adjusting the Decision Threshold to Handle Imbalanced Data in Machine Learning,” J. Chem. Inf. Model., vol. 61, no. 6, pp. 2623–2640, Jun. 2021, doi: 10.1021/acs.jcim.1c00160.

M. Ibrahim, M. Torki, and N. El-Makky, “Imbalanced Toxic Comments Classification Using Data Augmentation and Deep Learning,” in 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL: IEEE, Dec. 2018, pp. 875–878. doi: 10.1109/ICMLA.2018.00141.

Z. Mossie and J.-H. Wang, “Vulnerable community identification using hate speech detection on social media,” Information Processing & Management, vol. 57, no. 3, p. 102087, May 2020, doi: 10.1016/j.ipm.2019.102087.

A. Obadimu, E. Mead, M. N. Hussain, and N. Agarwal, “Identifying Toxicity Within YouTube Video Comment,” in Social, Cultural, and Behavioral Modeling, vol. 11549, R. Thomson, H. Bisgin, C. Dancy, and A. Hyder, Eds., in Lecture Notes in Computer Science, vol. 11549. , Cham: Springer International Publishing, 2019, pp. 214–223. doi: 10.1007/978-3-030-21741-9_22.

A. Abbasi, A. R. Javed, F. Iqbal, N. Kryvinska, and Z. Jalil, “Deep learning for religious and continent-based toxic content detection and classification,” Sci Rep, vol. 12, no. 1, p. 17478, Oct. 2022, doi: 10.1038/s41598-022-22523-3.

E. Yazgili and M. Baykara, “Cyberbullying and Detection Methods,” in 2019 1st International Informatics and Software Engineering Conference (UBMYK), Ankara, Turkey: IEEE, Nov. 2019, pp. 1–5. doi: 10.1109/UBMYK48245.2019.8965514.

F. Mohammad, “Is preprocessing of text really worth your time for online comment classification?” arXiv, Aug. 29, 2018. doi: 10.48550/arXiv.1806.02908.

“jigsaw-toxic-comment-classification-challenge.” Accessed: Jan. 05, 2024. [Online]. Available: https://www.kaggle.com/datasets/julian3833/jigsaw-toxic-comment-classification-challenge




DOI: https://doi.org/10.24167/proxies.v7i2.12471

Copyright (c) 2024 Proxies : Jurnal Informatika



View My Stats