Cyberbullying is becoming more common in the digital sphere and is a serious threat to mental health. People, across the world, are subjected to a variety of cyberbullying tactics as the use of social media and technology increases, including insults, false rumors, and online abuse. The severity of cyberbullying cannot be understated, despite its widespread occurrence, as it can result in depressing feelings and even suicidal thoughts. As people depend more and more on social media, cyberbullying has emerged as a major drawback. Victims may go through mental distress and frequently find it difficult to stop textual harassment. Intelligent systems must be put in place to address these problems on social media.
Introduction
I. INTRODUCTION
Cyberbullying has become a major concern in the current digital era, especially on social media platforms that act as the main distribution channels for this type of behavior. The sheer volume of user-generated content is too much for traditional manual moderation techniques to handle. By automating detection procedures and significantly increasing efficiency, machine learning offers a promising remedy.
Social networking sites are great resources for making connections with people, but excessive use of them has also made it easier for people to engage in unethical and illegal behavior in online communities. Notably, people are coming up with new ways to bully others online, particularly teenagers and young adults. Nearly 25% of parents say their child has experienced cyberbullying, according to Symantec.
Cyberbullying is so widespread that it can happen at any time and use the internet to reach anyone, anywhere. Cyberbullying can take many forms, including text, photo, or video, and can be posted covertly, making it difficult or even impossible to identify the original source.
Apps for online socializing provide a forum for sharing personal information and voicing ideas and opinions. Popular online media platforms that are accessible from a variety of devices, including phones, laptops, PCs, and tablets, are Facebook, Twitter, Instagram, and TikTok, to name a few. These days, social media serves noble purposes and boosts the global economy by generating many job opportunities, in addition to being essential in several fields such as coaching and entrepreneurship.
II. METHODS AND MATERIAL
In this study, we utilized Bidirectional Long Short-Term Memory (BiLSTM) for cyberbullying detection. BiLSTM, a type of recurrent neural network (RNN) architecture, was chosen for its ability to capture long-term dependencies in sequential data. It comprises two LSTM layers, processing input sequences in both forward and reverse order for a comprehensive understanding of the data.
Our dataset consisted of labelled instances of cyberbullying content. We prepared the data for BiLSTM input using preprocessing techniques such as tokenization, normalization, and feature extraction.
The BiLSTM model was configured with appropriate hyperparameters, determined through experimentation and validation on a separate development dataset. We trained the model using a backpropagation algorithm with stochastic gradient descent (SGD) optimization to minimize the loss function.
Performance evaluation of the BiLSTM model involved calculating standard metrics such as accuracy, precision, recall, and F1-score on a held-out test set. We also assessed the model's generalization ability to unseen data and its computational efficiency through cross-validation and runtime analysis, respectively.
Furthermore, we explored the interpretability of the BiLSTM model's predictions using techniques like attention mechanisms and feature importance analysis to understand the linguistic cues and patterns driving cyberbullying detection.
Overall, our study encompassed data preparation, model configuration, training, evaluation, and interpretation to effectively detect cyberbullying content in online environments using BiLSTM.
III. RESULTS AND DISCUSSION
The results of the experiment demonstrate that Bidirectional Long Short-Term Memory (BiLSTM) outperforms other classifiers in terms of accuracy for cyberbullying detection. Among the classifiers evaluated, BiLSTM achieved the highest accuracy of 81.31%. In comparison, traditional classifiers such as Bagging Classifier, Random Forest Classifier, and AdaBoost Classifier yielded accuracies ranging from 66.82% to 69.03%.
The superior performance of BiLSTM can be attributed to its ability to capture long-term dependencies and subtle linguistic cues in sequential data, which are crucial for accurately detecting cyberbullying content. Unlike traditional classifiers, which may struggle to effectively model the complex relationships present in text-based data, BiLSTM's bidirectional processing capability enables it to discern patterns in both forward and reverse sequences, leading to improved classification accuracy.
Conclusion
In conclusion, the adoption of Bidirectional Long Short-Term Memory (BiLSTM) technology for cyberbullying detection represents a significant advancement in creating safer online environments. BiLSTM\'s bidirectional processing capability enables a nuanced understanding of sequential data, facilitating the identification of subtle linguistic cues associated with cyberbullying. Its benefits, including improved accuracy, recognition of long-range dependencies, and flexibility in learning patterns, position BiLSTM as a valuable tool in the ongoing fight against online harassment.
References
[1] D. M A and D. K. Daniel, \"Cyberbullying Detection on Social Networks using LSTM Model,\" 2022 International Conference on Innovations in Science and Technology for Sustainable Development (ICISTSD), Kollam, India, 2022, pp. 293-296, doi: 10.1109/ICISTSD55159.2022.10010559.
[2] M. Gada, K. Damania and S. Sankhe, \"Cyberbullying Detection using LSTM-CNN architecture and its applications,\" 2021 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 2021, pp. 1-6, doi: 10.1109/ICCCI50826.2021.9402412.
[3] Alam, Kazi & Bhowmik, Shovan & Prosun, Priyo. (2021). Cyberbullying Detection: An Ensemble Based Machine Learning Approach. 710-715. 10.1109/ICICV50876.2021.9388499.
[4] Luo, Y., Zhang, X., Hua, J., Shen, W.: Multifeatured cyberbullying detection based on deep learning. In: 2021 16th International Conference on Computer Science & Education (ICCSE), pp. 746–751 (2021). IEEE
[5] Ahmed, Md.Tofael & Rahman, Maqsudur & Nur, Shafayet & Islam, Azm & Das, Dipankar. (2021). Deployment of Machine Learning and Deep Learning Algorithms in Detecting Cyberbullying in Bangla and Romanized Bangla text: A Comparative Study. 1-10. 10.1109/ICAECT49130.2021.9392608.
[6] Alam, K.S., Bhowmik, S., Prosun, P.R.K.: Cyberbullying detection: an ensemble-based machine learning approach. In: 2021 Third International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV), pp. 710–715 (2021)
[7] J. Yadav, D. Kumar and D. Chauhan, Cyberbullying Detection using Pre-Trained BERT Model, ICESC, pp. 1096-1100, doi: 10.1109/ICESC48915.2020.9155700. (2020)
[8] R. R. Dalvi, S. Baliram Chavan and A. Halbe, Detecting A Twitter Cyberbullying Using Machine Learning, ICICCS, pp. 297-301, doi: 10.1109/ICICCS48265.2020.9120893. (2020)
[9] Trana R.E., Gomez C.E., Adler R.F. (2021) Fighting Cyberbullying: An Analysis of Algorithms Used to Detect Harassing Text Found on YouTube. In: Ahram T. (eds) Advances in Artificial Intelligence, Software and Systems Engineering. AHFE 2020. Advances in Intelligent Systems and Computing, vol 1213. Springer, Cham. https://doi.org/10.1007/978-3-030-51328-3_2.(2020)
[10] Yadav, Y., Bajaj, P., Gupta, R.K., Sinha, R.: A comparative study of deep learning methods for hate speech and offensive language detection in textual data. In: 2021 IEEE 18th India Council International Conference (INDICON), pp. 1–6 (2021)
[11] Sahana, B., Sandhya, G., Tanuja, R., Ellur, S., Ajina, A.: Towards a safer conversation space: Detection of toxic content in social media (student consortium). In: 2020 IEEE Sixth International Conference on Multimedia Big Data (BigMM), pp. 297–301 (2020)
[12] Aind, A.T., Ramnaney, A., Sethia, D.: Q-bully: a reinforcement learning based cyberbullying detection framework. In: 2020 International Conference for Emerging Technology (INCET), pp. 1–6 (2020)
[13] Ketsbaia, L., Issac, B., Chen, X.: Detection of hate tweets using machine learning and deep learning. In: 2020 IEEE 19th International Conference on Trust, Security and Privacy in Computing and Communications (TrustCom), pp. 751–758 (2020).