Open Access Open Access  Restricted Access Subscription Access

AN APPROACH TO SENSITIVE CONTENT MODERATION USING BERT ALGORITHM

Maryala Harshitha, Cheekati Sanjay, Ms.Vedavathi. K, Dr.K. Rajitha, Mr. R. Mohan Krishna, Dr. K. Sreekala

Abstract


Hate speech is an ever-increasing menace among social media and online platforms. This covers harmful and offensive language directed towards an individual or group on the basis of race, gender, religion, or other identities. The alarming spread of hate speech creates toxic environments that have a serious collateral effect on individuals, including mental wellness and online safety. Most platforms have installed automatic systems to detect and remove hate speech, but fitness is often lacking. Traditional machine learning models like LSTM (Long Short-Term Memory) have been in use,


especially in hate speech detection. Although these were good models, they seem to struggle to understand deeper meaning in most of their words and sentences and specially when the given speech features sarcasm or indirect hate. We propose improved approach in our project using the BERT (Bidirectional Encoder Representations from Transformers) model- an state-of-the-art Natural Language Processing model, and unlike LSTM which processes the words in a sequence, BERT reads an entire sentence in one go and understands it both ways, thus making detection of hate speech that much more easier even in the most complex and trickiest of sentences. BERT was trained on


the social media comments dataset where both hate and neutral languages used. Thus with these results, this comparison of BERT to LSTMs shows that hate speech can be identified more accurately with less error using BERT. It can find those more nuanced patterns of hate speech that traditional models usually won't pick up. Achieving online safety is therefore the main aim of this project: installing a system with a more trustworthy detection scheme specific for the detection of hate speech. BERT can help platforms in minimizing harmful content more effectively, creating a more secure digital space for users. This work underlines the essence of adopting modern AI techniques to address real-world issues and improve communication on the web.


Full Text:

PDF

References


Z. Mansur, N. Omar, and S. Tiun, "Twitter hate speech detection: A systematic review of methods, taxonomy analysis, challenges, and opportunities," IEEE Access, vol. 11, pp. 16226–16249, Mar. 2023. [Online]. Available: https://ieeexplore.ieee.org/document/100 25718

J. Lu, H. Lin, X. Zhang, Z. Li, T.

Zhang, L. Zong, F. Ma, and B. Xu, "Hate speech detection via dual contrastive learning," IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 31, pp. 4024–4036, Jul. 2023. [Online].

Available: https://dl.acm.org/doi/abs/10.1109/TAS LP.2023.3294715

S. Nasir, A. Seerat, and M. Wasim, "Hate speech detection in Roman Urdu using machine learning techniques," in Proc. 2024 Int. Conf. Adv. Comput. Sci. (ICACS), Islamabad, Pakistan, 2024, pp. 1–7 Available:

https://doi.org/10.1109/ICACS60934.20 24.10473250bibsonomy.org

C. S. Wang, H. L. Yang, B. Y. Li,

and H. Y. Chen, "Can generative AI eliminate speech harms? A study on detection of abusive and hate speech during the COVID-19 pandemic," in Proc. 2023 IEEE Int. Conf. Consumer Electron.-Asia (ICCE-Asia), Busan, South Korea, Oct. 2023, pp. 1–4. [Online]. Available:

https://doi.org/10.1109/ICCE- Asia59966.2023.10326404National Taipei University of Technology+1ACM Digital Library+1

S. Riyadi, A. D. Andriyani, and S. N. Sulaiman, "Improving hate speech detection using double-layers hybrid CNN-RNN model on imbalanced dataset," IEEE Access, vol. 12, pp. 9660–9671, 2024. [Online]. Available: https://doi.org/10.1109/ACCESS.2024.3 487433ui.adsabs.harvard.edu

K. Guo, A. Hu, J. Mu, Z. Shi, Z.

Zhao, N. Vishwamitra, and H. Hu, "An investigation of large language models for real-world hate speech detection," in Proc. 2023 Int. Conf. Mach. Learn. Appl. (ICMLA), Jacksonville, FL, USA, Dec. 2023, pp. 1568–1573. [Online].

Available: https://doi.org/10.1109/ICMLA58977.2 023.00237NSF PARS

S. Almohaimeed, S. Almohaimeed, and L. Bölöni, "Transfer learning and lexicon-based approaches for implicit hate speech detection: A comparative study of human and GPT-4 annotation," in Proc. 2024 IEEE Int. Conf. Semantic Comput. (ICSC), Laguna Hills, CA, USA, Feb. 2024, pp. 142–147. [Online].

Available: https://doi.org/10.1109/ICSC59802.202 4.00028NSF PARS

F. T. Boishakhi, P. C. Shill, and M.

G. R. Alam, "Multi-modal hate speech detection using machine learning," in Proc. 2023 IEEE Int. Conf. Comput. Sci. Eng. (ICSE), Dhaka, Bangladesh, Jun. 2023, pp. 1–6. [Online]. Available: https://arxiv.org/abs/2307.11519arXiv


Refbacks

  • There are currently no refbacks.