By normalizing hate speech and maintaining a balance between responsible communication and free expression, it is essential to accurately gauge the severity of hate speech in order to moderate and regulate large online groups. The method at hand creates the new Multimodal Model for Hate Speech (MMHS) by utilizing knowledge from both the linguistic and visual domains. By achieving 0.350 lower Root Mean Squared Error (RMSE) and 0.132 and 0.012 higher Pearson and Cosine metrics, respectively, MMHS outperforms prior benchmarks and demonstrates state-of-the-art performance on the NACL dataset. Furthermore, user preference polls show that our forecasts are significantly preferred by 16.67% above those of Masud et al. By improving knowledge of online conversation and advancing the technological landscape of hate speech identification, the study makes it possible to implement more effective moderation techniques. https://link.springer.com/chapter/10.1007/978-3-031-78014-1_8 Share this: Click to print (Opens in new window) Print Click to share on Facebook (Opens in new window) Facebook Click to share on LinkedIn (Opens in new window) LinkedIn Click to share on Reddit (Opens in new window) Reddit Click to share on WhatsApp (Opens in new window) WhatsApp Click to share on Bluesky (Opens in new window) Bluesky Click to email a link to a friend (Opens in new window) Email Like this:Like Loading... Post navigation Investigating the Predominance of Large Language Models in Low-Resource Bangla Language over Transformer Models for Hate Speech Detection: A Comparative Analysis (MDPI) Generalizable Multilingual Hate Speech Detection on Low Resource Indian Languages using Fair Selection in Federated Learning (ACL Anthology)