In order to prevent toxic speech online, the present study investigates the possibilities of artificial intelligence (AI), specifically large language models (LLMs). The study focuses on hate speech classification and counterspeech generation, highlighting advancements in supervised, unsupervised, and generative approaches while recognizing important limitations, such as the amplification of biases from training data, implicit hate, and difficulties in capturing nuance. Reviewing efforts to create AI-powered counterspeech tools, the focus is on the challenges of creating human-like, constructive responses to offensive content. In order to ensure ethical and successful deployment in addressing online harms, the research finds that LLMs hold promise for counterspeech applications and provides guidelines for developers and policymakers. https://toda.org/policy-briefs-and-resources/policy-briefs/report-253-full-text.html Share this: Click to print (Opens in new window) Print Click to share on Facebook (Opens in new window) Facebook Click to share on LinkedIn (Opens in new window) LinkedIn Click to share on Reddit (Opens in new window) Reddit Click to share on WhatsApp (Opens in new window) WhatsApp Click to share on Bluesky (Opens in new window) Bluesky Click to email a link to a friend (Opens in new window) Email Like this:Like Loading... Post navigation Performance comparison of deep learning approaches for Indonesian twitter hate speech detection using IndoBERTweet embedding (Procedia Computer Science) Beyond Laws: Regulating Online Hate Through Collective Action (Tackling Hate)