A major issue with social media is online hatred, which has a detrimental impact on people, social media in general, and intolerance in society. To combat it, content moderation relies on dichotomous classifiers that only provide a score based on how likely a communication is to be hostile. However, the limitations of categorical classifiers have hindered social research investigating the factors that influence hatefulness degrees. Using ChatGPT (GPT-4), this study achieved two goals: (1) creating a continuous, 6-interval scale for rating hate communications that is scalable to huge datasets, and (2) comparing GPT-4 to trustworthy human raters on a limited but varied set of hate messages. Although GPT-4’s assessments were much more hostile than those of human raters, the results demonstrate parallelism in the analyses as well as convergent and discriminant validity. The GPT-4 and human raters’ ratings were more strongly associated than those obtained by other classifiers, according to further studies that compare the scores of the GPT-administered measure and human raters to those of other well-known hate classifiers. Limitations and implications for future study are discussed in the work’s conclusion. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5215167 Share this: Click to print (Opens in new window) Print Click to share on Facebook (Opens in new window) Facebook Click to share on LinkedIn (Opens in new window) LinkedIn Click to share on Reddit (Opens in new window) Reddit Click to share on WhatsApp (Opens in new window) WhatsApp Click to share on Bluesky (Opens in new window) Bluesky Click to email a link to a friend (Opens in new window) Email Like this:Like Loading... Post navigation NLP in the Digital Age: Combating Fake News, Hate Speech, and Ethical Risks for Social Integrity (SSRN) Two Weeks in Soft Security: Free Resources on Countering Extremism, Hate, and Disinformation, April 2025 (I/II)