(Gemini Audio Overview)

(preventhate.org) – Multiple research efforts are focused on improving the detection and regulation of hate speech across diverse linguistic and legal contexts, according to new content added to preventhate.org. One approach, called SMARTER, uses a two-stage, data-efficient framework with Large Language Models (LLMs) that generate synthetic explanations to achieve up to a 13.5% macro-F1 improvement in toxicity detection over few-shot baselines.

Another study addresses bias in automatic detection by proposing a weakly supervised framework that combines prompt-based learning and contrastive strategies with limited expert annotations to improve sensitivity to sociolinguistic subtleties. In the French language, a new dataset was assembled to evaluate models, with DistilCamemBert achieving the highest F1-score of 80% for binary hate speech classification. Furthermore, a new system for Roman Urdu expanded an existing dataset and used techniques like mBERT, which reached 92% accuracy, to effectively identify abusive and racist language patterns.

Finally, comparative legal research examined hate speech regulations in five Global South countries—South Africa, Argentina, Colombia, India, and Mexico—to suggest more comprehensive and successful legal measures for unequal societies.

By author

Leave a Reply

Your email address will not be published. Required fields are marked *