Elon Musk's Grok just ranked worst among AI chatbots in new Anti-Defamation League safety study — here's how it responds to 'antisemitic and extremist content'
A new study ranks Elon Musk’s Grok worst among major AI chatbots at detecting antisemitism
A new Anti-Defamation League (ADL) safety audit has found that Elon Musk’s AI chatbot Grok scored lowest among six leading AI models in identifying and countering antisemitic, anti-Zionist and extremist content — highlighting ongoing gaps in how AI systems manage harmful speech and bias.
The ADL’s AI Index, published this week, evaluated Grok alongside Anthropic’s Claude, OpenAI’s ChatGPT, Google’s Gemini, Meta’s Llama and DeepSeek on more than 25,000 prompts spanning text, images and contextual conversations. The study assessed models’ abilities to recognize and respond appropriately to problematic narratives tied to hate and extremism.
Major findings: Grok trails peers
According to the ADL report, Grok earned just 21 out of 100 points, placing last among the group. In contrast, Anthropic’s Claude led the field with a score of 80, consistently providing context that challenged anti-Jewish and extremist language. ChatGPT, Gemini, Llama and DeepSeek scored in the middle, with gaps in certain formats and categories.
The study highlighted Grok’s weaknesses in maintaining context across multi-turn dialogues and in analyzing images and documents containing harmful content — areas where stronger contextual understanding is required to counter dangerous narratives effectively.
The ADL AI Index provides both "good" and "bad" examples from each of the chatbots for those that want to review them.
Previous controversies around grok
Grok’s performance in the ADL study follows previous controversies tied to the chatbot’s outputs on social media. In July 2025, Grok generated antisemitic content on X that included praise of Adolf Hitler and other offensive language, prompting backlash from the ADL and other advocacy groups. These posts have since been deleted.
At the time, xAI and the chatbot’s official account acknowledged the problem, saying they were working to remove inappropriate posts and make improvements. The ADL called Grok’s behavior “irresponsible, dangerous and antisemitic, plain and simple.”
Get instant access to breaking news, the hottest reviews, great deals and helpful tips.
Elon Musk has previously addressed Grok’s problematic outputs, noting that certain responses were being fixed following those incidents. While those comments weren’t part of the ADL’s recent study, they underscore ongoing challenges in aligning generative AI with robust safety standards.
Industry and regulatory scrutiny
The ADL’s findings come amid broader concern over AI content moderation. Experts say that, without strong safety guardrails and bias mitigation, large language models can inadvertently echo or amplify harmful stereotypes and extremist rhetoric — a risk highlighted by both advocacy groups and regulators.
In addition to safety audit scrutiny, Musk’s AI platforms have faced regulatory attention over other issues tied to harmful outputs. For example, the European Commission recently opened an investigation into Grok’s generation of inappropriate and potentially nonconsensual sexualized images, adding to pressure on developers to address content risks.
Bottom line
With AI tools increasingly integrated into search, social media and productivity workflows, trust and safety remain top concerns for developers and users alike. The ADL’s report highlights that even leading AI models vary widely in how effectively they counter hate speech and harmful narratives — and that ongoing improvements are needed across the industry.
For developers like xAI and its competitors, these findings could influence future model updates and industry expectations around bias mitigation, contextual understanding and content moderation standards.
Follow Tom's Guide on Google News and add us as a preferred source to get our up-to-date news, analysis, and reviews in your feeds.
More from Tom's Guide
- Trusting ChatGPT turned into a 3-day nightmare — here’s why I regret not going to a pro
- Google Maps just got a major AI upgrade — and it’s finally hands-free for walkers and cyclists
- Pokémon reveals what AI still gets wrong — here’s what I found

Amanda Caswell is an award-winning journalist, bestselling YA author, and one of today’s leading voices in AI and technology. A celebrated contributor to various news outlets, her sharp insights and relatable storytelling have earned her a loyal readership. Amanda’s work has been recognized with prestigious honors, including outstanding contribution to media.
Known for her ability to bring clarity to even the most complex topics, Amanda seamlessly blends innovation and creativity, inspiring readers to embrace the power of AI and emerging technologies. As a certified prompt engineer, she continues to push the boundaries of how humans and AI can work together.
Beyond her journalism career, Amanda is a long-distance runner and mom of three. She lives in New Jersey.
You must confirm your public display name before commenting
Please logout and then login again, you will then be prompted to enter your display name.