Google doubles down on AI Overviews after telling users to eat rocks and put glue on pizza
Google admits to some 'unhelpful' results
Here at Tom’s Guide our expert editors are committed to bringing you the best news, reviews and guides to help you stay informed and ahead of the curve!
You are now subscribed
Your newsletter sign-up was successful
Want to add more newsletters?
Daily (Mon-Sun)
Tom's Guide Daily
Sign up to get the latest updates on all of your favorite content! From cutting-edge tech news and the hottest streaming buzz to unbeatable deals on the best products and in-depth reviews, we’ve got you covered.
Weekly on Thursday
Tom's AI Guide
Be AI savvy with your weekly newsletter summing up all the biggest AI news you need to know. Plus, analysis from our AI editor and tips on how to use the latest AI tools!
Weekly on Friday
Tom's iGuide
Unlock the vast world of Apple news straight to your inbox. With coverage on everything from exciting product launches to essential software updates, this is your go-to source for the latest updates on all the best Apple content.
Weekly on Monday
Tom's Streaming Guide
Our weekly newsletter is expertly crafted to immerse you in the world of streaming. Stay updated on the latest releases and our top recommendations across your favorite streaming platforms.
Join the club
Get full access to premium articles, exclusive features and a growing list of member rewards.
Google has doubled down on its AI-powered search results, saying the recent spate of weird and unhelpful responses are limited to just a handful of niche queries.
AI Overviews were announced as part of Google I/O earlier this month and put longer-form Gemini-powered text responses above results for more complex queries. These are questions that would require several website visits to answer.
Soon after launch a number of issues started to pop up including one suggesting the use of nontoxic glue to thicken up pizza sauce, another saying eating rocks can be good for you and a third suggesting smoking while pregnant is healthy — it is not. If you don't want AI Overviews we have a guide to blocking them from results.
In a statement, Liz Reid, Google's head of search said many of the more outlandish claims on social media couldn't be reproduced — including the rocks question — and of the real queries she said: "there are areas we need to improve."
What went wrong with Google AI Overviews?
Reid says the biggest problem with many of the results came down to how the AI model interpreted sarcasm and humor. "We saw AI Overviews that featured sarcastic or troll-y content from discussion forums," she wrote.
"Forums are often a great source of authentic, first-hand information, but in some cases can lead to less-than-helpful advice, like using glue to get cheese to stick to pizza." Which could be a problem for OpenAI's deal with Reddit.
Then you had the people running with this as a joke, further adding to the sarcastic and troll-y content on discussion forums.
Get instant access to breaking news, the hottest reviews, great deals and helpful tips.
There is also an element of the web stress testing the unexpected behaviors for Google. Reid says: "There’s nothing quite like having millions of people using the feature with many novel searches. We’ve also seen nonsensical new searches, seemingly aimed at producing erroneous results."
What's the solution to the AI Overviews problem?
Generative AI has no way of knowing what’s true, it only knows what’s popular. So, it often ends up surfacing answers from untrustworthy sources or parody accounts instead of the actual facts. In some cases, AI is also prone to “hallucination," where it just makes up false data to cover up its knowledge gap.
Google has always been a platform where users can find all sorts of information and disinformation. But without the ability to correlate the information against the source’s reputation, there’s no way to know if the answers you got are accurate.
Reid says they've already made some improvements to Overviews. "From looking at examples from the past couple of weeks, we were able to determine patterns where we didn’t get it right, and we made more than a dozen technical improvements to our systems," she said.
This includes improving the detection mechanisms for nonsensical queries and blocking AI Overviews from displaying. This includes removing satire and humor content from the source material for those queries.
They also cut back on user-generated content in responses, instead focusing on higher quality sources of material.
Finally, Google says it has implemented stronger guardrails, particularly for news and health content. "We aim to not show AI Overviews for hard news topics, where freshness and factuality are important. In the case of health, we launched additional triggering refinements to enhance our quality protections," said Reid.
More from Tom's Guide
- Google Search is getting 6 huge new Gemini AI upgrades — what you need to know
- Google Veo unveiled — new AI model makes high-def videos to take on Sora
- Gemini 1.5 Pro update is coming with a new app and live conversation features

Ryan Morrison, a stalwart in the realm of tech journalism, possesses a sterling track record that spans over two decades, though he'd much rather let his insightful articles on AI and technology speak for him than engage in this self-aggrandising exercise. As the former AI Editor for Tom's Guide, Ryan wields his vast industry experience with a mix of scepticism and enthusiasm, unpacking the complexities of AI in a way that could almost make you forget about the impending robot takeover.
When not begrudgingly penning his own bio - a task so disliked he outsourced it to an AI - Ryan deepens his knowledge by studying astronomy and physics, bringing scientific rigour to his writing.










