How Google’s “Near-Accurate” AI Overviews Fuel Misinformation
Researchers argue that such systems are “good enough,” especially given recent improvements between models like Gemini 2 and Gemini 3.
Topics
News
- How Google’s “Near-Accurate” AI Overviews Fuel Misinformation
- Automation Foundation Key to AI Success, Says Report
- Anthropic Keeps Its Latest AI Mythos Behind Closed Doors
- Oman Advances Smart Grid Ambitions with Digital Twin AI Initiative
- Intel Comes On Board for Musk’s Terafab AI Chip Project
- Big Tech Data Centers Confront Water, Energy, and Accountability Pressures
Google’s introduction of AI Overviews is changing how information is consumed online—pointing at what accuracy, scale, and user trust mean on the internet today.
An analysis requested by The New York Times of startup Oumi found that these summaries are correct roughly 90% of the time. At face value, that figure appears not bad at all, leading to large language models. But when applied to our dependence on Google to know things—more than five trillion searches annually—even a modest error rate translates into a colossal avalanche of misinformation: tens of millions of flawed responses per hour, burying the right information beneath.
One cannot deny that these AI systems are improving at what they are designed to do, but these “near-accurate” systems should not continue to be accepted at huge scales.
Assessing accuracy remains difficult. Oumi’s study found that over half of the responses deemed correct linked to sources that did not fully substantiate the claims made. This suggests a misleading gap in the information, appearing verified but lacking substance.
Some researchers argue that such systems are already “good enough,” especially given recent improvements between models like Gemini 2 and Gemini 3, where measured accuracy rose from 85% to 91% on the SimpleQA benchmark. Others suggest that user behavior should be prioritized over model performance, as these systems are being further developed by their owner companies.
The range of evidence and reportage around actions motivated by AI chatbots has been concerning. People rarely verify AI-generated responses: one report found only 8% consistently cross-check outputs, while another showed that users followed incorrect AI guidance nearly 80% of the time.
The researchers are calling it “cognitive surrender,” which in layman’s terms translates to mental laziness. In this context, even small error rates have bigger consequences.
Like other forces in the AI industry, Google acknowledges the limitation. It displays a disclaimer: “AI can make mistakes.” Internal testing of Gemini 3 reportedly found error rates as high as 28% in certain contexts.
Pratik Verma, CEO of Okahu says, “Never trust one source. Always compare what you get with another source.”


