How Google’s “Near-Accurate” AI Overviews Fuel Misinformation

Researchers argue that such systems are “good enough,” especially given recent improvements between models like Gemini 2 and Gemini 3.

Topics

  • Google’s introduction of AI Overviews is changing how information is consumed online—pointing at what accuracy, scale, and user trust mean on the internet today.

    An analysis requested by The New York Times of startup Oumi found that these summaries are correct roughly 90% of the time. At face value, that figure appears not bad at all, leading to large language models. But when applied to our dependence on Google to know things—more than five trillion searches annually—even a modest error rate translates into a colossal avalanche of misinformation: tens of millions of flawed responses per hour, burying the right information beneath.

    One cannot deny that these AI systems are improving at what they are designed to do, but these “near-accurate” systems should not continue to be accepted at huge scales.

    Assessing accuracy remains difficult. Oumi’s study found that over half of the responses deemed correct linked to sources that did not fully substantiate the claims made. This suggests a misleading gap in the information, appearing verified but lacking substance. 

    Some researchers argue that such systems are already “good enough,” especially given recent improvements between models like Gemini 2 and Gemini 3, where measured accuracy rose from 85% to 91% on the SimpleQA benchmark. Others suggest that user behavior should be prioritized over model performance, as these systems are being further developed by their owner companies. 

    The range of evidence and reportage around actions motivated by AI chatbots has been concerning. People rarely verify AI-generated responses: one report found only 8% consistently cross-check outputs, while another showed that users followed incorrect AI guidance nearly 80% of the time. 

    The researchers are calling it “cognitive surrender,” which in layman’s terms translates to mental laziness. In this context, even small error rates have bigger consequences.

    Like other forces in the AI industry, Google acknowledges the limitation. It displays a disclaimer: “AI can make mistakes.” Internal testing of Gemini 3 reportedly found error rates as high as 28% in certain contexts.

    Pratik Verma, CEO of Okahu says, “Never trust one source. Always compare what you get with another source.” 

    Topics

    More Like This

    You must to post a comment.

    First time here? : Comment on articles and get access to many more articles.