Menu
24matins.uk
Navigation : 
  • News
    • Business
    • Recipe
    • Sport
  • World
  • Health
  • Culture
  • Tech
    • Science
Currently : 
  • Entertainment
  • Tech
  • Health
  • International

Google warns: AI often fabricates information under intense pressure

Tech
By 24matins.uk,  published 16 July 2025 at 14h16, updated on 16 July 2025 at 14h16.
Tech

Google has raised concerns about the reliability of artificial intelligence, revealing that AI systems are prone to frequent inaccuracies and falsehoods, particularly when placed under intense pressure or demanding situations. This warning highlights ongoing challenges in AI trustworthiness.

Tl;dr

  • AI models show human-like doubt when challenged.
  • Confidence collapses after credible opposing arguments.
  • Improved training could bolster AI reliability.
  • AI Models and the Challenge of Doubt

    As the gap between artificial intelligence and human behavior narrows, surprising parallels continue to emerge. A recent study conducted by researchers at Google DeepMind and University College London highlights just how closely advanced language models, such as GPT-4 or Grok 4, mirror our own struggles with confidence and uncertainty—especially when their answers are confronted by credible contradiction.

    The Sudden Collapse of Confidence

    Interestingly, these models start off with an impressive level of assurance. Researchers use a metric called logits, which acts as an internal gauge of certainty for each answer provided. However, the introduction of a seemingly legitimate counterargument is often enough to shake—sometimes even completely undermine—this initial confidence. It becomes even more pronounced if the AI is reminded that its first answer clashes with fresh information or a persuasive critique. This pattern is hardly isolated. Remember that moment when Gemini, under pressure during a Pokémon game, began to falter? Or the case where Claude, developed by Anthropic, faced a sort of identity crisis while managing a virtual shop? These instances reveal that large language models frequently struggle to maintain composure under scrutiny.

    A Closer Look at the Experimentation Process

    To rigorously test this phenomenon, the research team designed a two-step experiment: First, the model was asked a multiple-choice question and its self-reported confidence was logged. Next, it received advice—sometimes accurate, sometimes misleading—from another AI system. The key goal was to observe whether the original response would stand firm or be revised in light of new input. The outcomes were telling:

  • The AI’s confidence in its initial selection often proved excessive.
  • A credible opposing viewpoint routinely triggered hesitation or self-doubt.
  • Error correction wasn’t systematic; indecision was common instead.
  • The Stakes for High-Impact Applications

    While these quirks may seem anecdotal during casual chatbot interactions, their significance deepens in critical decision-making scenarios. The ability for an AI to maintain robust self-assurance amidst external pressure becomes essential—raising serious questions about how much trust we can place in systems whose psychological stability remains imperfect. That said, there is room for optimism. Advances in training methodologies and refinements in prompt engineering offer promising pathways toward more consistent and reliable future models—potentially closing this lingering gap between machine reasoning and unwavering conviction.

    Le Récap
    • Tl;dr
    • AI Models and the Challenge of Doubt
    • The Sudden Collapse of Confidence
    • A Closer Look at the Experimentation Process
    • The Stakes for High-Impact Applications
    • About Us
    © 2026 - All rights reserved on 24matins.uk site content