Google DeepMind CEO Warns of AI’s Critical Inconsistency Problem
Google DeepMind CEO Demis Hassabis has issued a critical warning about artificial intelligence’s fundamental flaw: inconsistent performance. He revealed that even the most advanced AI systems can excel at elite mathematical competitions while failing basic elementary school problems.
Key Takeaways
- AI systems show brilliant performance in complex tasks but fail at simple problems
- This “jagged intelligence” threatens the timeline for achieving AGI
- Major tech leaders acknowledge significant hurdles remain before true AGI
The Jagged Intelligence Problem
Hassabis described current AI as having “uneven intelligences” or “jagged intelligences”—systems that excel brilliantly in some areas while being easily exposed in others. This characterization aligns with Google CEO Sundar Pichai’s term “AJI” (artificial jagged intelligence) coined earlier this year.
“It shouldn’t be that easy for the average person to just find a trivial flaw in the system,” Hassabis stated. He highlighted how Google’s Gemini models enhanced with DeepThink can win gold medals at the International Mathematical Olympiad but “still make simple mistakes in high school maths.”
Beyond Scaling: The Real Challenge
The DeepMind chief emphasized that solving this inconsistency requires more than just increasing data and computing power. “Some missing capabilities in reasoning and planning in memory” still need to be cracked, he explained. Hassabis called for better testing methodologies and “new, harder benchmarks” to precisely map AI strengths and weaknesses.
AGI Timeline Faces Reality Check
Despite predicting AGI’s arrival “in the next five to 10 years” in April, Hassabis now acknowledges significant hurdles remain. His concerns align with OpenAI CEO Sam Altman’s recent assessment following GPT-5’s launch, where Altman admitted the model lacks continuous learning capabilities—something he considers essential for true AGI.
The warnings highlight growing recognition among AI leaders that current systems’ propensity for hallucinations, misinformation, and basic errors must be addressed before achieving human-level reasoning. This serves as a cautionary tale reminiscent of social media platforms’ early failures to anticipate consequences at scale.



