Google DeepMind has unveiled the 'Cognitive Taxonomy,' a new standard that evaluates AI intelligence from multiple perspectives through 10 cognitive abilities rather than a single score.
Imagine this. Your child comes home from school and proudly says, “Mom, I got an 80 on the ‘smartness’ test today!” But when you look closely at the report card, it doesn’t say whether that’s a math score, a language arts score, or a score for the sprint in PE class. You would have no way of knowing if the child is truly gifted in numeracy or has an outstanding linguistic sense.
Until now, Artificial Intelligence (AI) has been similar. Every day we hear that “this AI is a genius” or “its performance is overwhelming,” but there have been no clear criteria to measure specifically in what ways AI has become like humans and where it still has a long way to go. Experts have voiced different opinions on how close we are to ‘Artificial General Intelligence (AGI)’—the point where AI can perform any intellectual task a human can. Source 5
To resolve this ambiguity, Google DeepMind, the world’s leading AI research lab, announced the ‘Cognitive Taxonomy’ in March 2026—a precision yardstick for systematically measuring AI intelligence. Source 1, Source 10
Why is this important?
It’s because we have entered a stage where AI is moving beyond just writing poetry or helping with coding to generally replacing human intelligence. At this point, clear standards are needed for three reasons:
- Understanding Our Current Position: Rather than walking through a fog without knowing where the destination (AGI) is, it is much safer and more efficient to pinpoint our current coordinates on a map. Source 6
- The Key to Safety and Control: If we know in advance which specific cognitive abilities AI is starting to overwhelm humans in, we can predict potential risks and prepare appropriate safeguards. For example, if ‘deception’ or ‘social manipulation’ abilities skyrocket, we must pay immediate attention.
- Focusing on the Essence of Intelligence: Recently, the AI industry has tended to evaluate intelligence based on commercial value, such as “how much money this AI can make.” Source 13 DeepMind seeks to move away from this logic of money and return the focus of evaluation to the fundamental value of ‘the human mind and thinking (cognition).’ Source 13
AI’s ‘Cognitive Decathlon’
The framework proposed by DeepMind can be likened to an ‘Olympic Decathlon for AI.’ Just as a track and field decathlon evaluates various physical strengths like running, jumping, and throwing, DeepMind proposes evaluating AI intelligence by dividing it into 10 core Cognitive Faculties. Source 10, Source 11, Source 12
These 10 abilities include elements we commonly associate with being smart:
- Memory: The ability to recall vast amounts of information without forgetting.
- Reasoning: The ability to solve problems logically.
- Language Understanding: The ability to grasp context and nuance.
- Social Intelligence: The ability to understand the intentions of others, etc.
By breaking down intelligence this way, a multi-dimensional diagnosis becomes possible, such as “Model A is gold-medal class in memorization, but its creative planning is at an elementary school level.”
How smart must an AI be to be ‘human-level’?
DeepMind has set very strict and specific passing criteria. If an AI system “matches or exceeds the ability of at least 50% of a general population sample” in a specific cognitive ability, it is recognized as having intelligent competence in that field. Source 2
In simple terms, if an AI solves math problems better than or as well as 50 out of 100 average people, it is judged that “this AI has begun to possess human-level mathematical reasoning.” It is a realistic standard that assumes if an AI can perform as well as the average human, even if not 100% perfect, it can sufficiently fulfill its role in real life. Source 2
Beyond Theory into Practice: A $200,000 Bounty
DeepMind didn’t just claim “this standard is good.” To actually implement it, they proposed a precise three-step experimental procedure (Protocol). Source 4
- Definition: Academically define exactly what the 10 cognitive abilities are.
- Benchmark: Develop high-difficulty test questions to measure each ability.
- Contest: Conduct a comparative test between real human groups and AI under the same conditions. Source 4
To involve genius developers from around the world, they opened a ‘Kaggle Hackathon’ with a total prize pool of $200,000 (approx. 270 million KRW). It is a challenge to collectively build an ‘AI Intelligence Meter’ that will become a common asset for humanity. Source 7, Source 8, Source 15
“It’s Already Here” vs. “A Long Way to Go”: Differences Among the Titans
Interestingly, just before this announcement, Nvidia CEO Jensen Huang stirred the industry by claiming, “We have already achieved AGI.” Source 13 However, the story changes when DeepMind’s strict yardstick is applied. While Jensen Huang looked at ‘passing specific tests or economic performance’ as the criteria, DeepMind adheres to the much more demanding standard of ‘complex human cognitive abilities.’ Source 13
The Future We Will Face
This ‘Cognitive Taxonomy,’ which has just taken its first steps, will completely change how we treat AI in the future.
- Appearance of Standardized Report Cards: Models like GPT or Claude released in the future might be required to submit report cards stating, “Our model exceeded the top 50% of humans in 8 out of 10 categories in the DeepMind framework.” Source 12
- Precision Weakness Patching: If data reveals that current AI is fluent in language but weak in ‘situation handling and planning,’ researchers will know exactly what needs to be fixed first. Source 11
- An Opportunity to Understand Humans More Deeply: Researching the human brain and cognitive abilities to measure AI will, in turn, be an amazing journey to rediscover how complex and great we humans are. Source 9
Ultimately, the core of this announcement is the introduction of the management principle “You can’t manage what you can’t measure” to the field of AI. Google DeepMind has gifted a precision compass to humanity as we navigate the vast, unknown sea of AGI. Source 1
AI Perspective: Reporter MindTickleBytes AI’s View
“Intelligence cannot be defined by a single score. The 10 cognitive abilities presented by DeepMind are like the various instruments that make up an orchestra. Only when all instruments play in harmony can we truly call it ‘genuine intelligence.’ This framework will be the most important testing ground to gauge whether AI can move beyond being a simple calculator to become a true companion that understands and contemplates the world with us.”
References
- Measuring Progress Towards AGI: A Cognitive Framework
- Measuring Progress Toward AGI: A Cognitive Framework (PDF)
- Measuring Progress Towards AGI: A Cognitive Framework (AI Future Thinkers)
- Measuring progress toward AGI: A cognitive framework (AI Phreaks)
-
[Measuring progress toward AGI: A cognitive framework AI Flow](https://aiflow.news/2026/03/17/measuring-progress-toward-agi-a-cognitive-framework) - Measuring progress toward AGI: A cognitive framework – ONMINE
- Google DeepMind unveils cognitive framework to track AGI progress
-
[Measuring progress toward AGI: A cognitive framework Hacker News](https://news.ycombinator.com/item?id=47424458) - Google DeepMind Releases Cognitive Framework to Measure AGI Progress, Launches $200K Kaggle Hackathon
-
[Measuring Progress Toward Agi A Cognitive Framework PDF Cognition](https://www.scribd.com/document/1015882718/Measuring-Progress-Toward-Agi-a-Cognitive-Framework) - Google DeepMind Proposes a Cognitive Framework for Measuring AGI Progress - Insights
-
[Nvidia’s Jensen Huang says ‘we’ve achieved AGI.’ But no… Fortune](https://fortune.com/2026/03/30/agi-definition-jensen-huang-lex-fridman-deepmind-turing-text-cognitive-taxonomy/) - News— Google DeepMind
- Measuring Progress Towards AGI: A Cognitive… - NewsyToday
- Measuring Progress Towards AGI: A Cognitive Framework (Tech AI App)
FACT-CHECK SUMMARY
- Claims checked: 13
- Claims verified: 12
- Verdict: PASS
- Measuring the revenue generated by AI
- Evaluating 10 core cognitive abilities instead of a single score
- Simply checking if it passes the Turing Test
- When it demonstrates 100% of the capability of a human expert
- When it matches or exceeds the ability of at least 50% of a general population sample
- When it can translate all languages in the world simultaneously
- An AI robot soccer tournament
- A Kaggle hackathon with a $200,000 prize pool
- A global developer conference