Exploring Humanity’s Last Exam for AI Intelligence Assessment

By | March 15, 2026

Humanity’s Last Exam: The AI Test That Could Stump Einstein?

Estimated reading time: 6 minutes

  • Ultimate benchmark: HLE features 2,500 expert-level questions.
  • Focus on reasoning: It assesses AI’s critical thinking and problem-solving skills.
  • High-stakes testing: Designed to challenge the best AI models available.
  • Implications for the future: Offers insights into AI capabilities and limitations.
  • Not a free lunch: Critiques highlight its limitations in real-world application.

The Conception of HLE: A Brainchild of Necessity

HLE didn’t just materialize from thin air. It was conceived by the Center for AI Safety and Scale AI, among others, in response to a notable issue: existing tests were as effective as trying to teach a cat to fetch. With the likes of MMLU (Massive Multitask Language Understanding) saturating the field, AI models were cruising through easier benchmarks. HLE was established as a high-stakes benchmark focusing on advanced reasoning rather than the boring old “recall this stuff” game.

The Nature paper titled “A benchmark of expert-level academic questions to assess AI capabilities” lays the groundwork for HLE, with its focus on multi-step reasoning in disciplines like mathematics, natural sciences, humanities, computer science, literature, and history. Basically, it takes the “intelligence” in “artificial intelligence” and gives it a workout.

The Structure of HLE: Questioning Everything (Almost)

Key Features

HLE is composed of a whopping 2,500 public questions, with an additional ~500 holdout questions that remain guarded like celebrity secrets. Here’s the breakdown:

  • Question Types:
    • Approximately 76% of the questions are short answers (which means AI can’t just regurgitate facts like parakeets).
    • About 24% are multiple-choice (because nothing says “you’re trapped” quite like a question with options).
    • Roughly 14% are multimodal, which means they require the brainpower to analyze both text and images.

Difficulty Criteria

The questions aren’t your run-of-the-mill trivia. They are original, possess a single verifiable answer, and are designed to stump those cutting-edge large language models (LLMs). A meticulous filtering process culled ~70,000 questions to a mere 6,000, ultimately resulting in the final public and private sets.

  1. Filtered from 70,000 to around 13,000 through expert peer review.
  2. Shrunk to ~6,000 after manual approval.
  3. Final split: 2,500 public and ~500 private questions.

The results were striking: even cutting-edge AI models stumbled on this exam. GPT-4o managed just 2.7% accuracy, while Claude 3.5 Sonnet scored 4.1%, and OpenAI’s o1 model topped out at roughly 8%. However, newer systems showed dramatic improvement—Gemini 3.1 Pro and Claude Opus 4.6 leaped to 40-50% accuracy, signaling rapid progress in the field.

Why HLE is Crucial: The Benchmark of Intelligence

While many benchmarks put AI’s capabilities on display, HLE takes matters a step further. It doesn’t just throw questions at AI models but assesses their capability to understand and work through complex reasoning tasks. Performance data reveals that even state-of-the-art LLMs fail to shine, showcasing low accuracy and a whopping gap between AI’s capabilities and human expertise.

This is where it gets spicy. HLE isn’t just another box-ticking exercise; it offers a glimpse into the future of AI development. Here’s a handy comparison of benchmark tests to illustrate the unique nature of HLE:

Benchmark Comparison Focus HLE Differentiation
MMLU 57 subjects, zero-shot knowledge Saturated; HLE emphasizes reasoning over recall.
MMLU-Pro+ Higher-order reasoning HLE uses expert-curated, more challenging problems.
GPQA Graduate-level STEM HLE offers a broader range of subjects.

Implications for the Future

HLE acts as a robust metric for tracking how far AI models actually progress. It’s a tool for scientists, policymakers, and educators to assess AI capabilities without implying that these systems possess full artificial general intelligence (AGI). Let’s face it: A high score on HLE doesn’t mean AIs are on the brink of leading revolutions. They might be great at formal exams but completely clueless about real-world nuances or the art of synthesizing disparate information.

Gotchas: The Limitations and Trade-offs

There’s no such thing as a free lunch, and HLE comes with its concerns. Testing structured problems is one thing; navigating the unpredictable waters of real-world scenarios is another. Critics argue that while HLE may serve as an impressive benchmark, it doesn’t capture the ability to handle messy, chaotic information that humans navigate instinctively every day.

A major limitation is that HLE focuses on closed-ended questions, which don’t lend themselves to AI creativity or synthesis of information in novel ways. Moreover, high scores could signal “inhuman” reasoning — and isn’t that what we need to worry about? Who wants an overconfident AI spouting answers as if it were a know-it-all? It begs the question: How much intelligence is too much?

What’s Next: The Future of AI Testing

The future holds intriguing possibilities. As discussions around the implications of HLE unfold, it’s becoming clear that this assessment tool will be critical in evaluating AI’s role in education, safety, and beyond. Hosting the benchmark at agi.safe.ai also opens avenues for educators and curious minds to engage with these public questions and potentially craft fresh, innovative learning experiences.

More research and iterations are needed, especially in exploring how well AI can integrate disparate pieces of information and engage in creative solutions. As AI models grow ever more sophisticated, the means of testing their capabilities must evolve with sophistication that matches their potential.

In essence, HLE is not the end but rather the beginning of a comprehensive understanding of AI capabilities — a vital stepping stone toward figuring out just how smart these artificial minds can get. If the last exam is a sign of what’s to come, the future of AI tests is bound to be as unpredictable as a cat on a hot tin roof.

FAQ

  • What is HLE? HLE stands for Humanity’s Last Exam, a benchmark aimed at assessing AI’s advanced reasoning capabilities.
  • How many questions are in HLE? HLE consists of 2,500 public questions and ~500 holdout questions.
  • What subjects does HLE cover? It spans a range of disciplines including mathematics, natural sciences, humanities, and more.
  • Why is HLE important? It challenges AI models to demonstrate understanding and problem-solving skills rather than mere recall.
  • What are the limitations of HLE? Critics argue it does not effectively evaluate AI’s ability to navigate real-world scenarios and may focus too heavily on closed-ended questions.

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.