The End of Traditional Benchmarks and the Need to Evaluate AI in the Real World
AI models outperform humans in isolated tests but fail in practice. Experts advocate for a shift toward benchmarks based on workflows and human collaboration.
The global artificial intelligence industry is experiencing a growing paradox: while models reach impressive milestones in laboratory tests, their practical implementation in real-world environments often results in inefficiency and frustration. For decades, the metric for AI success was based on direct comparisons between machines and humans in isolated tasks, such as writing code or solving mathematical problems. However, this approach, while seductive for its simplicity and ability to generate headlines, ignores the fact that AI does not operate in a vacuum, but rather within complex, collaborative, and often chaotic ecosystems.
The limits of laboratory testing
The current landscape of AI evaluation is dominated by static benchmarks that focus on binary answers—right or wrong—and processing speed. This method creates an illusion of competence, where a model may show 98% accuracy in a controlled environment but fail miserably when integrated into a hospital or a legal department. The fundamental flaw lies in the fact that these tests ignore organizational dynamics, interpersonal interaction, and the evolutionary nature of human decisions, which rarely depend on a single isolated data point.
The fallacy of performance in complex environments
Research conducted between 2021 and 2024 at healthcare institutions in the UK, the United States, and Asia clearly demonstrates the disconnect between technical performance and operational utility. Doctors using AI tools approved by regulatory bodies, such as the FDA, often discover that, instead of accelerating diagnosis, the technology introduces delays. This occurs because hospital workflows require coordination among radiologists, oncologists, and nurses, in addition to compliance with specific regulatory standards. AI, having been tested outside of this context, becomes an obstacle rather than an assistant.
A new approach called HAIC
To mitigate these risks and avoid what has become known as the “AI graveyard”—where expensive technologies are abandoned after failing in implementation—the proposal for HAIC (Human-AI, Context-Specific Evaluation) benchmarks has emerged. Unlike traditional tests, this methodology proposes a radical shift in how we evaluate success:
- Change in the unit of analysis: Evaluating the performance of teams and workflows, not just individual software.
- Expansion of temporal scale: Analyzing the impacts of AI over weeks or months, rather than in a single interaction.
- Measurement of organizational outcomes: Focusing on the quality of coordination and the ability to detect errors, rather than just speed.
- Analysis of systemic effects: Considering the direct and indirect consequences of AI implementation across the entire production chain.
Impacts on the market and society
The insistence on metrics that do not reflect reality creates regulatory blind spots and wastes significant financial and technical resources. When organizations invest in solutions that do not deliver as promised, it leads to an erosion of public and internal trust in the technology itself. Governments and companies that rely on superficial benchmarks to decide on AI adoption end up assuming disproportionate risks, operating with data that lacks ecological validity. The transition to contextual evaluations is, therefore, an economic and ethical necessity to ensure that AI is a tool of sustainable value.
The future of AI evaluations
The path forward requires developers and managers to abandon the obsession with isolated accuracy rankings in favor of stress tests in real-world environments. The future success of an AI will not be measured by its ability to beat a human at a game of chess or a math test, but by its ability to integrate productively into a human team, contributing to complex and collective decisions. The next generation of benchmarks must, by necessity, be as complex and dynamic as the work environments these tools aim to transform, ensuring that technological innovation translates into real human progress.