The AI Understanding Paradox: Are We Looking in the Wrong Places?

Table of contents
- The Actor Analogy: Performance vs. Understanding
- The Missing "Why" in AI Understanding
- The Experience Gap: What AI Cannot Know
- The Projection Problem: Finding Cats Where There Are Only Patterns
- The Limits of Interpretability: Technical Solutions vs. Philosophical Questions
- Beyond Mechanistic Explanations: The Interpretability Paradox
- Conclusion: Rethinking Our Approach to AI Understanding

In our relentless pursuit of artificial intelligence that can match or exceed human capabilities, we may be overlooking fundamental questions about the nature of understanding itself. Recent research into Large Language Models (LLMs) has revealed fascinating insights about their behavior - including the concept of a "jailbreak tax," which examines the tradeoff between an AI's usefulness and the bypassing of its safety measures.
But beneath these technical investigations lies a more profound question: Are we truly understanding what we're working with? And are we asking the right kinds of questions about these increasingly sophisticated systems?
The field of AI interpretability has emerged precisely to address this gap - to peek inside the "black box" of neural networks and understand how they make decisions. Explainable AI (XAI) initiatives strive to make AI systems' outputs transparent and their reasoning processes clear to humans. Yet despite these valuable technical efforts, a philosophical chasm remains.
The Actor Analogy: Performance vs. Understanding
Consider a masterful Shakespearean actor who delivers lines with perfect timing and emotional resonance. They know what to say and how to say it to move an audience. But this doesn't necessarily mean they understand why the character feels certain emotions, the historical context of the play, or its deeper philosophical meanings.
Similarly, LLMs excel at predicting what words should come next based on patterns they've observed in vast datasets. They demonstrate impressive information about what happens in text. But do they understand why?
The Missing "Why" in AI Understanding
This "why" information - causal relationships, intentions, subjective experiences, and underlying meaning - forms the core of human understanding. When we examine the internal workings of an LLM - its layers of numbers and computations - are we actually finding this "why" information, or merely projecting our interpretation onto mathematical patterns?
The concept of "alignment" - making AI systems safe and beneficial - raises even deeper questions. We train these systems to avoid harmful outputs and to be helpful and honest. But where do human values and safety guidelines truly reside? Are they inherent properties of a system, or are they concepts we project onto the world based on our shared experiences and consciousness?
The Experience Gap: What AI Cannot Know
If our values are indeed projections rooted in human experience, can we genuinely expect to find them residing within a system that operates purely on statistical relationships between text tokens?
Consider these thought experiments:
Can an AI that has only processed text descriptions of flavors truly taste chocolate?
Can an AI that has only read about emotions genuinely feel joy or sorrow?
We can teach an AI to recognize patterns in light wavelengths and correlate those patterns with color words humans use. But the subjective experience of seeing "red" - that's a fundamentally different kind of reality.
The Projection Problem: Finding Cats Where There Are Only Patterns
When researchers identify what they call a "cat detector circuit" within an AI's neural network, are they describing something the AI is, or simply finding a pattern in its internal state that correlates with our human concept of a cat - a concept that originates outside the AI?
The complexity of modern AI models resembles an intricate roadmap of a vast city. Researchers diligently study these maps, identifying key intersections and pathways - revealing how information flows. But understanding traffic patterns isn't the same as understanding why people live in that city, what their hopes and dreams are, or the history that shaped its culture.
The Limits of Interpretability: Technical Solutions vs. Philosophical Questions
The field of interpretability research has made remarkable strides in recent years. Techniques like attention visualization, feature attribution, and mechanistic interpretability have helped us map the internal processes of neural networks. Researchers can now identify specific neurons or circuits that activate in response to particular inputs, trace information flows, and describe functional components within these complex systems.
Yet interpretability research faces a fundamental challenge: it can show us how an AI system processes information, but struggles to reveal whether the system truly understands in any human-like sense. When we examine activation patterns and describe them as "detecting cats" or "identifying toxic content," are we discovering genuine understanding, or merely convenient human labels for statistical correlations?
Similarly, explainable AI approaches that generate natural language explanations for model decisions may provide plausible-sounding rationales, but do these explanations reflect the actual internal processes, or are they post-hoc justifications that align with human expectations?
Current approaches in AI research may be focused on mapping the "how" while assuming they are capturing the "why." Researchers look for meaning in the statistical patterns carved by data, perhaps expecting to find human concepts embedded within.
Is this quest to find human-like understanding within systems built on probabilistic token prediction a productive path? Or are we engaging in a kind of digital alchemy, projecting our own complex reality onto a fundamentally different kind of machine?
Beyond Mechanistic Explanations: The Interpretability Paradox
The paradox at the heart of interpretability efforts is this: the more technically sophisticated our explanations become, the more they may distance us from the human understanding we seek to compare them with. Human understanding is embodied, contextual, and deeply connected to our lived experience - qualities that may not have direct parallels in the mathematical operations of neural networks.
When we identify a "concept direction" in a model's latent space or trace the flow of gradients through a network, we're creating useful technical descriptions. But these descriptions exist in a fundamentally different domain than human conceptual understanding. The map is not the territory; the explanation is not the understanding.
Conclusion: Rethinking Our Approach to AI Understanding
Perhaps the most crucial step in navigating AI's future isn't just building more complex models or refining existing interpretability methods. Perhaps it's about pausing to ask more fundamental questions:
What is understanding in the context of AI?
Where does meaning truly reside?
What kinds of systems are even capable of embodying the values we deem essential for beneficial use?
How should we interpret the results of mechanistic interpretability research?
What would constitute genuine understanding in an artificial system?
Until we grapple with these deeper philosophical questions alongside our technical investigations, we risk developing increasingly sophisticated tools while missing fundamental aspects of what makes understanding meaningful. We might be building elaborate maps while forgetting the terrain they represent.
The true challenge of AI may not be purely technical, but conceptual - requiring us to reconsider our assumptions about what it means to understand, to value, and ultimately, to be. Interpretability research isn't just about making AI explainable - it's about understanding the very nature of understanding itself.
Subscribe to my newsletter
Read articles from Gerard Sans directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by

Gerard Sans
Gerard Sans
I help developers succeed in Artificial Intelligence and Web3; Former AWS Amplify Developer Advocate. I am very excited about the future of the Web and JavaScript. Always happy Computer Science Engineer and humble Google Developer Expert. I love sharing my knowledge by speaking, training and writing about cool technologies. I love running communities and meetups such as Web3 London, GraphQL London, GraphQL San Francisco, mentoring students and giving back to the community.