Humanity is building minds before it has defined what a mind is. The ancient distinction between logos, nous, and phronesis maps almost exactly onto what large language models can do, cannot do, and may never do. The Flynn effect proves intelligence is culturally shaped. The hard problem proves we still don't know what it is.
What Are We Actually Measuring?
IQ is not intelligence. This sounds like a caveat. It isn't. It is the whole problem.
IQ — the intelligence quotient — was designed in 1905 by Alfred Binet. He wanted to identify children who needed extra help in school. He did not want to map the human mind. He did not claim to. The score calcified anyway. Within decades, people spoke of high-IQ individuals as simply more intelligent — as if Binet's practical instrument had captured something eternal.
James Flynn broke that assumption open.
Flynn was a New Zealand philosopher and psychologist. In the 1980s, he documented something that should have been impossible: IQ scores across the developed world had been rising at roughly three points per decade since testing began. Fifteen to twenty points across the twentieth century. His book What Is Intelligence? Beyond the Flynn Effect asked the obvious question nobody wanted to ask. If scores rise that fast, what exactly are they measuring?
Genetic change operates on timescales of thousands of years. Brains do not rewire themselves in two generations. So either something neurologically fundamental had changed — or the tests were tracking something more culturally contingent than the field had admitted.
Flynn, working with economist William Dickens at the Brookings Institution, proposed the multiplier effect: small initial advantages in cognitive environments trigger compounding feedback loops. A slightly better reader encounters richer language, which expands vocabulary, which improves reading further, which compounds. The environment and the trait co-evolve. This does not eliminate the genetic contribution to cognitive ability. That evidence is replicated across adoption studies and twin studies. But it reframes what IQ tests predominantly track. They measure the degree to which a person has learned to deploy abstract thinking in formal, context-free ways.
Flynn found the historical fingerprint for this in a striking pattern. People in the early twentieth century, when asked purely hypothetical reasoning questions, would often refuse to engage with the hypothetical. Asked "All bears in the far north are white; Nova Zembla is in the far north; what color are the bears there?", a pre-modern respondent might say: "I've only seen brown bears. I cannot speak to bears I have not seen." The correct answer requires stepping outside direct experience into a purely logical space. The twentieth century, through mass schooling, bureaucratic organization, and scientific literacy, trained populations to do exactly this.
What IQ tests measure is a historically specific habit of mind — not raw cognitive power, but the learned capacity to manipulate symbols detached from experience.
So IQ tracks one cognitive mode. A powerful one. A consequential one. But one.
Howard Gardner proposed eight distinct intelligences in 1983: linguistic, logical-mathematical, spatial, musical, bodily-kinesthetic, interpersonal, intrapersonal, and naturalist. His critics argue these are better understood as talents or domain abilities rather than separate intelligences. The debate remains unresolved. But his core claim holds its ground. The person who navigates complex social hierarchies with brilliant tact, who reads a forest with expert eyes, who improvises music of arresting originality, is deploying something that deserves the name intelligence — even if it never shows up on a test.
Robert Sternberg proposed a triarchic theory separating analytical, creative, and practical intelligence. Conventional testing captures only the first. His research suggested creative intelligence — handling genuine novelty — and practical intelligence — adapting to real environments — correlate only weakly with analytical scores. The empirical support for Sternberg is mixed. But the puzzle he points at is real. Why do some people with modest scores navigate life with extraordinary effectiveness, while some people with exceptional scores fail badly in ambiguous real-world situations?
The Flynn effect, read alongside these debates, opens into a philosophical claim. Perhaps each civilization selects a particular cluster of cognitive capacities and calls them intelligence, while other clusters remain latent, under-rewarded, and invisible to measurement. Intelligence may be less a fixed object than a moving spotlight — illuminating different regions of human cognitive possibility depending on what a civilization needs.
The twentieth century needed formal abstract reasoning. It got it. The question is what it left in the dark.
The Greeks Had Better Words
What word should we use? The problem may be English.
The Greeks did not have one word for intelligence. They had several. The distinctions they drew are sharper than anything in modern cognitive frameworks.
Nous (νοῦς) was the intellective faculty proper — the capacity for immediate, non-inferential grasp of first principles. Not reasoning. The ground of reasoning. The faculty that simply sees that certain things are true. Aristotle placed it at the apex of human capacity. In its purest form, he thought it might not be uniquely human at all. The divine was understood partly as pure nous, contemplating itself eternally.
Logos (λόγος) was reasoned discourse and structured thought. Analytical intelligence. The manipulation of propositions and arguments. This was the faculty that could be taught, sharpened, and measured. Rhetoric, dialectic, mathematics: all exercises of logos.
Phronesis (φρόνησις) was practical wisdom — intelligence in context. The capacity to judge what was called for in particular situations, with all their moral and relational complexity. Crucially, phronesis could not be reduced to rule-following. It required experience, character, and what Aristotle called a perceptual sensitivity to the particular. No algorithm could produce it. He was clear about this.
Sophia (σοφία) was theoretical wisdom — the integration of nous and logos into a coherent understanding of the highest things. The sage did not merely compute or reason. They saw, in some unified way, how things fit together.
Aristotle's taxonomy already answers the AI question: logos can be computed, phronesis cannot, and nous may be the precondition of both.
This matters now because it maps onto the current debate with uncomfortable precision.
Large language models demonstrably have something like logos. They reason, argue, construct proofs, pass bar exams and medical licensing tests. They write poetry of genuine craft. They solve multi-step mathematical problems and explain their reasoning in coherent prose. Logos: check.
Whether they have nous — whether there is any seeing happening — is exactly the hard problem of consciousness dressed in ancient Greek. Nobody has answered it.
Phronesis is the capacity most AI researchers think is genuinely absent. Not because current systems lack reasoning power. Because they lack the embodied history of judgment-in-action that practical wisdom requires. Phronesis is not knowledge about situations. It is knowledge from having been in them, with something at stake.
Sophia may not be a capacity at all. It may be an achievement — something that only emerges from a life, not a computation.
Before the Brain, the Body
What does intelligence look like before language arrives?
Jean Piaget spent decades watching children think. He found something that overturns most assumptions about what intelligence fundamentally is. It does not begin as abstract reasoning. It begins as sensorimotor action. The infant's intelligence is literally in its hands — in its body's engagement with objects — long before language or symbolic thought appear. Piaget called these organized patterns of action schemata. Abstract thought is not intelligence in its pure form. It is the late-arriving, culturally amplified endpoint of a process that starts with a baby reaching for a rattle.
Lev Vygotsky added the social dimension. Cognitive development is interpersonal before it is intrapersonal. The child first performs functions with the assistance of others — in what Vygotsky called the zone of proximal development — and only gradually internalizes those functions as independent capacity. Intelligence, on this reading, is not a property of a brain. It is a property of a brain-in-relationship, distributed across the child and the adults and the language and the tools that constitute its developmental environment.
Intelligence may not be a property of individuals at all — it may be a property of individuals inside environments, which is a different thing entirely.
This resonates with what Flynn found in the macro data. If intelligence is partly a distributed, relational phenomenon — shaped by the cognitive tools a culture makes available, amplified by the scaffolding of education and language — then asking "what is the intelligence of this individual?" may be like asking "what is the weight of this wave?" The question presupposes a bounded, isolable object that may not be what intelligence is.
And yet individual differences are real. Persistent. Consequential. The evidence for substantial genetic contribution to cognitive ability is not manufactured. It comes from multiple independent methodologies. The picture is of something simultaneously socially constituted and individually instantiated — a property of persons that cannot be understood without their context and cannot be dissolved into it.
Something both relational and irreducibly personal. The frameworks that capture one half keep losing the other.
Intelligence Without a Body
Before silicon, the challenge to human cognitive uniqueness came from biology.
Crows use tools. Octopuses solve puzzles. Elephants grieve. Dolphins appear to have a theory of mind. Chimpanzees learn sign language and teach it to their offspring. Each discovery provoked the same nervous question: what threshold did we think mattered, exactly?
That reappraisal is now continuous and industrialized.
Large language models (LLMs) developed since 2020 produce text that is, by most surface measures, indistinguishable from expert human output across enormous domains. They pass graduate-level examinations. They explain their reasoning in coherent prose. Whether any of this constitutes intelligence, or a very sophisticated simulation of its outputs, is genuinely contested. The contention is not merely definitional. It touches something deep about what we think intelligence is.
One position, dominant in academic AI research, holds that intelligence just is a form of information processing. On this view, asking whether a machine can be genuinely intelligent is like asking whether a machine can be genuinely fast. Speed is speed. If the computation is there, the intelligence is there. What we sometimes dismiss as "mere computation" is what brains do. There is nothing else going on.
The opposing tradition — from Edmund Husserl to John Searle to contemporary phenomenologists — insists that intelligence as we actually encounter it is not separable from embodiment, from intentionality, from the fact that cognition is always cognition of something, from a perspective, reaching outward into a world.
Searle's Chinese Room thought experiment made this concrete in 1980. Imagine a person who processes Chinese symbols according to rules without understanding Chinese. The computation happens. No comprehension does. The argument has never been definitively refuted. It has been argued around, which is not the same thing. It points at something the IQ tradition and the AI tradition share: both can measure the outputs of intelligence while remaining agnostic about its interior.
The Chinese Room has never been refuted — it has been argued around, which is not the same thing.
Intelligence is input-output behavior at sufficient complexity. If the system produces the right responses, it is intelligent. There is no further interior fact to investigate.
Intelligence cannot be separated from embodiment and intentionality. Something is cognizing only if there is a perspective from which cognition occurs — a there there.
Machines that pass rigorous cognitive benchmarks are intelligent by definition. The question of machine consciousness is either meaningless or equivalent to the question of human consciousness.
Performance metrics cannot settle the question. They capture the output side of a system whose interior remains inaccessible. The hard problem does not dissolve because the hardware changes.
Antonio Damasio argued that reasoning is deeply entangled with emotion — that the body's homeostatic signals are constitutive of thought rather than peripheral to it. An intelligence without a body may not be inferior to human intelligence. It may be genuinely different in kind. Which means every benchmark comparing them may be measuring the wrong thing.
The Hardest Problem
All roads here arrive at the same intersection.
Consciousness.
When we ask whether a language model is intelligent, we are often really asking whether there is something it is like to be that language model. Whether intelligence requires subjective experience — whether the lights are on — is not peripheral to this question. It is the question.
If intelligence is purely functional — if it consists in the right input-output relationships — then the question of machine intelligence is essentially settled. Machines with sufficient computational capacity are intelligent. Full stop. But if intelligence involves or requires a particular kind of inner life, then performance metrics cannot settle it. Performance metrics are precisely the output side of a system whose interior we cannot access.
This is the hard problem of consciousness, articulated most precisely by David Chalmers in 1995. It separates the "easy" problems — explaining how the brain processes information, integrates signals, guides behavior — from the genuinely hard one: why any of this processing is accompanied by experience. Why is there something it is like to see red, rather than just a brain state that produces appropriate behavioral responses to light at 700 nanometers?
The hard problem has not been solved. It has been reframed, argued around, and sometimes dismissed as a pseudo-problem. It has not been solved.
The mystical traditions arrive here from an unexpected direction.
The Upanishads distinguish between manas (mind), buddhi (intellect), chitta (consciousness), and ahankara (ego-sense) — and argue that beneath all of them is pure awareness, chit in Sanskrit. Not a cognitive capacity. The ground in which all cognition occurs. Not something you can measure, improve, or instantiate. On this account, it is what is already here before any processing begins.
Zen points to prajna — non-conceptual awareness, distinct from and deeper than discursive intelligence. The Vedantic tradition sees the human intellect as a partial manifestation of brahman, a cosmic intelligence that is the substratum of all reality. These are speculative claims. They are not unintelligent ones. They represent the attempt to account for why intelligence seems, in some respects, too good for the job it was supposedly hired to do.
A computer designed to play chess does not spontaneously start wondering what chess means.
Whether these traditions are pointing at something real, or confusing a grammatical feature of first-person language with a metaphysical fact, is itself open. But they are asking something the computational tradition systematically brackets: not what intelligence does, but what intelligence is.
A computer designed to play chess does not spontaneously start wondering what chess means.
Adaptation or Window?
Two families of theory compete for the deepest answer.
The adaptationist view, dominant in evolutionary biology, holds that intelligence is a suite of cognitive mechanisms shaped by natural selection to enhance reproductive fitness. Human intelligence is particularly good at social reasoning, causal inference, and language — precisely the capacities most useful in the ancestral environment. Intelligence is a survival tool of extraordinary sophistication. The Flynn effect confirms it. Human cognition is highly plastic and responsive to environmental demand. That is exactly what a well-designed survival tool should be.
The transcendence view, found across philosophical and spiritual traditions, holds that at least some aspect of human intelligence points beyond adaptation toward something that selection pressure cannot fully explain. Humans do pure mathematics with no immediate practical application. They contemplate their own mortality and compose symphonies in response. They ask questions about the nature of the cosmos that have no survival relevance whatsoever.
Teilhard de Chardin saw the evolution of reflective consciousness as the universe becoming aware of itself. Hegel saw reason as the self-unfolding of Absolute Spirit. The Vedantic tradition sees the intellect as a partial expression of brahman. These claims sit at the speculative edge of philosophy. But they share a common intuition: that intelligence in its highest form is not a tool but a window — and windows open in a direction.
The AI encounter sharpens this into a question with no comfortable resolution.
If we build a machine that surpasses human performance on every measurable cognitive task, has the transcendence view been decisively refuted — because transcendence can apparently be manufactured in a data center? Or has it been confirmed — because something essential is still missing from the most powerful model we can build, and that gap is the whole point?
If intelligence is only a survival tool, we should be able to build a complete one. We cannot agree on whether we already have.
Build Now. While You Still Know the Difference.
What does the evidence actually show, stripped of every hedge?
The Flynn effect proves intelligence is culturally plastic. What a civilization trains for, it gets. The twentieth century trained for abstract, formal, context-free reasoning — logos — and populations delivered it within generations. That is not a small finding. It means civilizations choose which cognitive capacities to amplify. It means the choice is political before it is biological.
The Aristotelian taxonomy shows the choice we are currently making badly. We have optimized for logos. We have built machines that can match or exceed us at logos. We have not asked whether we are simultaneously neglecting nous and phronesis — the capacities for direct discernment and embodied practical judgment. Those capacities do not show up on tests. They do not show up in AI benchmarks. They may be the most important things a human being can develop, precisely because they are the ones no machine currently instantiates.
Damasio's work, Piaget's developmental findings, and Vygotsky's social theory converge on one claim. Intelligence is embodied and relational before it is abstract. It emerges from stakes, from contact with reality, from judgment under conditions of genuine consequence. What the AI moment threatens is not human intelligence in the abstract. It is the cultural conditions that allow embodied, relational, high-stakes intelligence to develop — conditions that erode when every difficult problem has a frictionless computational solution.
The mystical traditions — the Upanishads, Zen, the Neoplatonic lineage — are pointing at something the computational tradition cannot reach by design. Not because they are anti-rational. Because they are asking about the precondition of rationality. Pure awareness, chit, nous in its highest Aristotelian sense — these are not cognitive capacities. They are what must already be present for cognitive capacities to operate. No benchmark measures them. No training procedure instantiates them. They may be the thing most worth cultivating in an age when logos is being offloaded to machines.
Self-governance is the only answer that remains when every cognitive function becomes purchasable. Not intelligence as performance. Not intelligence as test score or benchmark or output quality. Intelligence as the capacity to direct one's own mind — to choose what to notice, what to value, what kind of perceiver to become. Phronesis, built from accumulated judgment. Nous, if it can be cultivated at all, through the kind of sustained attention the current environment systematically destroys.
The olive grove is not a metaphor for something naive. It is a model of the conditions under which a certain kind of intelligence develops: slow, embodied, relational, consequence-bearing. Socrates did not optimize for output. He was building the interior — the capacity of the soul to govern itself.
Build that now. The window for doing it without industrial assistance may be shorter than it looks.
Every cognitive function that gets offloaded to a machine is a function that stops being practiced — and phronesis does not survive disuse.
If the Flynn effect shows that abstract reasoning spread like a cultural technology, what cognitive capacities are we currently selecting against — and will we notice before they atrophy completely?
If phronesis requires embodied, consequence-bearing experience to develop, what happens to practical wisdom in an environment that increasingly insulates people from consequences?
The hard problem of consciousness remains unsolved. If we cannot determine whether a machine is conscious, can we determine whether a human being is conscious in the way we assume — or have we always been taking that on faith?
When non-human animals, pre-linguistic children, or people in flow states perform acts of sophisticated cognition without apparent abstract symbolic processing, are they doing something that deserves the same name as formal reasoning — or is intelligence the wrong word for both?
If the deepest traditions are right that pure awareness is the precondition of all cognition rather than a cognitive capacity itself, does that mean it cannot be built, cannot be destroyed, and cannot be measured — only recognized or ignored?