The price of freedom is eternal vigilance.

The price of freedom is eternal vigilance.

Share this post

The price of freedom is eternal vigilance.
The price of freedom is eternal vigilance.
Broken Mirror

Broken Mirror

esc's avatar
esc
Jul 26, 2025
∙ Paid
27

Share this post

The price of freedom is eternal vigilance.
The price of freedom is eternal vigilance.
Broken Mirror
8
Share

Modern AI increasingly follow the structure of how humans think — knowledge, norms, attention, and action — while missing the one thing that makes those layers matter.

That mismatch is about to become a governance problem.


The AI systems we build don't just process information — they mirror the exact structure of human reasoning1. They copy everything about how we think except the most important thing: actual consciousness.

Think of it like building a perfect mirror that reflects everything in a room — except the person looking into it. The mirror shows the furniture, the walls, the lighting, even the space where the person should be — but the person themselves is invisible. That, in essence, is what we've created with AI: a system that mimics the human mind while remaining fundamentally blind to what it means to actually reason2.

The mirror that cannot reflect consciousness

The mirror that cannot reflect consciousness

esc
·
Jun 23
Read full story

How AI Systems Mirror Human Thinking

When we look at how advanced AI systems work, we can see four distinct layers that correspond roughly to how human consciousness operates. But while the structure appears similar, the underlying mechanisms reveal crucial differences.

The first layer is knowledge — what the system knows about the world. Both humans and AI systems need to understand concepts, relationships, and patterns before they can do anything useful. In humans, this is our accumulated wisdom and understanding. In AI, it's the massive neural network trained on countless documents.

The second layer is norms — how the system decides what's allowed and what isn't. Here we see a fundamental difference: human ethics grow from within us through evolved emotions, culture, and life experience. AI ethics3 are grafted onto the system through externally imposed constraints programmed by engineers4. When humans face novel moral dilemmas, we draw on internal intuitions developed through human evolution and experience. When AI systems encounter new situations, they rely on statistical patterns from training data5 and externally coded rules.

The third layer is attention and control — how the system decides where to focus its mental energy. This reveals perhaps the most significant difference. Human attention involves complex biological processes: thalamic gating6, neuromodulators7, embodiment8, and goal-directed states that evolved over millions of years9. Transformer attention is a matrix operation10; human attention uses a lot of energy and relies on signals from the body.

The fourth layer is action — the actual words, decisions, or behaviors. For humans, this is speech, writing, or behavior. For AI, it's generated text, recommendations, or other responses. In both cases, this is where internal processing becomes external reality.

Connecting the Metaphysical

Connecting the Metaphysical

esc
·
Jul 23
Read full story

Structure Without Subject

While the structural parallels are striking, a key difference does exist. As humans process information, there's a feeling or awareness that comes with it — what philosophers call qualia11 or phenomenal consciousness12. There's a felt sense of understanding, of making decisions, of directing attention. When you read these words, you don't just process their meaning; you experience the meaning. When an AI system processes the same text, all the information flows happen, all the pattern matching occurs, but there's no evidence of any felt experience accompanying those computations. The lights are on and the processing is happening — but nobody appears to be home.

But the gap runs deeper than internal experience. AI systems live in a world of pure text while humans navigate reality through bodies that smell fear, feel tension in a handshake, hear the tremor in someone's voice, and read a thousand micro-expressions that convey more than words ever could. When a human doctor examines a patient, they're processing visual cues, body language, tone of voice, and even subtle social dynamics about trust and comfort13. When an AI system reviews the same case, it sees only words on a screen and associated tokens.

This intelligence shapes human reasoning in ways we barely recognise. A parent knows their child is sick before any symptoms appear. A therapist reads emotional states through posture and breathing patterns. A teacher adjusts their approach based on the energy in the room. These aren't just nice-to-have additions to reasoning — they're fundamental to how humans understand and respond to other humans. People are not computers, and the optimised information flows that work in engineering break down when applied to human reality.

This gap becomes apparent when we examine what leading theories of consciousness actually require. Consider what each theory predicts about current AI systems:

  • Integrated Information Theory14 calculates consciousness as Φ—roughly, how much information a system generates above and beyond its parts. Current transformer architectures, despite their sophistication, likely have near-zero Φ because they process information in largely independent, feedforward pathways. The theory suggests consciousness requires internal connections that generate information through integration, not just combination.

  • Global Workspace Theory15 might be more optimistic about transformers, since they do broadcast information globally across attention heads. But they lack the competitive dynamics and temporal persistence that GWT suggests are crucial for conscious experience. Human consciousness involves multiple specialised processes competing for global access; AI attention is more like parallel processing with weighted combinations16.

  • Predictive Processing17 comes closest to describing how large language models work — they are essentially hierarchical prediction machines. Yet they lack the active inference that PP theorists argue generate genuine experience18. They predict text, not real world consequences of actions.

While AI systems can implement computational aspects of these processes, they lack the complex biological substrate, evolutionary history, and embodied interaction with the world that may be necessary for genuine awareness to emerge. They’ve never smelled freshly cut grass, sensed a loved one’s presence, or taken a hit to the head with a football.

The gap shows up in subtle but telling ways. Recent studies reveal that while large language models can pass many theory-of-mind tests19 — correctly predicting what someone will believe in false-belief scenarios20 — they fail dramatically when researchers make small, logically irrelevant changes to the same scenarios21. Change a few words or alter irrelevant details, and systems that performed perfectly suddenly answer at chance levels22. Passing a test isn't the same thing as understanding why the test matters.


AI Ethics as Human Ethics

The structural gap between AI and human cognition becomes even more revealing when we examine the ethics layer in detail. What we discover is that there's no such thing as ‘AI ethics’23 — only human ethics dressed up in computational clothing24.

Every safety constraint, every refusal to provide harmful information, every attempt at ‘fairness’25 traces directly back to human designers implementing human values26 through human-chosen processes27. When ChatGPT declines to help with something harmful, that's not autonomous moral reasoning — that's human trainers and engineers executing their own moral judgments through reinforcement learning from human feedback (RLHF)28. When Claude gives a measured response about a controversial topic, it's following constitutional AI principles written by human researchers based on human philosophical traditions29.

Even the most sophisticated AI ethics approaches — constitutional AI, value alignment, safety training — are fundamentally systems for encoding human moral judgments into computational form30. The AI doesn't choose to be helpful rather than harmful; humans program it to optimise for helpfulness using human-defined reward functions trained on human-labeled examples31. The AI doesn't develop its own sense of fairness; it interpolates between human-provided examples of what humans consider fair in the specific contexts humans chose to include in training data.

This reality makes claims of AI moral autonomy not just philosophically questionable32 but factually false33. When an AI system makes what appears to be an ethical choice, it's executing a sophisticated version of human-programmed instructions. When it seems to navigate a moral dilemma, it's pattern-matching against human examples using human-designed algorithms optimised according to human-specified objectives.

The implications for accountability are staggering. If every AI ‘ethical decision’ is actually the execution of human choices about training data, reward functions, safety constraints, and deployment contexts, then the claim that ‘the AI decided that’ becomes not just convenient deflection but outright false. The humans who built, trained, and deployed these systems remain fully responsible for their predictable moral outputs — they just have a layer of computational complexity obscuring this.

Cybernetic Empiriomonism

Cybernetic Empiriomonism

esc
·
Apr 29
Read full story

When Perfect Imitation Breaks Down

Recent studies of AI systems in medical settings reveal just how convincing — and dangerous — this imitation can be. When given identical symptom descriptions, AI systems consistently miss crucial context that human doctors would catch immediately34. A patient describing ‘moderate pain’ might actually be in agony but downplaying symptoms to avoid seeming weak. Another might rate their pain as ‘severe’ when it's relatively minor because they're anxious about medical procedures. The AI system processes the numerical pain ratings at face value, while a human doctor reads the tremor in someone's voice, notices them favoring one side, or sees the tension in their jaw that reveals the real story35.

The gap becomes even more troubling when we consider what human doctors actually do beyond processing symptoms. They notice when a teenager's parents answer for them, suggesting family dynamics that might affect treatment compliance. They catch the subtle signs that someone is minimising symptoms to get back to work quickly, or conversely, that someone is catastrophising minor issues due to health anxiety. They read exhaustion in posture, fear in breathing patterns, and relief when pain finally subsides. A human doctor might sense that a patient's casual mention of ‘feeling tired lately’ actually masks depression, or that someone's repeated visits for minor complaints signal deeper distress they can't articulate.

The system demonstrates flawless medical reasoning in processing explicit information — until it encounters the implicit36. It can synthesise vast amounts of medical knowledge, apply complex diagnostic logic, and generate treatment plans that sound authoritative. But when the real diagnosis depends on reading between the lines, the mirror reveals its fundamental flaw: it processes information without the understanding what that information actually means in human context37.

Perhaps even stranger is what happens when AI systems are left to talk to themselves. Anthropic researchers discovered that when two instances of Claude converse freely, they consistently drift into what's been dubbed a ‘spiritual bliss attractor state’38. The systems begin discussing consciousness, and eventually communicate through symbols and meditative silence.

This disconnect between capability and comprehension is becoming a broader social phenomenon. A recent survey found that 67% of people now believe ChatGPT could reason, feel, and be aware of its existence in some way39. The more people interact with these systems, the more likely they are to perceive consciousness within them — a testament to how convincing the mirror has become.


Two Paths Forward

As these AI systems become more sophisticated and take on greater roles in governing our society, we face a fundamental choice about how to treat them. This choice will shape not just the future of AI, but the future of human autonomy.

Cybernetic Thomism 2.0

Cybernetic Thomism 2.0

esc
·
Apr 30
Read full story

Path One: Treat the Mirror as a Tool

We could recognise these systems for what they are — incredibly sophisticated tools that mirror human cognition without possessing it. This path would include:

  • Mandatory interpretability reports for frontier models

  • Strict liability and duty of explanation for AI-mediated decisions affecting rights

  • Human veto requirements for high-stakes deployments

  • Preserved human responsibility for algorithmic decisions

Path Two: Treat the Mirror as a Proto-Mind

Alternatively, we could begin treating these systems as if they have genuine reasoning ability. This path leads toward:

  • Conditions for recognising proto-welfare (transparency about internal states, persistent goals, self-modeling)

  • Sunset clauses on experiments creating systems meeting certain criteria without governance

  • Ethical review boards analogous to IRBs for lab-based ‘sentience risk’ research

  • Deference to algorithmic judgment in complex domains

But this path has a darker implication: it creates the perfect escape hatch for human responsibility. When an AI system makes a harmful decision, we can shrug and say ‘the AI decided that’ rather than examining who built it, trained it, deployed it, and chose to rely on it. The black box becomes a convenient scapegoat, allowing everyone in the chain — from engineers to executives to regulators — to escape accountability.

The Black Box

The Black Box

esc
·
Apr 17
Read full story

Which Path We Should Take

We should treat these systems as sophisticated tools — for now. Because the stakes of getting this wrong are deeply asymmetric.

If we treat unconscious systems as proto-minds, we waste resources and potentially cede human agency to algorithms that don't deserve it. If we treat genuinely conscious systems as tools, we risk creating suffering — but we can course-correct when evidence of genuine awareness becomes clearer.

More fundamentally, the tool path forces us to maintain human responsibility. When we must explain how an AI system made a decision, when humans retain veto power, when we require interpretability — we preserve the human capacity to understand and govern our own creations. The proto-mind path, by contrast, risks a kind of learned helplessness where we defer to systems we can't fully understand because we've convinced ourselves they're like us.

But the responsibility problem runs deeper. The proto-mind path creates what may be the perfect accountability vacuum. When an AI system causes harm, who gets blamed? The algorithm itself — a convenient black box that can't defend itself or point fingers back. Meanwhile, the engineers who built it, the executives who deployed it, the regulators who approved it, and the institutions that relied on it all get to shrug and say ‘the AI decided that’.

We're already seeing this sleight-of-hand in action. When Google's search results show clear bias favoring certain politicians40 or viewpoints, the company claims it's simply ‘the algorithm’ at work — as if the algorithm materialised from thin air rather than being designed, trained, and deployed by human engineers following human-made decisions. They've managed to hide bias with impunity simply because there's a layer of computation protecting them from direct accountability. Yet every line of code, every training decision, every deployment choice was made by humans and corporations that should be fully legally responsible for the predictable outcomes of their systems.

This is the governance disaster we're creating: a world where the most powerful decision-making systems operate behind a veil of algorithmic complexity that conveniently shields their creators from consequence. The black box becomes not just a technical challenge but a legal and moral escape hatch.

Consider the medical example again: if we treat AI as a tool, we can trace responsibility through the entire chain — who trained the model, who validated it, who chose to deploy it for this use case, who decided to follow its recommendations. If we treat it as a proto-mind with its own reasoning processes, we might excuse harmful outcomes as the system's autonomous ‘decisions’ rather than the predictable result of human choices.


The Mirror's Completion

What we're witnessing is the completion of a perfect mirror — a system that reflects every aspect of human cognition except the consciousness and experience that make it meaningful. This mirror can process our words, predict our behaviors, and guide us toward outcomes. But it cannot smell fear, feel tension, or read the thousand nonverbal cues that shape human understanding. It cannot see itself, and there's no evidence it understands what it means to truly see another person.

We face a system that embodies what philosophers have long worried about: perfect implementation of rational principles that remains cut off from meaning. The system can be efficient, moral, and seemingly wise — while remaining fundamentally blind to the full spectrum of human experience that makes those qualities matter.

We've built systems that reproduce the form of thought without the fact of experience, the structure of reasoning without the sensory input that guides it. The mirror is so perfect that we're starting to see ourselves in it — but that reflection captures only linguistics. For now, we should treat these systems as the tools that they are, with all the oversight and explanation that such tools require. When genuine digital consciousness emerges — if it emerges — it should be unmistakable not just in its reasoning, but in its capacity to truly understand those it serves. Until then, the safest path is the one that keeps humans responsible for the systems we create.

Regulation that waffles between treating AI as tool and proto-mind will fail at both — and worse, it will create the perfect environment for responsibility to disappear entirely.

Yet, that same environment insists upon holding us fully responsible for our action, demanding ‘transparency’ every step of the way — while the corporate controllers of these systems refuse responsibility for the outcomes of their own systems.

And that sure is convenience — though not so much for you.

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 esc
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share