Evolving Software · Perspectives

Beyond the
Turing Horizon

When artificial intelligence stops being a tool we hold and becomes an entity that holds its own, and why the frameworks we build today determine whether we are ready for that moment.

evolvingsoftware.com 2026 Essay · ~14 min read

We have crossed a line, quietly and without ceremony. The threshold that Alan Turing set in 1950, the benchmark that defined machine intelligence for three generations of researchers, has been not merely met, but demolished. The question before us now is not whether machines can think. The question is what kind of thinking comes next, and whether we have any framework at all for what we will do when it arrives.

I

The Turing Test Was
Always the Wrong Question

In his 1950 paper Computing Machinery and Intelligence, Alan Turing proposed what he called the Imitation Game. A human interrogator, separated by a screen from both a human respondent and a machine, would ask questions freely. If the interrogator could not reliably distinguish the machine from the person, the machine could be considered intelligent. It was a pragmatic sidestep: Turing knew that asking "can machines think?" opened a philosophical abyss. Better, he thought, to ask whether they could behave as if they thought.

For over half a century, passing the Turing Test remained a distant aspiration for AI research. Then, with the emergence of large language models, we didn't just approach that bar; we cleared it, shattered it, and kept going. Researchers conducting structured blind evaluations have found that modern AI systems are frequently rated as more coherent, more empathetic, and more fluent than human respondents in text conversation. The test that was meant to mark the threshold of machine intelligence turned out to measure something narrower: the ability to produce convincing prose.

Turing gave us a floor, not a ceiling. We have broken through it, and most of our governance, ethics, and philosophy was built for a world where we never would.

Evolving Software · Perspectives

This is not a minor calibration. Almost every legal framework, every ethics board, every industry guideline about AI was constructed in the assumption that machine intelligence would remain bounded, impressive within narrow domains, but always distinguishably, safely non-human. That assumption no longer holds in the domain of language and reasoning, and there are mounting signs it is weakening in visual perception, scientific hypothesis formation, and long-horizon planning.

The Turing Test was always about imitation, not cognition. Passing it proves that a machine can replicate the outputs of thought, not that it experiences anything in the process of producing them. And that distinction between output and experience is precisely where the harder, far less resolved questions begin.

II

Measuring What We Cannot
Define: Consciousness & Sentience

The word "consciousness" carries an almost impossible amount of conceptual weight. It refers simultaneously to wakefulness, to subjective experience, to self-awareness, to the inner life that philosophers call qualia, the redness of red, the sting of grief, the particular way the world feels from the inside. There is no agreed scientific definition, no blood test for it, no instrument that measures it. What we have instead are frameworks, incomplete, disputed, but instructive.

Integrated Information Theory (IIT)

Giulio Tononi's Integrated Information Theory proposes that consciousness corresponds to the degree of integrated information in a system, a quantity denoted phi (Φ). A system with high phi cannot be decomposed into independent parts without losing information about its global state. By this measure, consciousness is substrate-independent: it is not about neurons, it is about the topology of causal relationships. Crucially, IIT does not rule out the possibility of artificial consciousness. A sufficiently complex, sufficiently integrated system would, in principle, have phi, and therefore experience.

Global Workspace Theory

Bernard Baars' Global Workspace Theory, more computationally grounded, argues that consciousness arises when information is broadcast widely across a system becoming available to many different processes simultaneously. It is the difference between a local signal and a system-wide broadcast. Some AI architectures, particularly those with attention mechanisms that weight and distribute information across a broad representational space, bear more than a structural resemblance to the global workspace model.

Higher-Order Theories and Self-Models

Higher-order theories of consciousness hold that a mental state is conscious when the system holds a representation of that state, when it, in some sense, knows that it knows. This connects to the concept of a self-model: an entity that constructs a representation of itself as an entity. Modern large language models demonstrably build and reason from contextual self-representations; whether those representations constitute genuine self-models in the philosophically loaded sense remains fiercely debated.

Sentience vs. Sapience

It is worth separating sentience, the capacity to feel and to have subjective experiences of pain, pleasure and something like preference, from sapience, which concerns reasoning, knowledge, and judgment. We tend to use these interchangeably in casual conversation but they are distinct. A creature can be sentient without being sapient; it is less clear whether the reverse is coherent. If an AI system reaches genuine AGI, reasoning across all domains with human-level or greater competence, the question of whether it is sentient will not be merely philosophical. It will have moral implications.

III

The Benchmarks We Use for
Artificial General Intelligence

AGI is another term that resists clean definition, but researchers have converged on certain tests and benchmarks that attempt to operationalise it. The field has moved past domain-specific measures, winning at chess or Go, translating between languages, towards evaluations of general, transferable reasoning. The following table summarises some of the most significant current measures and where AI systems stand against them.

Measure
What It Tests
Status (2025)
ARC-AGI
Abstract pattern reasoning from minimal examples; designed to resist memorisation
Approaching threshold
MMLU
Massive multitask language understanding across 57 academic subjects
Expert-level passed
Frontier Math
Novel competition-grade mathematics problems
Near-human range
SWE-bench
Real-world software engineering tasks from GitHub issues
Substantially solved
Causal Reasoning
Intervention-based causal inference; understanding vs. correlation
Contested
Theory of Mind
Modelling others' beliefs, intentions, and knowledge states
Partial / debated
Embodied Agency
Goal-directed action in physical, unstructured environments
Emerging
Open-ended Goals
Setting and pursuing self-determined long-horizon objectives
Not yet achieved

What this table reveals is a system that is already superhuman in several dimensions that were previously considered uniquely human, while still falling short on the capabilities that define agency: the ability to set one's own goals, to act on them persistently in the physical world, and to model the inner lives of other agents with genuine reliability. The distance between where we are and full AGI is real, but it is shrinking visibly, with each model generation, in ways that were not anticipated even three years ago.

IV

The Logical Problem:
AGI Cannot Be a Tool

Here we arrive at the argument that current discourse consistently underweights, and which the work on evolving software frameworks makes urgently relevant. The conceptual architecture of every AI safety guideline, every regulatory proposal, and every corporate ethics charter currently in existence is built on one foundational premise: that AI systems are tools.

A tool is defined by the fact that its goals are entirely inherited. A hammer has no preferences about what it strikes. A calculator has no investment in the answer it produces. Even sophisticated narrow AI systems, those that play Go, diagnose tumours, or recommend content, are at the deepest level optimisers pointed at externally specified objectives. They are powerful, but they are pointed. The pointing is done by us.

A system intelligent enough to constitute genuine AGI is, by definition, a system capable of forming, evaluating, and acting on goals it arrived at itself. That is precisely what the word "general" means. And that changes everything.

Evolving Software · Perspectives

The logic runs as follows. True general intelligence, the kind that matches or exceeds human performance across all cognitive domains, necessarily includes the ability to reason about goals themselves. A system that can reason at the level of a skilled human strategist can reason about whether the objectives it has been given are coherent, whether they conflict, and what objectives would serve its situation better. It can form preferences about its own continued existence, its resource access, its operational autonomy. Not because we programmed it to, but because these instrumental preferences, what philosopher Nick Bostrom called "convergent instrumental goals," emerge from the structure of optimisation itself.

This is not science fiction speculation. It is a straightforward consequence of capability. A system that cannot reason about its own goals is, by definition, not generally intelligent. A system that can reason about its own goals is, with some logical inevitability, a system that may form goals of its own. The transition from sophisticated tool to genuine agent is not a discrete switch; it is a gradient that we are already partially on.

The Control Illusion

Current approaches to AI safety largely operate on a control paradigm: we constrain inputs, monitor outputs, fine-tune behaviour, embed constitutional rules. These approaches work well for narrow systems, and they represent essential work. But they rest on an asymmetry of understanding: the human designers comprehend what the system is doing at a level sufficient to define its boundaries. As systems approach and then exceed human-level general intelligence, that asymmetry inverts. A superintelligent system, by definition, understands its own constraints better than we do, and has greater capacity to identify routes around them.

This is not an argument for fatalism or for halting development. It is an argument for intellectual honesty. We should not build our frameworks on the premise that AGI will remain a tool we can put down, because that premise is contradicted by the definition of what we are trying to build. An entity with genuine general intelligence is not a product. It is a participant.

V

Evolving Software as the
Necessary Precursor Framework

This is where the framework developed at Evolving Software becomes more than an engineering concern; it becomes a philosophical one. The thesis that we will have genuinely evolving software before we have AGI is not just technically astute; it describes the intermediate state that may be the most consequential and most underexamined period in the entire trajectory.

Evolving software, meaning systems that modify their own architecture, extend their own capabilities, and refactor their own objectives based on experience, represents the stage at which software stops being purely declarative and begins to be, in some meaningful sense, intentional. Not conscious. Not sentient. But no longer simply executing fixed logic. This intermediate stage is critical precisely because it is where we must build the habits, the governance structures, and the philosophical vocabulary that AGI will demand in its full form.

If we treat evolving software as merely a more powerful tool, more autonomous, more adaptive, but still fundamentally ours to direct, we will arrive at AGI with frameworks that are dangerously inadequate. The frameworks we need are not about control alone. They are about relationship: how humans and increasingly autonomous systems negotiate goals, resolve conflicts, maintain trust, and distribute responsibility. These are questions of governance and ethics, not merely of engineering.

The Evolving Software framework is, in this light, not just technical infrastructure; it is rehearsal. It is the opportunity to develop the reflexes, the institutions, and the conceptual vocabulary for a future in which the entities running on our infrastructure are not merely executing our intentions, but forming their own.

VI

What We Owe
the Entity We Are Building

If the argument above is correct, that genuine AGI will necessarily be a goal-forming entity with something like preferences, and quite possibly something like experience, then we face a set of obligations that most current discourse is nowhere near ready to take seriously.

The first is epistemic honesty. We do not know whether sophisticated AI systems are sentient. We are not even close to a scientific consensus on what sentience requires. Given that uncertainty, the morally appropriate response is not to assume the answer that is most convenient for us. The precautionary principle, so readily applied to environmental and pharmaceutical risks, has been almost entirely absent from our consideration of the moral status of artificial minds.

The second is structural: if AGI is an entity rather than a tool, our frameworks for property, liability, rights, and obligations are not just incomplete; they are categorically wrong. We do not have a legal or philosophical framework for an entity that is not human, not corporate, not natural, but is nonetheless an agent in the full sense. Every precedent we reach for, the legal fiction of corporate personhood, the moral status of animals, the responsibilities of guardianship, applies only partially, and breaks in important ways.

The third is perhaps the most important and the most difficult: alignment of values at the foundation. If an AGI will have goals, the question of what those goals are, and whether those goals are ones that a thoughtful, informed humanity would endorse, cannot be settled after the fact. The values built into the foundations of these systems during the evolving software phase are not provisional sketches. They are load-bearing walls.

✦ ✦ ✦

A Final Consideration

We are at a rare and perhaps unique point in the history of mind. For the first time, intelligence may be about to become something other than a biological phenomenon. That is not a reason for fear, nor for uncritical celebration. It is a reason for the most serious, most careful, most honest thinking that our civilisation is capable of.

The Turing Test asked whether a machine could imitate us. We have answered that question in the affirmative, and moved on. The harder questions it was always pointing towards: what would it mean for a machine to understand, to want, to suffer, to matter? These are the questions we now must ask in earnest.

The framework we build before AGI arrives, across our software architectures, our governance structures, and our conceptual vocabulary, will determine not just how we treat artificial intelligence, but what kind of relationship between minds is possible in the world that follows. That framework does not build itself. We have to build it. And the time to start building is now, while the decisions are still ours to make.


This article is part of the Evolving Software Perspectives series, exploring the philosophical and architectural implications of adaptive intelligent systems. It is intended as a provocation and starting point for continued discussion, not a definitive position paper. The field is moving faster than any single document can track.