There is a quiet agreement at the heart of modern AI development one so deeply embedded it rarely gets examined. The assumption is this: if a system produces outputs indistinguishable from human reasoning, it is reasoning. It sounds defensible. It may even sound obvious. But follow the logic far enough and you'll find it folds back on itself into a perfect, airtight loop a tautology masquerading as a theorem.
Big Tech has bet the house on this loop. The argument goes roughly: scale the data, scale the parameters, scale the compute and eventually, intelligence emerges. But emergence from what? From human text. Validated by what? Human-constructed benchmarks. Defined as what? The ability to match human-generated answers. This is not a path to Artificial General Intelligence. This is an extraordinary, multi-billion-dollar exercise in circular reasoning.
Two significant bodies of empirical research examining the structural limits of autoregressive transformers and the ceiling effects in imitation learning have quietly confirmed what the philosophy already suggested. The architecture itself is the problem. And scaling does not escape it.
The Three-Layered Tautology
The tautological trap isn't a single flaw. It operates simultaneously at three levels logical, epistemic, and evaluative each reinforcing the others. Together, they form a closed system that cannot, by its own nature, step outside itself.
Layer One: The Logic Loop
The foundational claim of modern LLM development is that intelligence is the ability to produce correct sequences of information. The model is then trained specifically to produce those sequences. The conclusion that the model is therefore intelligent commits the precise fallacy it claims to solve.
This isn't philosophical hair-splitting. Empirical research has documented exactly what this circularity produces in practice. A landmark study of autoregressive reasoning found that models frequently learn surface patterns rather than genuine algorithms, with teacher-forcing enabling what researchers call "Clever Hans cheating" the model exploits revealed prefixes rather than developing any real planning capability.[1] In controlled path-finding tasks, both Transformer and Mamba architectures achieved accuracy no better than random guessing despite achieving perfect in-distribution training performance.
The model hasn't solved anything. It has retrieved a high-probability state and we've mistaken retrieval for reasoning.
Pattern Matching vs. Genuine Reasoning Elicit Autoregressive Review, 2024
Perhaps the most striking demonstration of this came from studying OpenAI's o1 a model explicitly optimised for reasoning, positioned as a step beyond standard LLMs. Even here, probability sensitivity persisted. When tested on low-probability task variants, accuracy collapsed to 47%. On high-probability variants, it reached 92%.[2] The model hadn't learned to reason. It had learned to recognise and reproduce the most statistically likely answer which, on familiar problems, often resembles correct reasoning. On unfamiliar ones, it doesn't.
Layer Two: The Human Ceiling
The second tautology is epistemic. LLMs are trained on the sum of human knowledge. If AGI let alone ASI is defined as intelligence that surpasses and generalises beyond human ability, then a system whose training distribution is entirely bounded by human output has a hard theoretical ceiling: us.
The model's conception of truth is defined by the statistical consensus of its training corpus. It cannot generate a new scientific law that doesn't exist somewhere in that distribution, because anything outside that distribution is technically an error an artefact to be corrected, not a discovery to be celebrated.
The imitation learning research confirms this structurally. Orca, one of the most sophisticated imitation-trained models, was trained on five million ChatGPT responses and one million GPT-4 responses. It achieved competitive performance eventually approaching parity with ChatGPT. But it consistently, fundamentally could not exceed GPT-4.[3] The teacher is the ceiling. The student cannot exceed what it was shown.
Orca demonstrated that imitation learning "can approach but not exceed teacher model capabilities." Despite training on millions of GPT-4 explanations, it remained bounded by the teacher's performance a fundamental architectural constraint absent from symbolic reasoning approaches, which can achieve perfect accuracy within their formal specification. >100% improvement over baseline yet hard-capped below its own teacher.
This isn't a temporary limitation of current models. It is a structural consequence of the paradigm. You cannot imitate your way to something no one has ever thought before. You cannot pattern-match your way to a hypothesis the training set doesn't contain. The model knows what we know, precisely because we told it what we know. That's the tautology: the knowledge boundary is a mirror of our own.
Layer Three: The Benchmarking Paradox
The third tautology completes the circuit. We evaluate these models using benchmarks the Bar Exam, GSM8K, Big-Bench Hard, the GPQA constructed from the same body of human knowledge the models trained on. A model that achieves high scores on the Bar Exam is being graded on an exam where the answer key was arguably part of its training distribution. The test is the training data, restructured.
The research literature calls this "level-1 reasoning" the retrieval of embedded knowledge rather than genuine causal or logical inference.[4] The distinction becomes stark when models are tested on truly fresh data. In a landmark causal reasoning study, LLM performance degraded monotonically as benchmark corpora became more temporally distant from training data. On CausalProbe-2024 a fresh corpus nearly unseen by the models performance dropped significantly compared to earlier benchmarks on the same task type.
We are essentially grading a student on a test where they were allowed to memorise the answer key beforehand then announcing they have achieved human-level intelligence.
The grokking research adds an uncomfortable nuance here. It turns out that transformers can eventually develop something resembling genuine generalisation but only after "extended training far beyond the point of overfitting."[5] Most evaluations terminate far before this point, observing only memorisation and incorrectly concluding that the task cannot be solved. But even when grokking occurs, the capability remains constrained to tasks structurally similar to those in training. Composition tasks still fail. Multi-hop inference still degrades. The architecture is not liberated by grokking it is merely trained to memorise at a higher level of abstraction.
The Synthetic Data Escape Hatch (And Why It Fails)
At this point, the standard counter-argument arrives: synthetic data. If models can train on their own outputs, surely they can bootstrap beyond the human knowledge ceiling? Can't they generate new knowledge by training on their own generations?
No. And the research explains precisely why.
When a model trains on its own outputs, it amplifies its existing statistical biases. There is no external ground truth to correct against, no novel signal to incorporate. The model begins to echo its own distributions until the "intelligence" becomes a caricature of itself a process researchers call model collapse. The training signal converges on the model's own blindspots, and those blindspots compound. You don't transcend the distribution by recursively sampling from within it.
The autoregressive literature calls this the "information retention bias" the model's KV cache retains unnecessary sequence history, leading to representations optimised for reconstruction rather than abstraction.[6] Models excel at interpolating between patterns they've seen. They structurally resist the kind of compression and abstraction that would allow genuinely novel inference. The Bottlenecked Transformer which addresses this through periodic transformation of the KV cache achieved up to 3.5× parameter efficiency gains by forcing the model toward predictive features rather than raw memorisation. The vanilla architecture actively resists this without surgical intervention.
Autoregressive training on "A→B" fails to induce recognition of "B←A." The asymmetry isn't incidental it reflects fundamental gradient flow dynamics in cross-entropy optimisation. A model trained that Lincoln was assassinated cannot reliably infer what event ended Lincoln's life. With vanilla training, this sits near random-guess probability. Mitigable, yes but only by explicitly patching a flaw the paradigm introduced.
What True Reasoning Would Require
AGI if it means anything must deal not in the probable but in the possible. Not "what usually comes next in this distribution" but "what has never been thought before but could be structurally true." That distinction is not a matter of scale. It is a matter of architecture and epistemology.
The research hints at what genuine reasoning looks like, and it isn't gradient descent on the next token. The most striking results in the literature came not from scaling, but from hybrid approaches systems that couple learned pattern recognition with explicit logical verification. The THOUGHT-LIKE-PRO framework, which verifies reasoning trajectories through a Prolog engine before translating them into natural language, achieved near-perfect accuracy on formal logic benchmarks: 98.19% on ProofWriter, 100% on PrOntoQA.[7] This is not imitation learning. This is a system where correctness is guaranteed by external ground truth, not inferred from statistical likelihood.
The difference is not subtle. A language model learns the form of logical argument. A symbolic verifier determines whether an argument is actually valid. These are categorically different operations and conflating them is precisely the confusion that has led Big Tech to mistake fluency for thought.
The abductive gap the ability to form a plausible hypothesis from genuinely incomplete information, to arrive at a belief that wasn't present in the training distribution remains unaddressed by the autoregressive paradigm. It is not a problem that more parameters solve. It is a problem that requires a fundamentally different epistemic architecture.
The Uncomfortable Conclusion
None of this is to say that LLMs aren't extraordinary technological achievements. They are. The ability to compress and surface patterns across the breadth of human knowledge is genuinely remarkable, and genuinely useful. For countless practical tasks, they represent a step-change in capability.
But useful is not general. Impressive is not intelligent. And pattern retrieval however sophisticated, however fast, however vast the pattern library is not reasoning.
The empirical record is clear: performance degrades on fresh data, collapses on truly novel reasoning tasks, and remains fundamentally bounded by the teacher's ceiling in imitation settings. Process-level instability causes exponential decay of decision advantage with execution length.[8] Long-horizon reasoning the kind AGI would require degrades systematically as the chain extends. These are not bugs to be patched. They are structural consequences of the autoregressive objective itself.
Big Tech is building the world's most perfect mirror and calling it a window. They have constructed a closed epistemic loop human knowledge in, human-pattern output, graded against human-generated tests and named the output of that loop "intelligence." The loop is elegant. It is enormously profitable. It is a scientific and engineering triumph.
And it will never be AGI. Because A=A is not a discovery. It is a definition. And no amount of compute will make a tautology into a theorem.
The Final Word
The path to genuine machine intelligence if it exists runs through a different kind of architecture: systems that can form hypotheses the training data doesn't contain, verify reasoning against ground truth that isn't itself statistical, and operate with epistemic humility about the boundaries of their own knowledge. Until then, we are watching the most expensive, most sophisticated, most convincingly human echo chamber in history and calling it the future of mind.
Academic Sources Referenced
- Bachmann, G. & Nagarajan, V. (2024). The pitfalls of next-token prediction. ICML. arXiv:2403.06963
- McCoy, R.T. et al. (2024). When a language model is optimized for reasoning, does it still show embers of autoregression? arXiv:2410.01792
- Mukherjee, S. et al. (2023). Orca: Progressive Learning from Complex Explanation Traces of GPT-4. arXiv:2306.02707
- Chi, H. et al. (2025). Unveiling Causal Reasoning in Large Language Models: Reality or Mirage? NeurIPS. arXiv:2506.21215
- Wang, B. et al. (2024). Grokked Transformers are Implicit Reasoners. arXiv:2405.15071
- Oomerjee, A. et al. (2025). Bottlenecked Transformers: Periodic KV Cache Abstraction for Generalised Reasoning. arXiv:2505.16950
- Tan, X. et al. (2024). Thought-Like-Pro: Enhancing Reasoning via Self-Driven Prolog-based Chain-of-Thought. arXiv:2407.14562
- Liao, H-J. (2026). Intrinsic Stability Limits of Autoregressive Reasoning: Structural Consequences for Long-Horizon Execution.