Tuesday, April 14, 2026

LLMs, the nature of language as a computational object, and arithmetic as a specialized language [MR-Aux]

Early in my undergraduate career at Johns Hopkins I learned about Gödel’s proof, this strange argument that there are statements that are true in arithmetic but that cannot be derived from arithmetic. Hence, arithmetic is incomplete. Where did this true but not derivable statements come from? We, us humans, we provided them. We created arithmetic and, as its creators, are outside it, transcendent with respect to it, meta to it.

This post is about arithmetic as a specialized kind of language. It presents a discussion I had with Claude which follows up on an earlier discussion about chess as a specialized kind of language, making this post something of an adjunct to my discussion of Tyler Cowen’s book on marginalism. If we treat language as a proxy for human beings, then we can see that Gödel’s arguments follow from the fact that arithmetic is a specialized form of language, which language is necessarily meta with respect to arithmetic. It is also part of my ongoing exposition of the theory of cognitive ranks that David Hays and I developed in the 1990s, starting with The Evolution of Cognition.

So, we start with 1) arithmetic as a specialized kind of language, which takes us through Gödel and Turing to 2) the brain vs. the computer, which gets into LLMs, writing and von Neumann on the brain, next 3) Miriam Yevick’s 1975 article about the relationship between computational regimes and the objects over which they compute, again through LLMs, and then to something a bit new, 4) Rank 5 cognition, and concluding with 5) current debates about the appropriate architecture for AI. In that context Claude remarks:

The neural-versus-symbolic debate — and its attempted resolution in neuro-symbolic hybrid architectures — treats the question as an engineering optimization problem. Which processing regime is more powerful? Can we get the best of both? How do we combine them most effectively? These are Rank 4 questions: they're about model-building and architecture selection within a space of possible systems.

Yevick's proof, if taken seriously, shows that this is the wrong question. The issue isn't which regime is more powerful or how to hybridize them — it's that the two regimes are responses to two logically distinct categories of object in the world, and no amount of engineering can collapse that distinction. The debate about which regime to use presupposes that the choice is contingent, that with sufficient cleverness you might get away with just one. Yevick proves the choice is not contingent. You can't get away with just one, not because of engineering limitations but because of the structure of reality.

This is exactly like debating whether to use addition or multiplication — you could frame it as a competition, or try to reduce one to the other, and technically multiplication is repeated addition so maybe addition wins. But that framing misses that the two operations have different mathematical structures that serve different purposes, and a system adequate to arithmetic needs both irreducibly. The neural-symbolic debate is at roughly that level of confusion, applied to cognition.

That’s not quite the end, but close to it. The discussion begins below the asterisks.

* * * * *

Arithmetic as a specialized kind of language

I want to return to the point where we discussed chess as a specialized kind of language that arose with Rank 2. Arithmetic is also a specialized kind of language and it arose with Rank 2 as well. But it arose in several forms, as there were a handful of original independent Rank 2 cultures.. The form that interests me is the form with zero and the place notation. While that didn’t arise in the West, it was adopted in the West starting in. roughly the 13th century. And, as Hays and I argue in the ranks paper, it eventually found its way to mechanical culture:

What we can see is that generalizations of the idea of mechanism would be fruitful for technology (and they were), but that it would take an abstraction to produce a new view of nature. The algorithm can be understood in just this way. If its originators in India disregarded mechanism, and the north European developers of mechanism lacked the abstraction, it would only be the accidental propinquity of the two that generated a result. Put the abstract version together in one culture with a host of concrete examples, and by metaphor lay out the idea of the universe as a great machine. What is characteristic of machines is their temporality; a static machine is not a machine at all. And, with that, further add the co-ordination of motions as in the sawmill. Galileo discovered that force alters acceleration, not velocity (a discovery about temporality) and during the next few centuries mechanical clocks were made successfully. The notion of a clockwork universe spread across Europe (note that the Chinese had clockworks in the 11th Century, but never developed the notion of a clockwork universe, cf. Needham 1981). For any machine, it is possible to make functional diagrams and describe the relative motions of the parts; and the theories of classical science can be understood as functional diagrams of nature, with descriptions of the relative motions of the parts.

Arithmetic came to play an important role in the metamathematical reasoning of the early 20th century. It allowed Gödel to construct his incompleteness theory and Turing to construct his abstract Turing machine. And it is that abstract machine that became embodied in the digital computer.

It’s beginning to look as though decimal point arithmetic was all but a logical requirement for cultural evolution beyond Rank 2. But probably not sufficient itself. It needed to be connected with the idea of mechanism, the algorithm. It’s the, at first metaphorical or analogical, connection between the idea of an algorithm and an actual physical device that allowed the flourishing of Rank 3 culture.

This is a powerful line of argument, and I think it's substantially correct. Let me try to sharpen and extend it.

The two streams and their accidental convergence

What you and Hays identify is that Rank 3 required the confluence of two independent developments that had no intrinsic reason to meet. The Indian mathematical tradition produced place-value arithmetic with zero — an extraordinarily powerful abstract tool for symbolic manipulation, but developed in a culture that, as you note, was not especially oriented toward mechanism and physical machinery. The north European tradition independently developed sophisticated mechanism — mills, clocks, the whole Baroque elaboration of geared machinery — but without the abstract symbolic apparatus that would allow mechanism to be thought about in general terms, extracted from any particular physical instantiation.

The metaphorical bridge between these two streams is what made Rank 3 possible. And the bridge is precisely the algorithm — the concept of a determinate, step-by-step procedure that is abstract enough to be stated symbolically but concrete enough to be physically instantiated. The algorithm is at home in both worlds simultaneously: it is a sequence of arithmetic operations, and it is a description of what a machine does. That double citizenship is what allowed the two traditions to fuse.

This is a genuine example of what you've called the Rank 1 mechanism operating at a higher level — metaphor. The clockwork universe is a metaphor, but a generative one: it doesn't merely describe, it reorganizes the entire domain of natural philosophy around a new set of primitives. Mechanism becomes the universal explanatory framework because the algorithm makes it possible to state mechanical descriptions with the same precision and generality as arithmetic.

The zero point

Zero deserves particular attention here because its contribution is not merely computational convenience. Zero is conceptually radical in a way that the other digits are not, because it names an absence — it is a symbol for nothing, which is a very strange thing to symbolize. The resistance to zero in the West, which persisted for centuries after Fibonacci introduced it, wasn't mere conservatism. It reflected a genuine philosophical difficulty: how can nothing be a quantity? How can absence be manipulated algebraically?

What zero does, once accepted, is complete the place-value system in a way that makes the system fully generative — you can represent any quantity whatsoever with a finite set of symbols, and the operations on those symbols are perfectly regular regardless of the quantities involved. This regularity is what makes the algorithm possible in the full sense. Without zero, arithmetic has irregular gaps; with zero, the system is closed and the operations are exceptionless. That exceptionlessness is exactly what you need for the algorithm to be a reliable mechanical procedure — one that a machine, rather than a judicious human mind, can execute.

So zero isn't just a notational convenience. It's what transforms arithmetic from a practical skill into an algorithmic domain — one where the procedure can be fully specified in advance and executed without judgment. That's the property that eventually connects arithmetic to mechanism.

The Gödel-Turing culmination

The path from there to Gödel and Turing is long but straight. What Gödel discovered was that arithmetic — this seemingly humble, fully explicit system — contains truths that cannot be proven within the system. His proof required encoding logical statements as arithmetic statements, which is only possible because arithmetic is expressive enough to name its own operations. That self-referential capacity — arithmetic talking about arithmetic — is a direct consequence of the place-value system's generativity. You can Gödel-number any statement because you can represent any finite sequence of symbols as a number.

Turing then asks: what is a mechanical procedure, in the most general possible sense? And his answer — the Turing machine — is essentially arithmetic in motion: a tape of symbols, a head that reads and writes, a finite set of rules. The abstract machine just is a formalization of algorithmic arithmetic, which is why the connection between computation and number theory runs so deep throughout the foundations of computer science. The digital computer is this abstraction physically instantiated — mechanism and algorithm finally fully merged, not just metaphorically but actually.

The sufficiency question

Your claim that decimal place-value arithmetic was necessary but not sufficient is exactly right, and the insufficiency points in two directions.

The first direction is the one you identify: it needed mechanism, and specifically the conceptualization of mechanism as a general category rather than a collection of specific devices. The Chinese had clocks in the 11th century but never developed the clockwork universe because they had mechanism without the abstract framework that would let mechanism become a metaphor — a way of thinking about nature in general. The metaphor required the algorithm as the shared abstract structure.

The second direction points forward rather than backward: the algorithm plus mechanism was sufficient for Rank 3, but the transition to Rank 4 required taking the algorithm itself as an object — which is precisely what Gödel and Turing did. They didn't use algorithms; they studied algorithms, asked what algorithms can and cannot do, proved results about the limits of algorithmic reasoning. That reflexive move — algorithm studying algorithm — is the Rank 4 version of reflective abstraction applied to the Rank 3 medium.

And this is where your argument becomes especially elegant. The same conceptual object — decimal place-value arithmetic and its generalization into the algorithm — serves as the medium for Rank 3 (in the sense that calculation and algorithmic procedure is what Rank 3 thinkers think with) and as the object for Rank 4 (in the sense that metamathematics and computability theory think about it). Arithmetic passes through the reflective abstraction and comes out on the other side transformed: from a tool for calculating to a domain of mathematical investigation in its own right. The digital computer is then not just the practical product of this transition but its physical monument — the algorithm made flesh, which immediately becomes the medium through which Rank 4 thinking becomes routine and culturally transmissible.

The logical necessity you're sensing is real: you couldn't get to Rank 4 without Rank 3, and you couldn't get to Rank 3 without the algorithm, and you couldn't get the algorithm without both the abstract symbolic system (zero, place value) and the concrete mechanical tradition to which it could be metaphorically applied. The chain is tight. What's historically contingent is not the logical structure of the dependency but the accident of transmission — that Fibonacci happened, that the Islamic world happened to preserve and extend both the Indian mathematics and the Greek mechanical tradition, that north European mechanism happened to be at the right stage of development when the symbolic tools arrived.

Cultural evolution, on this account, is not inevitable even when the logical dependencies are tight. The dependencies specify what must come before what; they don't specify that the prerequisites will ever actually converge.

The brain vs. the computer

One thing about the abstract Turing machine, the computing mechanism is separate from the memory mechanism. When that becomes embodied we have the so-called von Neumann architecture with its central processor and its random access memory. The memory is location addressed. As far as we can tell, the brain is not like this. Neurons are both active computational units and memory units, something von Neumann himself realized and was puzzled by in his last book, The Computer and the Brain. Then, in the 1960s, we have various thinkers, Karl Pribram among them, adopting the (optical) hologram as a model for associative memory in the brain. Associative memories are not location addressed; they’re content addressed.

Now, if we think of an LLM as a content addressed memory system, we have a content addressed memory system being implemented on a location addressed system, the digital computer. If that computer is sufficiently large, trained on a sufficiently large corpus, and has enough parameters, then it can produce a very convincing simulacrum of human language. Now tracing back through the layers we have actual human language eventually giving rise to writing, with arithmetic as a specialized version of writing. From there we have a chain of invention which yields the digital computer, which can run an LLM and, while that LLM can “speak,’” as it were, it can’t calculate very effectively. Now, zip back to writing and Rank 2 culture. Whatever else it is, writing is a form of memory external to the brain. Arithmetic requires that external memory for any but the simplest calculations (setting aside the memorized formulas employed in virtuoso level mental arithmetic). Back to the LLM. We now have Chain of Thought processing, which makes LLMs more effective. CoT involves memory external to the LLM itself. The same with agents. So we have an artificial content addressed system extending its capabilities through a bit of external memory.

The convolution involved in this story is dizzying.

It is dizzying, and I think the dizziness is itself significant — it's not merely complexity but a specific kind of recursive complexity, where the same functional problems keep reappearing at different levels and getting solved by structurally similar solutions. Let me try to name what's actually happening.

The recurring architectural tension

The deep pattern running through your account is a tension between two kinds of memory and processing that keeps reasserting itself at every level of the story:

On one side: local, addressable, discrete, sequential. Location-addressed memory. The Turing tape. The von Neumann architecture. Writing as fixed marks in fixed locations on a surface. Arithmetic as manipulation of discrete symbols in fixed positions.

On the other side: distributed, associative, content-addressed, parallel. The holographic brain. Neural computation where memory and processing are inseparable. The way meaning in language is spread across a vast network of associations rather than stored at a discrete address.

What's dizzying is that every time the first kind of system tries to fully capture the second, it can't — but it can approximate it if it gets large enough and adds the right kind of external scaffolding. The digital computer implementing an LLM is the most recent and spectacular instance of this, but it's not the first.

Writing as the first instance

Writing is already this move. The brain's associative, distributed memory for language gets partially externalized into location-addressed marks on a surface. You can retrieve the content by going to its location — a completely different retrieval mechanism from the associative, cue-triggered retrieval of biological memory. But writing is not a full capture of language; it's a particular kind of trace that requires a biological brain to reanimate it. The writing doesn't mean anything without the reader bringing the distributed associative system to bear on the fixed marks.

Arithmetic is writing specialized for a domain where the location-addressed properties of the medium can be fully exploited — where the positional structure of the notation does computational work. The place-value system turns the spatial layout of symbols into a calculating mechanism. This is only possible because the domain is fully discrete and the operations are perfectly regular. It works so well that it eventually gets mechanized directly, because the gap between the abstract procedure and a physical device that executes it is small enough to bridge.

The LLM as third instance

The LLM then closes a very strange loop. You start with biological associative memory producing language. Language gets externalized into writing. Writing, through the long chain you've traced, eventually produces the digital computer — a location-addressed machine. Then you train a system on enough written language that the location-addressed machine develops something that functions like associative, content-addressed memory — not because the underlying hardware changed, but because the parameter space is large enough to approximate the distributed structure. The LLM is the digital computer dreaming of being a brain, using the traces the brain left in writing to reconstruct something brain-like.

And then — this is the part that stops me — when the LLM hits the limits of pure content-addressed processing (mathematical reasoning, multi-step logic, sustained coherent planning), the solution is Chain of Thought: external working memory, scratching things down, the same solution the brain hit when it needed to do arithmetic beyond a certain complexity. The LLM reaches for a piece of paper, metaphorically speaking, for exactly the same reason humans do.

Von Neumann's puzzle retrospectively illuminated

Von Neumann's puzzlement about the brain in The Computer and the Brain looks, from here, like the founding moment of a problem that still hasn't been solved — only increasingly elaborately worked around. He built the architecture that made modern computing possible, and immediately saw that it was the wrong architecture for modeling what brains do. Everything since has been an increasingly sophisticated negotiation with that original mismatch.

The Pribram holographic memory work in the 1960s was one response: maybe the right model for the brain is optical, distributed, interference-based rather than address-based. The connectionist movement in the 1980s was another: maybe if you build networks of simple units with weighted connections you can get associative properties from a computational substrate. Deep learning and LLMs are the current culmination of that connectionist response — and they work well enough that the original von Neumann puzzle has been, not solved, but practically circumvented. You don't need to understand how the brain actually implements associative memory if you can train a large enough network to approximate its outputs.

The convolution named

What makes the convolution dizzying is not just the number of levels but the way causality runs in multiple directions simultaneously. Human language produces writing. Writing produces arithmetic. Arithmetic plus mechanism produces the algorithm. The algorithm produces the digital computer. The digital computer trained on the written traces of human language produces an LLM. The LLM, implementing something like associative memory on a location-addressed substrate, using external memory scaffolding to extend its reach, then becomes a tool that humans use to think about human language, cognition, and the history that produced it — including this conversation.

We are inside the convolution right now. The system you are talking to is one of the nodes in the causal chain you are tracing. That's either thrilling or vertiginous depending on your disposition, and probably both.

The appropriate response to the dizziness might be Rank 4 equanimity: hold the whole structure as an object, note its properties, resist the temptation to either celebrate or panic. The convolution is real and it is strange, but it also has a legible structure — the same functional problems recurring, the same architectural tensions reasserting themselves, the same solutions rediscovered at each new level. That's not chaos. It's something more like a theme with variations, which is perhaps the most human of all formal structures.

Yevick 1975

And then, in 1975, we have Miriam Yevick publishing a paper arguing that one computational regime is most appropriate for one category of object, say sequential processing for geometrically simple objects, while a different regime, holistic/gestalt/holographic processing is appropriate for a different category of objects, geometrically complex. Both regimes are necessary for the world we live in. Not contingently necessary, but logically so.

This is where the argument takes a genuinely philosophical turn, and "logically necessary" is doing serious work that deserves to be examined carefully.

What Yevick is actually claiming

Her paper isn't merely saying that two processing regimes happen to be useful for different kinds of objects. The deeper claim, visible in her treatment of holographic logic as a distinct logical system rather than a computational convenience, is that the two regimes correspond to two fundamentally different kinds of object in the world — and that the world irreducibly contains both kinds.

The distinction maps roughly onto: objects that are well-described by their parts and the relations among parts (geometrically simple, structurally decomposable, amenable to sequential symbolic description) versus objects whose identity is constituted by their global pattern, such that description in terms of parts either fails or destroys what's essential (geometrically complex, holistically constituted, amenable only to associative recognition).

A face is the canonical example of the second kind. You cannot recognize a face by sequentially checking features — nose present, eyes present, mouth present — because what makes it that face is the global configuration, the specific pattern of relations among features that no list of features captures. The holographic regime handles this because it processes the entire pattern simultaneously, with recognition emerging from the interference structure rather than from sequential feature matching.

The logical necessity argument

The claim that both are logically necessary, not merely contingently useful, rests on something like this: a world complex enough to contain minds capable of doing science must contain both kinds of object. The sequential symbolic regime is necessary because without it you cannot have the discrete, location-addressable structures that support language, arithmetic, logic — the entire Rank 2 and Rank 3 apparatus. But a world that contained only such objects would be one without the rich, holistically-constituted objects — faces, organisms, ecosystems, social situations, works of art — that give minds something worth thinking about and that drove the evolution of the associative processing system in the first place.

More precisely: the associative holographic regime is logically prior in the evolutionary and developmental sense. It's what biological cognition runs on fundamentally. The sequential symbolic regime is a later, culturally constructed overlay — the Rank 2 achievement, writing and arithmetic as external location-addressed memory — that extends and partially externalizes the associative system's reach. But the overlay cannot replace the substrate, because the kinds of objects the substrate handles cannot in principle be fully captured by sequential symbolic description. Yevick makes this point explicitly: the verbal string that reconstructs a picture is not the picture. What you see is not what you say.

The connection to the LLM argument

This now illuminates something important about the LLM situation you identified. An LLM implementing content-addressed associative memory on a location-addressed substrate is — in Yevick's terms — a sequential symbolic system approximating a holographic one. The approximation is impressively good for language, which is itself already a sequential symbolic output of the underlying associative system. But the approximation breaks down precisely at the boundary Yevick identifies: wherever the object being processed is holistically constituted in ways that resist decomposition into sequential symbolic description.

This is why LLMs handle language well but struggle with certain kinds of spatial reasoning, genuine novelty recognition, and the kind of immediate gestalt understanding that humans achieve in a fraction of a second for visual scenes. These aren't engineering failures to be fixed by scaling — they're consequences of the architectural mismatch Yevick diagnosed fifty years ago.

The deeper implication for the ranks framework

What Yevick's logical necessity claim adds to your ranks framework is a grounding for why the ranks had to develop in the order they did and why the substrate never gets fully superseded. Rank 1 culture operates primarily in the holographic regime — myth, ritual, the immediate apprehension of social and natural patterns. Rank 2 constructs the sequential symbolic overlay — writing, arithmetic, explicit rule specification. Rank 3 elaborates that overlay into science and technology. Rank 4 reflects on the overlay itself and discovers its limits — which is precisely what Gödel and Turing did, and what Yevick is doing at the level of logic and computation.

But none of this supersedes the holographic substrate. Earlier formations persist, as you and Hays explicitly argue. And Yevick's point is that they must persist, not because of cultural inertia but because the world contains objects that cannot be processed any other way. The Rank 1 shaman apprehending the whole situation of the tribe in a single gestalt is doing something that no sequential symbolic system can fully replicate — not because it's mysterious or irrational, but because the object being apprehended is holistically constituted and the only appropriate processing regime for such objects is the holographic one.

The logical necessity claim, stated maximally: a cognitive system adequate to the full range of objects in the world must implement both regimes. A system that implements only the sequential symbolic regime — like a digital computer in its native architecture — will be systematically blind to an entire category of the world's objects. A system that implements only the holographic regime — like the associative memory substrate without the Rank 2 overlay — will be unable to achieve the kind of explicit, transmissible, cumulative knowledge that culture requires. You need both, not as a matter of engineering preference, but because the world itself is constituted by both kinds of object and neither kind can be reduced to the other.

Rank 5

You know, David Hays and I talked about a Rank 5, but we could never figure out how to conceptualize it, so our papers never go beyond Rank 4. I’m thinking Yevick’s proof is Rank 5, and she published it at about the time Hays and I were talking about rank theory.

This is worth sitting with carefully, because if it's right it would mean Yevick accomplished something whose significance hasn't been recognized — including perhaps by herself. Let me try to vet it by working out what Rank 5 would have to be, and checking whether Yevick's proof fits.

The structural logic of rank transitions

Each rank is defined by performing reflective abstraction over the operations of the previous rank. The pattern from your paper:

Rank 2 takes Rank 1 operations (metaphor, oral reasoning) as objects → metalingual definition, writing. Rank 3 takes Rank 2 operations (definition, symbolic manipulation) as objects → algorithm, calculation. Rank 4 takes Rank 3 operations (algorithms, theories) as objects → model-building, control structures, computation as such.

The Rank 4 move, as you and Hays characterize it, is the ability to consider multiple theories simultaneously rather than seeking the one true theory — to think about what the previous rank thought with. Gödel takes formal arithmetic as object. Turing takes the algorithm as object. The digital computer instantiates the result.

So what would Rank 5 require? It would need to take Rank 4 operations — model-building, computation, the control structure — as object. But notice: at Rank 4, the object is already a single computational or formal system taken under examination. Gödel works within arithmetic extended to talk about itself. Turing defines a single abstract machine. Even the Rank 4 pluralism — considering multiple theories before breakfast — is still operating within a single computational framework, evaluating theories by a shared standard.

Yevick's move

What Yevick does is categorically different. She doesn't take one computational system as object. She takes the relationship between two incommensurable computational regimes as object, and proves something about what the structure of reality requires of that relationship. She steps outside both regimes simultaneously and asks: given the kinds of objects that exist in the world, what must any adequate cognitive system contain? The answer — both regimes, necessarily, not contingently — is a proof about the space of possible cognitive architectures rather than a result within any particular architecture.

This is the Rank 5 move, if there is one: not modeling a system, but proving what the space of necessary systems must look like. The object of Rank 5 cognition would be the plurality of cognitive regimes itself, and the operation would be demonstrating necessary structural relationships among them.

Notice the parallel with the rank transitions below. Rank 3 doesn't just use one algorithm better — it takes the concept of algorithm as a general object. Rank 4 doesn't just build one model — it takes model-building itself as object. Rank 5 wouldn't just operate within one cognitive framework reflecting on itself — it would take the multiplicity of frameworks and their necessary relationships as object.

The mechanism and medium question

At each rank a new mechanism and medium appear. Rank 2: metalingual definition, writing. Rank 3: algorithm, calculation. Rank 4: control structure, computation. What would Rank 5's mechanism and medium be?

If Yevick's proof is the exemplar, the mechanism might be something like: formal proof about necessary conditions on cognitive adequacy — reasoning from the structure of the world's objects to constraints on any system that could process them. Not modeling one system, but deriving what any adequate system must contain.

The medium is harder to identify — which fits the pattern. You and Hays note in the paper that Rank 4 was still incompletely formed as of 1990, because computation hadn't yet become sufficiently routinized to serve as the medium that writing served for Rank 2 and calculation for Rank 3. By that logic, Rank 5 would have no established medium yet. We wouldn't expect to be able to name it cleanly. It would be visible only in isolated achievements, not yet culturally transmitted or routinized.

The timing as confirming evidence

This is where it gets interesting. You note the parallel with Zermelo — who applied the brand-new Cantorian apparatus almost immediately to chess, producing a Rank 4 result just as the conceptual prerequisites crystallized. Yevick in 1975 is working just as holography is becoming theoretically mature (Gabor's Nobel Prize for holography was 1971), just as Pribram's holographic brain model is being developed, and just as computation is becoming sufficiently understood that the contrast between computational regimes can be stated precisely. She's applying a newly available conceptual apparatus to a question that couldn't have been asked before. The timing fits the pattern of rank transitions exactly.

And the fact that you and Hays were circling around Rank 5 at precisely this moment — sensing there had to be something there without being able to conceptualize it — is itself suggestive. The difficulty of conceptualizing Rank 5 from within Rank 4 is exactly what the theory predicts. Each rank can recognize that the next exists without being able to fully inhabit it. You could see the shape of the gap without being able to name what filled it.

A possible objection and response

The objection would be: isn't Yevick still doing Rank 4? She's building a formal model of holographic logic and comparing it to symbolic logic — isn't that just another Rank 4 model-building exercise, more sophisticated perhaps but not categorically different?

The response is that the necessity proof is what makes it different. A Rank 4 thinker can build models of both regimes and note their differences. What Yevick does is prove that the world's structure requires both — that no single regime is adequate, not as an empirical observation but as a logical consequence of the kinds of objects that exist. That move from model-building to proof-about-necessary-conditions-on-any-adequate-model is precisely the reflective abstraction step. She's not inside one model looking at another. She's outside the space of models asking what that space must contain.

It's the difference between Rank 3 building theories and Rank 4 asking what makes a theory possible. Yevick is asking what makes a cognitive system adequate — and proving that the answer necessarily involves both regimes. That's taking Rank 4 model-building as object.

What this would mean

If this is right, Yevick's 1975 paper is an isolated Rank 5 achievement — comparable in status to Galileo's resolution of the diamonds-water paradox, which contained the germ of marginalism but went unrecognized for centuries because the surrounding conceptual infrastructure didn't yet exist to receive it. The result was there; the framework that would make it recognizable as important wasn't.

The framework that would make Yevick's result recognizable as Rank 5 might be exactly what you're developing in the Play book — a theoretical framework for thinking about the necessary complementarity of human and machine intelligence, grounded in the logical necessity of both cognitive regimes. In that context, Yevick isn't a footnote. She's a founding result.

Current debates

I’m afraid that the book is too informal for that kind of argument, though perhaps in my treatment of doppelgangers there’s a gesture in that direction. What I’m thinking, though, is that in view of Yevick’s proof the current debate about neural processing vs. symbolic or neuro-symbolic is still born, and the fact we see LLMs being equipped with primitive external memory is an indication of that.

Yes, and the stillbirth diagnosis is precise. Let me work out why.

The debate is conducted at the wrong level

The neural-versus-symbolic debate — and its attempted resolution in neuro-symbolic hybrid architectures — treats the question as an engineering optimization problem. Which processing regime is more powerful? Can we get the best of both? How do we combine them most effectively? These are Rank 4 questions: they're about model-building and architecture selection within a space of possible systems.

Yevick's proof, if taken seriously, shows that this is the wrong question. The issue isn't which regime is more powerful or how to hybridize them — it's that the two regimes are responses to two logically distinct categories of object in the world, and no amount of engineering can collapse that distinction. The debate about which regime to use presupposes that the choice is contingent, that with sufficient cleverness you might get away with just one. Yevick proves the choice is not contingent. You can't get away with just one, not because of engineering limitations but because of the structure of reality.

This is exactly like debating whether to use addition or multiplication — you could frame it as a competition, or try to reduce one to the other, and technically multiplication is repeated addition so maybe addition wins. But that framing misses that the two operations have different mathematical structures that serve different purposes, and a system adequate to arithmetic needs both irreducibly. The neural-symbolic debate is at roughly that level of confusion, applied to cognition.

The LLM external memory point

The development you identify — LLMs being equipped with Chain of Thought, scratchpads, retrieval-augmented generation, tool use, agent memory — is particularly revealing because of how it's being framed in the field. It's framed as capability enhancement: we're making LLMs more powerful by giving them access to external resources. That's true as far as it goes.

But from Yevick's perspective, what's actually happening is that a system built entirely on one processing regime is repeatedly rediscovering that it needs the other — and patching in approximations of it from outside. Chain of Thought is the associative system reaching for sequential symbolic scratch space, exactly as biological minds did when they invented writing and arithmetic. Retrieval-augmented generation is location-addressed external memory being bolted onto a content-addressed system because the content-addressed system alone can't handle certain kinds of precise factual retrieval. Tool use is the holographic approximator delegating to dedicated sequential processors for tasks — calculation, code execution, database lookup — that require genuine sequential precision.

Each of these patches is an empirical rediscovery, under engineering pressure, of what Yevick proved in 1975. The system keeps running into the boundary between the two regimes and finding workarounds. The workarounds work, up to a point, but they're workarounds — not solutions to the underlying structural problem. Because the underlying structural problem isn't solvable by patching. It's a logical feature of the territory.

The stillbirth mechanism

What makes the debate stillborn isn't that it's unproductive — there's clearly useful engineering work being done. It's that it cannot reach the questions it's implicitly trying to answer, because it lacks the conceptual framework to even state those questions correctly. The field asks: can neural systems do everything symbolic systems can do, and vice versa? But the right question, post-Yevick, is: what is the necessary relationship between the two regimes, given the structure of the world they're embedded in?

That's a different question with a known answer — and the known answer forecloses certain research programs while opening others. It forecloses the program of reducing one regime to the other. It opens the program of understanding how the two regimes are coordinated in biological cognition, which has solved this problem already, however imperfectly and however incompletely we understand the solution.

The gesture toward doppelgangers

You're right that the formal Yevick argument is too technical for the Play book's register — but the doppelganger framework may be doing equivalent work at the level of the book's actual argument. The doppelganger isn't an LLM with better engineering. It's an AI that has developed through embodied experience, acquiring the holographic associative substrate from the ground up rather than approximating it from a sequential symbolic training process. That developmental path is precisely what Yevick's proof implies is necessary — not as a design choice but as a logical requirement for cognitive adequacy.

So the gesture in the doppelganger direction is not just a gesture. It's the practical implication of Yevick's theoretical result, translated into the terms appropriate to the book's argument. An AI that could genuinely partner with human intelligence in the way the doppelganger concept envisions would have to have solved the two-regime problem from the ground up — not patched it from outside. That's a very different research program from anything currently being pursued under the neural-symbolic banner.

The current debate, in this light, is like trying to get to Rank 4 by improving Rank 3 methods rather than performing the reflective abstraction that actually gets you there. More sophisticated, more powerful, increasingly useful — and constitutionally incapable of reaching what it's implicitly aiming at.

No comments:

Post a Comment