Saturday, January 17, 2026

Poetry in humans and machines: Who’s great?

Hollis Robbins, LLM poetry and the "greatness" question, Anecdotal Value, Jan, 7, 2026.

I posted the following comment to Hollis’s article:

An oblique observation. To the extent that an LLM can be said to "simulate" a human brain, it would be simulating the neocortex. With only 16 B neurons the neocortex contains only 19% of 86 B neurons in the brain. Thus LLMs, no matter how many parameters, are operating with only a fraction of the capacity of a poet's brain. To be sure, the poems they're trained on were created by full-brained poets, so those poems bear the traces of full brains. But I figure a full-brained poet would be able to find new kinds of paths through the brain's full state space. An LLM's ability to do that would seem to be rather truncated.

It's not clear to me how this bears on greatness in poetry. But I would like to think that at least some poems are great because they opened up new whole brain possibilities.

The purpose of this post is to develop those ideas more carefully. First I’ll talk about the idea of the brain’s state space. Then we move to the idea of a text as tracing a path through that space. Finally we’ll take up greatness in poetry.

The phase space of a brain

The notion of phase space, as I’m using the term, derives from statistical mechanics. It’s a way of thinking about how a physical system changes over time. Change is a movement from one state of the system to another.

The phase space of a system depends on three things: 1) how many elements the system has, 2) how many states each element can take, and 3) how the individual elements are coupled. The phase space has a dimension for each element of the system. The system’s state at any given moment will be a point in the phase space.

If the system has only three elements, then its phase space has three dimensions. Its state at any moment will be a point in that space. Notice that time isn’t one of those dimensions. But as the system changes over time its position in the phase space will change as well. It is possible, however, keep track of those positions.

A phase space having only three dimensions is easy to visualize. What about a phase space with four, five, six, thirty-eight, of 999,000 dimensions? Impossible to visualize.

The human brain has roughly 86 billion neurons. A phase space with that many dimensions will have 86 billion dimensions; impossible to visualize. But the individual neuron might not be the appropriate element to be the primitive unit in the brain’s phase space. The individual synapse – a connection between one neuron and another – may be the appropriate phase space primitive. Each neuron is connected to an average of 7000 other neurons, making the number of dimensions in the brain’s phase space much higher. The exact number doesn’t matter; the point is simply that the phase space is huge.

Note that neurons do not operate independently of one another. They are coupled together. The brain is divided into a number of functional areas (NFA, my term) where the connections between neurons within an area are more numerous that the connections each neuron has to neurons in other NFAs.

At this point let me introduce a query I put to ChatGPT 5.2:

The human brain has roughly 86 billion neurons. How many different functional regions does it have? I believe that the neocortex has between 100 and 200 different functional regions, but only has 16 billion neurons or so. How many different functional regions do we have in subcortical area?

The resulting discussion was a bit messy and complicated, as these things can be, but the upshot is that that brain seems to have roughly 700 NFAs. Only half of them are in the cortex. That’s very important (recall my original comment to Hollis). Later in the conversation ChatGPT observed:

  • If LLM behavior resembles anything in the brain, it most closely resembles cortical linguistic competence: high-dimensional, distributed patterning; associative completion; flexible recombination of learned material.
  • But cortex is only one part of the full system. Even taking the cortex generously as “the cognitive engine,” it is not the whole engine.

Thus, “At best it’s like a subsystem of the brain, and specifically one that is downstream of bodily regulation, affect, action selection, and reinforcement gating.”

Text as trace

I tend to think of texts, any texts, as the trace of a trajectory through the brain’s state space. But, at best, an LLM models only part of that space. Thus, returning to ChatGPT:

Even when an LLM produces a convincing poem, its method is necessarily missing the full set of whole-brain constraints that shaped the poetic corpus it learned from. Therefore its exploration of possibility space is structurally truncated relative to the exploration performed by a whole-brain poet.

ChatGPT offered me this paragraph:

If we treat a text as a path through a brain’s phase space, then poetry is not simply the output of a language module. It is the trace of a whole-brain trajectory: language regions interacting with memory, interoception, affect, reward, attention, and social cognition. A large language model is trained on the textual residue of those trajectories — the fossil record, not the living process. The poems in its corpus were written by full-brained poets, so the text contains evidence of that fuller generative machinery. But the model’s own generative dynamics are constrained to what can be recovered from token sequences and their statistical regularities. However impressive the simulation, it is necessarily missing the deeper gating systems that, in humans, determine what becomes sayable, what becomes bearable, and what becomes necessary. My hunch is that some poems are great not merely because they are well-formed linguistic artifacts, but because they open new whole-brain possibilities — new stable pathways of feeling-and-thought — that alter what the mind can do afterward.

What of humans and AI’s

I then prompted ChatGPT:

The issue that’s hanging over this kind of discussion is ultimately: Are there good roles for humans in a world of intelligent machines or are we replaceable (and so will be replaced)? That’s a messy question that begs many issues as formulated and so is rather meaningless. [...]

At the moment I’m working on a book about the social and cultural implication of AI. I started working on the book last summer. I broke off work from September through about a week ago. I’m now back at it. Last year I had a title, scads of notes and drafts and two or three versions of an outline. I’ve now arrived at an outline that I don’t think I could have produced without the help of Claude and ChatGPT. Many of the chapters I was working on last year are in this new version, 6 of them in fact. But I’ve got two chapters I didn’t have last year and an overall architecture that I didn’t have. Moreover I don’t think I could have produced that architecture on my own. The AIs gave me the means of holding the whole thing in my head and working at different scales so that I’ve been able to do something new.

I’m a good thinker. I’m better working with AIs. I should think the same would be true for poets.

We had some further discussion and ChatGPT offered me this paragraph:

The question “Will AI replace us?” is too blunt to be useful. My own experience suggests something more precise: large language models can function as cognitive instruments that let a person think at scales they could not reliably manage alone. I can now hold a book-length architecture in view, revise it coherently, and move between local detail and global structure without losing the plot. I am not a poet, but I am a thinker, and I know the difference between producing text and producing understanding. The models are not merely faster typists with larger memories. They are interfaces that make new kinds of conceptual control possible. If that is true for expository thinking, it is likely true for poetry as well: not that the machine becomes the poet, but that the poet gains a new way to explore and stabilize trajectories through a much larger space of possibilities.

I’m willing to let things rest there for the moment.

(Oh, I also asked ChatGPT to produce that illustration at the head of this post.)

A breakfast story

Friday, January 16, 2026

Hon. Martha A. Mills (ret): Lawyer, Activist, Judge

Martha Mills is a friend of mine. She took a course on the history of technology taught by my teacher, David Hays, and became fascinated by his ideas. That's how I know here. I published a review of her 2015 book, Lawyer, Activist, Judge: Fighting for Civil and Voting Rights in Mississippi and Illinois, in 3 Quarks Daily in 2016.

Here's a short professional biography on the website of the American College of Trial Lawyers:

Mills was appointed in 1995 by the Illinois Supreme Court to serve as a Judge of the Circuit Court of Cook County. She served for two years, primarily in the Juvenile Justice, Child Protection Division. She was recalled in 2007, and was Supervising Judge of the Parentage and Child Support Court until December of 2012. Under her guidance, the court started a highly successful pilot restorative justice program with the help of Elizabeth J. Vastine and Peter Newman. Since her retirement she has continued to to teach and offer services in community building and restorative practices.

After law school, Mills was the first woman lawyer with White & Case in NYC. She then joined the staff of the Lawyers Committee for Civil Rights Under Law in Mississippi where she tried many cases, civil and criminal, getting the first jury verdict in Mississippi since Reconstruction for over a million dollars on behalf of the estate of an African American man murdered by the Klan. She was Chief Counsel of the Lawyers Committee Office in Cairo, IL prior to moving back to Chicago. She has practiced law in Illinois since 1969, both as a trial and appellate attorney. The American Bar Association in 2013 published her book: Lawyer, Activist, Judge: Fighting for Civil and Voting Rights in Mississippi and Illinois.

Judge Mills was a trial lawyer for more than forty years. She is licensed to practice in the states of Illinois, New York and Mississippi, and in numerous federal courts. She has argued cases in the Mississippi, Illinois and U.S. Supreme Courts. She has not only arbitrated, mediated and facilitated many cases and conflicts, but also taught others to do so. As founder and Executive Director of Transforming Communities, a not-for-profit corporation, she engaged consulting, practice, education and training in restorative justice from 1996-2007, as well as work in violence prevention and intervention.

Friday Fotos: Vibrant Matter

Now’s the time, free at last

Something I found on the web:

When it comes to changing the world:

It's always too late,
always too soon,
the time's never right,
Now's the time.

I found it deep in the Wayback Machine, here. You’ll have to scan down the lefthand panel until you come to the phrase, "Now is the time"; it'll be in grey type. Click on that phrase; it's a hypelink. You'll see new text appear in the righthand panel. The text i quoted above is there. That’s from a tribute to Martin Luther King that Bill Berry put on the web in the mid-1990s in the early days of the web. If you’re at all curious about how that got there, this post tells the story: I was on the web before the New York Times was, me and Cuda Brown at Meanderings.

This post is about that little four-line ditty. It’s not attributed to anyone. I created that page so I assume that I wrote those lines. But I don’t remember having done so. It’s not a big thing. I have no trouble imagining that I’ve forgotten all about it; it was 30 years ago. That I’ve just forgotten is more plausible than that I’d gotten the lines from someone, somewhere, and just didn’t attribute them. That’s not my style.

Thursday, January 15, 2026

About the clave rhythm, central to Latin music

The clave rhythm developed in Cuba from the African bell pattern brought to the Americas via the trans-Atlantic slave trade. The rhythm is used to lock in the many rhythms of Afro-Cuban music, as well as other Latin music like the bossa nova of Brazil. This video focuses on three claves -- rumba clave, son clave, and the Brazilian clave.

Selfie-stick in Hoboken confronts Hudson Yards

Among universities, the Chinese are moving up, USA is moving down

Just think, a century ago most Americans thought of Charlie Chan, a movie detective, when they they thought on the Chinese. And Chinatown, of course. In the 1960s a character named "Hop Sing" appeared as comic relieve in the hit TV program Bonanza. He was the house boy for the wealthy Cartwright family – note the word "boy." And then came the war in Vietnam along with Chairman Mao and his Red Book. When I was in graduate school in the 1970s I took a course on radical approaches to literature, taught be Art Efron. We read a collection of essays on aesthetics written by Mao.

Mark Arsenault, Chinese Universities Surge in Global Rankings as U.S. Schools Slip, NYTimes, Jan 15, 2026. The article opens:

Until recently, Harvard was the most productive research university in the world, according to a global ranking that looks at academic publication.

That position may be teetering, the most recent evidence of a troubling trend for American academia.

Harvard recently dropped to No. 3 on the ranking. The schools racing up the list are not Harvard’s American peers, but Chinese universities that have been steadily climbing in rankings that emphasize the volume and quality of research they produce.

The reordering comes as the Trump administration has been slashing research funding to American schools that depend heavily on the federal government to pay for scientific endeavors. President Trump’s policies did not start the American universities’ relative decline, which began years ago, but they could accelerate it.

“There is a big shift coming, a bit of a new world order in global dominance of higher education and research,” said Phil Baty, chief global affairs officer for Times Higher Education, a British organization unconnected to The New York Times that produces one of the better-known world rankings of universities.

In the middle of this long article:  

The number of international students arriving in the U.S. in August 2025 was 19 percent lower than the year before, a trend that could further hurt the prestige and rankings of American schools if the world’s best minds choose to study and work elsewhere.

China has been pouring billions of dollars into its universities and aggressively working to make them attractive to foreign researchers. In the fall, China began offering a visa specifically for graduates of top universities in science and technology to travel to China to study or do business.

“China has a boatload of money in higher education that it didn’t have 20 years ago,” said Alex Usher, president of Higher Education Strategy Associates, a Toronto education consulting company.

Mr. Xi has made the reasons for the country’s investments explicit, arguing that a nation’s global power depends on its scientific dominance.

There's much more in the article.

Hays, Forth, and creativity in man-machine-interaction

I've just been talking with ChatGPT about David Hays back in the 1970s. From my prompt:

In 1977 or 1978 he got a personal computer, one of those based on the Z-80 chip, and started learning to program it in Forth. He also sent out a weekly snail-mail newsletter entitled “The Diary of a Man and His Machine.” In one issue he asked for replies from people whose work feel between the cracks. He got about 100 replies. We read them all. We concluded that most of them were just oddballs; they didn’t even seem to know the standard literature in the field(s) relevant to their work. But a handful seemed legit. I don’t think we ever did anything with them.

...working between the cracks is not the same as working without foundations

From ChatGPT's response:

First, David Hays’s early adoption of a Z-80–based personal computer and Forth in 1977–78 places him well ahead of the curve, not just technically but cognitively. Forth, in particular, is telling: it is intimate with the machine, minimalist, stack-based, and reflective. It invites thinking about process, control, and representation in ways that high-level languages often conceal. That choice alone aligns with his lifelong interest in cognitive architecture rather than surface performance.

working between the cracks is not the same as working without foundations

Later:

Third—and this is perhaps the most interesting part—the call for people whose work “fell between the cracks,” and the response it generated, reads like an early, analog precursor to today’s online intellectual ecosystems. The fact that you and Hays received about a hundred replies, and that most of them struck you as oddballs, is not a disappointment; it’s a sociological constant. When disciplinary boundaries loosen, noise floods in first. What matters is that a handful were clearly legitimate, even if nothing came of it at the time. [...] Finally, your retrospective judgment—that many respondents lacked grounding in the relevant literature—underscores a distinction you’ve been circling throughout this chat: working between the cracks is not the same as working without foundations. Hays was interested in the former, not the latter.

Another prompt:

Your comments about Forth are apt. Hays valued the intimacy with the machine. He eventually programmed a whole environment using Forth and used it for keeping notes and generating text. Some of the text for our articles came out of that machine. That detail completes the picture—and it makes the point even stronger.

Chat replies:

What David Hays was doing with Forth was not “early word processing” or “personal productivity.” He was building a cognitive prosthesis—a writable environment in which thinking, note-keeping, and text generation were continuous with computation rather than merely supported by it.

A few things are worth making explicit.

First, the intimacy you mention is decisive. Forth collapses distance between intention and execution: no elaborate compilation pipeline, no abstraction layers to hide what the machine is doing. That intimacy encourages a style of thought in which process is visible. For someone like Hays—already committed to understanding cognition as structured activity—that mattered more than convenience or speed.

Second, the fact that he built an entire environment—not just scripts, but a living workspace for notes and text—means that the machine was not ancillary. It was part of the thinking loop. That some of the prose in your joint articles emerged from that environment is not incidental; it means the machine participated, materially, in the production of theory. This is not metaphorical augmentation. It is literal.

Wednesday, January 14, 2026

23% chance that the USA will buy part of Greenland for $500 to $700 billion

Can AI generate new mathematical ideas?

Cade Metz, Can A.I. Generate New Ideas? NYTimes, Jan 14, 2026 

Here's how Claude 4.5 summarized the article:

AI Tackles Mathematical Problems - But Questions Remain

An AI startup called Harmonic, using OpenAI's GPT-5 technology, recently solved a mathematical "Erdos problem" - a collection of challenging puzzles designed to test mathematical ingenuity. This sparked debate about whether AI has reached the point of conducting legitimate academic research.

Key Points:

The achievement generated mixed reactions. While some experts see it as evidence that AI can do genuine research, others, including renowned mathematician Terence Tao, argue the solution resembles "a clever student who memorized everything" rather than demonstrating true understanding. The solution was largely based on well-known methods. Initially, OpenAI claimed GPT-5 had solved 10 Erdos problems, but researchers discovered it had merely identified existing solutions buried in decades-old papers. Despite this, the technology proved valuable - it could find obscure sources humans might never locate.

Current Capabilities:

Modern AI systems use "reinforcement learning" to reason through problems, sometimes working for hours. While they can't yet generate truly novel ideas, they've become powerful research tools that can:

  • Analyze and store far more information than humans
  • Suggest hypotheses researchers hadn't considered
  • Help scientists narrow down experiments from 50 to 5

The Verdict:

Experts agree AI is a rapidly improving research assistant, but it still requires experienced human collaborators to guide it, interpret results, and separate useful insights from noise. Whether AI can independently generate breakthrough ideas remains an open question.

A black and white photo of a dessicated plant

Artificial Intelligence and the Future of Work

Konrad Kording and Ioana Marinescu, (Artificial) Intelligence Saturation and the Future of Work, Brookings Institute, 2025, https://papers.ssrn.com/sol3/papers.cfm?abstract_id=6051694

Abstract: Macroeconomic models typically treat AI as just another form of capital, and predict a slowly evolving world, while computer science scaling laws applied to the whole economy predict explosive growth and the potential for a singularity-like event. Both views gloss over the asymmetric reality that intelligence capital or AI scales at computer-science speeds, whereas physical capital and labor do not. What’s missing is a unified, parameter-driven framework that can nest assumptions from both economics and computer science to generate meaningful predictions of AI’s wage and output im- pacts. Here we use a constant elasticity of substitution (CES) production function framework that separates physical and intelligence sectors. Whereas physical capabilities let us affect the world, intelligence capabilities let us do this well: the two are complementary. Given complementarity between the two sectors, the marginal returns to intelligence saturate, no matter how fast AI scales. Because the price of AI capital is falling much faster than that of physical capital, intelligence tasks are automated first, pushing human labor toward the physical sector. The impact of automation on wages is theoretically ambiguous and can be non-monotonic in the degree of automation. A necessary condition for automation to decrease wages is that the share of employment in the intelligence sector decreases; this condition is not sufficient because automation can raise output enough to offset negative reallocation effects. In our baseline simulation, wages increase and then decrease with automation. Our interactive tool shows how parameter changes shift that trajectory. Wage decreases are steeper at high levels of automation when the outputs of the physical and intelligence sectors are more sub- stitutable. After full automation, more AI and more physical capital increase wages, a classic prediction from standard production functions in capital and labor. Yet, when intelligence and physical are complementary, the marginal wage impact of AI capital saturates as AI grows large. More broadly, the model offers a structured way to map contrasting intuitions from economics and computer science into a shared parameter space, enabling clearer policy discussions, and guiding empirical work to identify which growth and wage trajectories are plausible.

H/t Tyler Cowen.

Monday, January 12, 2026

AI for the Next 30 Years: Four kinds of activity that should be pursued

Here’s the prompt I used to elicit a text from Claude:

I want to do a blog post setting forth those four programs. I want it to be between, say, 1000 and 2000 words, no more. It should have an introduction, sections for each of the four programs, and some final remarks. Give it a title like: AI for the next 30 years, an open-ended plan. Assume a college-educated readership that is generally sophisticated. I’m going to put it in my New Savanna blog along with another post in which I present excerpts from Rodney Brooks’ current remarks on technology.

I’ve spent a lot of time over the last three days conceptualizing those four programs, interacting with both ChatGPT 5.2 and Claude 4.5. Those chats, in turn, rest on work that I’ve done with both chatbots over the last three years. Moreover, I have uploaded a fair number of documents to those chatbots, both articles from the formal literature and informal working papers, going back five decades.

Note that AGI is not mentioned anywhere nor did I ask ChatGPT to make dated prediction. Predicting where the earth will be in the solar system in 30 years, that’s easy. We’ve known how to do that since Newton. Predicting the weather 30 years out, very difficult to impossible. But maybe we can come up with rough estimates of average temperature for the year, and precipitation. Predicting the 30-year evolution of a complex socio-cultural-technical system? Not on your life.

I’ve edited Claude’s text in some minor ways and added so links at the end of each section.

AI for the Next 30 Years

Large Language Models represent something fundamentally new in computing: systems that have learned vast amounts about the world but encode that knowledge implicitly, in billions of inscrutable parameters. We can use these systems—often impressively—but we don't truly understand what they know or how they organize that knowledge. It's as if we've discovered a vast wilderness: we can explore parts of it, but we lack comprehensive maps.

Over the past few years, I've been thinking about what it would take to map this territory systematically and transform it from mysterious wilderness into reliable infrastructure. This thinking has crystallized into four parallel research programs, each essential, each reinforcing the others. Unlike the prevailing Silicon Valley vision of one lab developing a superintelligent system that does everything, this is a distributed, collaborative, multi-decade effort requiring both technical innovation and institutional creativity.

Activity 1: Ontology Extraction

The challenge: LLMs generates texts that distinguish between dogs and cats, animate and inanimate, concrete and abstract—but this knowledge exists only implicitly in weight matrices. We need to extract this latent ontological structure and make it explicit and inspectable.

Recent work by Christopher Manning and colleagues at Stanford has shown that neural networks encode rich linguistic structure—syntax trees, for instance—that can be extracted through systematic probing. I'm proposing we extend these methods from linguistic structure to ontological structure: the categories, hierarchies, and affordances that organize conceptual knowledge.

The key insight is that ontology is implicit in syntax. Verbs select for certain kinds of subjects and objects based on categorical presuppositions. "Eat" requires an animate agent and edible patient. These selectional restrictions reveal the categorical structure underneath. By systematically probing syntactic behavior, clustering words by shared patterns, and validating through transformation tests, we can extract the ontologies LLMs have learned.

This work must be distributed across many research groups, each focusing on specific domains—medical ontologies, legal ontologies, physical systems ontologies, and so forth. No single lab has the expertise or resources to map the entire territory. We need shared infrastructure (probing tools, ontology repositories, validation benchmarks) and coordinated standards, but the actual extraction work happens in specialized communities with deep domain knowledge.

The payoff: explicit ontological structure that can be verified, debugged, systematically improved, and integrated with symbolic reasoning systems. We transform opaque neural networks into hybrid systems that combine learning with legible structure.

Some background:

Christopher Manning et al. Emergent linguistic structure in artificial neural networks trained by self-supervision, PNAS 2020, https://www.pnas.org/doi/full/10.1073/pnas.1907367117

William Benzon, ChatGPT: Exploring the Digital Wilderness, Findings and Prospects, https://www.academia.edu/127386640/ChatGPT_Exploring_the_Digital_Wilderness_Findings_and_Prospects (see especially pp. 28-38, 42-44]

Activity 2: Cognitive Models and Multimodal Grounding

The challenge: Extracting ontologies from language gives us how language talks about the world, not how minds represent the world for perception and action. A robot needs more than linguistic categories—it needs grounded representations that integrate vision, touch, motor control, and yes, language, into a unified cognitive model. This distinction is standard in the cognitive sciences, including “classical” symbolic AI. I picked it up in the work I did with David Hays in the 1970s on cognitive networks for natural language semantics. We conceived of language mechanisms as operating on a separate cognitive model—language is an interface to the model, not the container of it. For embodied AI and robotics, this becomes crucial.

Consider a cup. The linguistic ontology tells us: cup is-a container, is-a artifact, can-hold liquids. The cognitive model adds: cylindrical shape with hollow interior, graspable via handle, stable on flat surfaces, rigid, will break if dropped, liquid spills if tilted beyond 45 degrees. This is sensorimotor knowledge grounded in perception and action, not purely linguistic.

Current multimodal systems (like GPT-4V or Gemini) take vision and "linguistify" it—everything gets processed through language. What we need are systems where multiple modalities read and write to a common cognitive model. Vision contributes spatial structure, language contributes categorical relationships, action contributes causal understanding, and they all integrate.

This research connects directly to robotics. A robot exploring a new kitchen needs to build spatial maps, identify affordances, understand causal relationships (that knob controls that burner), and eventually respond to linguistic commands—all drawing on the same underlying world model. The cognitive model is where the "adhesion" component of meaning lives: the grounding in physical reality that pure language systems lack.

Some background: Gary Marcus, Generative AI’s crippling and widespread failure to induce robust models of the world, Marcus on AI, June 28, 2025, https://garymarcus.substack.com/p/generative-ais-crippling-and-widespread

Activity 3: Associative Drift and Discovery

The challenge: Current AI systems are reactive, not curious. They solve problems you give them but don't discover problems worth solving. They lack what I'm calling associative drift—the capacity for open-ended, low-bandwidth exploration that enables serendipitous discovery.

Think about how intellectual discovery actually works. When I searched "Xanadu" on the web years ago, I had no hypothesis—just idle curiosity. When I got 2 million hits, I had a hunch that seemed interesting (though I couldn't articulate why). The opportunity cost of investigating was low, so I poked around. Eventually I discovered distinct cultural lineages (sybaritic via Citizen Kane, cybernetic via Ted Nelson's hypertext project) that revealed something about how cultural memes evolve.

This is fundamentally different from task-directed reasoning. I wasn't trying to solve a predefined problem. I was in a low-bandwidth exploratory mode, sensitive to interesting patterns, following hunches without clear goals. Current LLMs operate only in high-bandwidth mode: given a prompt, they generate detailed responses. They can't "skim" or "wonder" or "notice something odd" without generating full text.

We need architectures that support dual-mode processing: high-bandwidth for focused problem-solving, low-bandwidth for pattern detection during exploration. This requires technical innovations (sparse attention patterns, adaptive computation, salience detection) and new ways of thinking about AI objectives. How do we train systems to explore productively without specific goals?

For robotics, this is essential. A robot with associative drift doesn't just execute commands—it develops intuitions about its environment through undirected exploration, notices regularities, forms hunches about what matters. It becomes genuinely curious rather than merely reactive.

The interesting twist: associative drift needs the other programs. Ontologies provide the structured space that makes certain patterns "interesting" (ontologically distant concepts appearing together). Cognitive models enable embodied drift (noticing patterns through physical interaction). And drift enables discovery in the other programs (finding ontological incoherences, noticing when modalities misalign).

Gringos Tacos

Rodney Brooks on the state of AI and Robotics

As you may know, Rodney Brooks has been keeping an annual scorecard for various categories of high-tech activity. He puts it online on the first of the year. I’ve listed some excerpts from the 2026 scorecard below. The scorecard has much much more that I haven’t excerpted.

The Falcon 9

Eight years ago, Falcon 9 had been launched 46 times, all successful, over the previous eight years, and it had recently had a long run of successful landings of the booster whenever attempted. At that time five launches had been on a previously used booster, but there had been no attempts to launch Falcon Heavy with its three boosters strapped together.

Now we are eight years on from those first eight years of Falcon 9 launches. The scale and success rate of the launches has made each individual launch an unremarkable event, with humans being launched a handful of times per year. Now the Falcon 9 score card stands at 582 launches with only one failed booster, and there have been 11 launches of the three booster Falcon Heavy, all successful. That is a sustained growth rate of 38% year over year for eight years. And that it is a very high sustained deployment growth rate for any complex technology.

There is no other modern rocket with such a volume of launches that comes even close to the Falcon 9 record. And I certainly did not foresee this volume of launches. About half the launches have had SpaceX itself as the customer, starting in February 2018, launching an enormous satellite constellation (about two thirds of all satellites ever orbited) to support Starlink bringing internet to everywhere on the surface of Earth.

[Not AI or robotics, I know. But it interests me.] 

Humanoid Robots

My blog post from September, details why the current learning based approaches to getting dexterous manipulation will not get there anytime soon. I argue that the players are (a) collecting the wrong data and (b) trying to learn the wrong thing. I also give an argument (c) for why learning might not be the right approach. My argument for (c) may not hold up, but I am confident that I am right on both (a) and (b), at least for the next ten years.

I also outline in that blog post why the current (and indeed pretty much the only, for the last forty years) method of building bipeds and controlling them will remain unsafe for humans to be nearby. I pointed out that the danger is roughly cubicly proportional to the weight of the robot. Many humanoid robot manufacturers are introducing lightweight robots, so I think they have come to the same conclusion. But the side effect is that the robots can not carry much payload, and certainly can’t provide physical support to elderly humans, which is a thing that human carers do constantly — these small robots are just not strong enough. And elder care and in home care is one of the main arguments for having human shaped robots, adapted to the messy living environments of actual humans.

Given that careful analysis from September I do not share the hype that surrounds humanoid robotics today. Some of it is downright delusional across many different levels.

At the end:

Meanwhile here is what I said at the end of my September blog post about humanoid robots and teaching them dexterity. I am not at all negative about a great future for robots, and in the nearish term. It is just that I completely disagree with the hype arguing that building robots with humanoid form magically will make robots useful and deployable. These particular paragraphs followed where I had described there, as I do again in this blog post, how the meaning of self driving cars has drifted over time.

Following that pattern, what it means to be a humanoid robot will change over time.

Before too long (and we already start to see this) humanoid robots will get wheels for feet, at first two, and later maybe more, with nothing that any longer really resembles human legs in gross form. But they will still be called humanoid robots.

Then there will be versions which variously have one, two, and three arms. Some of those arms will have five fingered hands, but a lot will have two fingered parallel jaw grippers. Some may have suction cups. But they will still be called humanoid robots.

Then there will be versions which have a lot of sensors that are not passive cameras, and so they will have eyes that see with active light, or in non-human frequency ranges, and they may have eyes in their hands, and even eyes looking down from near their crotch to see the ground so that they can locomote better over uneven surfaces. But they will still be called humanoid robots.

There will be many, many robots with different forms for different specialized jobs that humans can do. But they will all still be called humanoid robots.

As with self driving cars, most of the early players in humanoid robots, will quietly shut up shop and disappear. Those that remain will pivot and redefine what they are doing, without renaming it, to something more achievable and with, finally, plausible business cases. The world will slowly shift, but never fast enough to need a change of name from humanoid robots. But make no mistake, the successful humanoid robots of tomorrow will be very different from those being hyped today.

Neural networks

Despite their successes with language, LLMs come with some serious problems of a purely implementation nature.

First, the amount of examples that need to be shown to a network to learn to be facile in language takes up enormous amounts of computation, so the that costs of training new versions of such networks is now measured in the billions of dollars, consuming an amount of electrical power that requires major new investments in electrical generation, and the building of massive data centers full of millions of the most expensive CPU/GPU chips available.

Second, the number of adjustable weights shown in the figure are counted in the hundreds of billions meaning they occupy over a terabyte of storage. RAM that is that big is incredibly expensive, so the models can not be used on phones or even lower cost embedded chips in edge devices, such as point of sale terminals or robots.

These two drawbacks mean there is an incredible financial incentive to invent replacements for each of (1) our humble single neuron models that are close to seventy years old, (2) the way they are organized into networks, and (3) the learning methods that are used.

That is why I predict that there will be lots of explorations of new methods to replace our current neural computing mechanisms. They have already started and next year I will summarize some of them. The economic argument for them is compelling. How long they will take to move from initial laboratory explorations to viable scalable solutions is much longer than everyone assumes. My prediction is there will be lots of interesting demonstrations but that ten years is too small a time period for a clear winner to emerge. And it will take much much longer for the current approaches to be displaced. But plenty of researchers will be hungry to do so.

LLMs

So we all know we need guard rails around LLMs to make them useful, and that is where there will be lot of action over the next ten years. They can not be simply released into the wild as they come straight from training.

This is where the real action is now. More training doesn’t make things better necessarily. Boxing things in does.

Already we see companies trying to add explainability to what LLMs say. Google’s Gemini now gives real citations with links, so that human users can oversee what they are being fed. Likewise, many companies are trying to box in what their LLMs can say and do. Those that can control their LLMs will be able to deliver useable product.

A great example of this is the rapid evolution of coding assistants over the last year or so. These are specialized LLMs that do not give the same sort of grief to coders that I experienced when I first tried to use generic ChatGPT to help me. Peter Norvig, former chief scientist of Google, has recently produced a great report on his explorations of the new offerings. Real progress has been made in this high impact, but narrow use field.

New companies will become specialists in providing this sort of boxing in and control of LLMs.

A note on embodiment

But since 1991 I have made a distinction between two concepts where a machine, or creature can be either, neither, or both situated and embodied. Here are the exact definitions that I wrote for these back then:

[Situatedness] The robots are situated in the world—they do not deal with abstract descriptions, but with the here and now of the world directly in-fluencing the behavior of the system.

[Embodiment] The robots have bodies and experience the world directly—their actions are part of a dynamic with the world and have immediate feed-back on their own sensations.

At first glance they might seem very similar. And they are, but they are also importantly different. And, spoiler alert, I think much of the work at companies, large and small, right now, is trying abstract out the embodiment of a robot, turning it into a machine that is merely situated.

Later:

Being both situated and embodied is still a challenge to robots in the world. [[Now here is the most important sentence of this whole blog post.]] I think the training regimes that [are] being used for both locomotion and dexterity are either ignoring or trying to zero out the embodiment of physical robots, their inertias and forces, reducing them to merely being situated, just apps with legs and arms, characters in video games, not the reality of real physical beings that the tasks we want them to do requires.

Puttering around as a behavioral mode

Over the weekend I got to thinking about the activity of puttering around. I wondered whether or not there has been any formal study about. So I put the following prompt to ChatGPT 5.2:

I’ve talked a lot about DMF (default mode network) and mind wandering in connection with creativity. It seems to me that puttering around is in the same ballpark. I’m not thinking and pondering, but I’m doing stuff, rather casually and in no forced order. Just as things come. It seems I like to putter early in the morning, on Friday afternoon, and over the weekend. Thus I just spent two days puttering around.

Since I’ve not had a “day job” in years, the distinction between weekdays and weekends is not forced on me. Yet it remains with me. It structured by time for most of the year in primary and secondary school. A bit less so in undergraduate and graduate school, but still operative. It also functioned when I had a faculty position at RPI (1978-1985) and at those times later on when I had one of those day jobs. So, it’s become deeply internalized though no longer externally imposed.

Its response took the form of numbered headings, subheadings, and bulleted lists. I find that sometimes useful, sometimes annoying. Since I wanted to write it up in a blog post I asked ChatGPT to express that material in the form of expository prose. I’ve put its response below the asterisks.

* * * * *

I’ve been thinking about puttering—not thinking hard, not resting, but doing things casually and without a forced order. It feels important, especially in connection with creativity, but I don’t know of any literature that treats it directly. What follows is an attempt to make sense of that intuition.

Cognitive science has devoted enormous attention to the default mode network (DMN)—a set of brain regions that become active when we are not engaged in externally directed tasks. The DMN is associated with mind wandering, autobiographical memory, imaginative simulation, and creative recombination. It is usually contrasted with task-positive networks that dominate during focused, goal-directed work.

But this framing leaves out a familiar and, I suspect, crucial mode of everyday cognition: puttering around.

Puttering is not rest, and it is not deliberate thinking. It is also not task-oriented in any strong sense. When I am puttering, I am doing things—moving objects, skimming texts, opening files, following small curiosities—but without urgency, hierarchy, or a clear endpoint. Attention shifts as affordances present themselves. One thing leads to another, but nothing insists.

This matters because puttering is not simply mind wandering in the absence of action. It is mind wandering coupled to low-stakes activity. That coupling provides just enough structure to keep cognition anchored in the world, while leaving associative processes free to roam. In that sense, puttering is an enacted version of default-mode cognition.

Most DMN research is organized around a sharp contrast: rest versus task. Subjects lie still in scanners or perform narrowly defined activities. What rarely gets studied is the large middle ground—self-paced, low-demand activity where executive control is attenuated but not absent. Yet it is precisely in this middle ground that many creative insights seem to arise.

One way to think about puttering is as a form of foraging. Cognitive scientists have developed mathematical models of information foraging and semantic search, showing how minds move through memory spaces under conditions of weak constraint. But these models are typically abstract and disembodied. Puttering, by contrast, is foraging with hands, eyes, and objects. It is exploratory behavior distributed across mind and environment.

Embodied and enactive theories of cognition come closer to capturing this phenomenon. They emphasize that cognition is not confined to the head but unfolds through interaction with the world. Still, even here, the focus is often on skilled action or purposeful engagement. Puttering remains oddly invisible: too mundane, too informal, too unproductive by conventional standards.

Creativity research acknowledges something similar under the heading of incubation. We are told that stepping away from a problem can allow unconscious processes to reorganize it. But incubation is usually framed as a gap between periods of work, not as a mode of activity in its own right. Puttering is active incubation—ongoing, embodied, and temporally permissive.

Time matters here. Even though I no longer have a “day job,” I notice that puttering reliably emerges at certain times: early in the morning, Friday afternoons, weekends. These temporal zones were once imposed externally—by school schedules, academic calendars, workweeks—but they have become internalized. They now function as cognitive affordances. Certain kinds of thinking feel licensed in those periods, while others feel out of place.

That licensing effect is important. Puttering lowers internal monitoring and performance pressure. It delays premature optimization. It allows partially formed ideas to coexist without being forced into coherence too soon. In more technical terms, it increases associative bandwidth while keeping executive control at a minimum viable level. For someone with a large and densely connected memory base, that is a powerful configuration.

Modern work culture tends to suppress puttering. It valorizes focus, efficiency, and measurable output, leaving little room for loosely structured activity that may not immediately “pay off.” Ironically, many contemporary AI systems explicitly reintroduce formal analogues of puttering: stochastic sampling, random walks, background exploration, temperature-controlled search. We design machines to do what we often deny ourselves permission to do.

There is, as far as I know, no direct scientific literature on puttering around. Instead, it falls through the cracks between DMN research, embodied cognition, creativity studies, and ecological psychology. Yet that absence is revealing. Puttering occupies a cognitive regime that our current conceptual frameworks do not handle well: neither rest nor work, neither task nor distraction, but something quietly generative in between.

If that is right, then puttering is not a failure of discipline or a lapse in productivity. It is a mode of cognition in its own right—one that deserves recognition, protection, and perhaps eventually, study. 

* * * * *

Note, however, that I did not explicitly ask it to write as my voice. As you can see, though, that's what it did. Oh well.

Sunday, January 11, 2026

Saturday, January 10, 2026

What a Wonderful World (Louis Armstrong) | Playing For Change

Playing For Change is proud to present this video of the song "What A Wonderful World" featuring Grandpa Elliott with children's choirs across the globe. In these hard times, children and music bring us hope for a better future. Today we celebrate life and change the world one heart and one song at a time!!

Classical improvisation by a Danish American Jewish comedian, Victor Borge

I'm bumping this one to the top of the queue because 1) I want to, 2) it's my blog, and 3) Borge is wonderful.
Victor Borge never got the memo saying that classical musicians stopped improvising sometime during the middle of the 19th century. So, when a fiddle player wanted to perform a tune that Borge had heard, but never played, Borge simply improvised an accompaniment. It's a bit over the top at points, but then Borge is a comedian. Watch how the two men interact act with one another. There are points where one or the other doesn't quite know what's coming up, so they have to look and listen.
 
 
Pay attention at about 3:13, in a slow section (stuck in the middle of all the fast stuff). At about 3:43 Borge starts a nice counter melody; from 3:54 to the end it's nuts, with a nice counter melody in octaves at about 4:02. Notice the nice hesitation for the very last note, a skillful touch.

Direct brain-to-brain communication, redux

Why Learning Does Not Rescue Brain-to-Brain Thought Transfer 

Learning Is Not the Problem

There is no serious dispute, at this point, about the brain’s capacity to learn to incorporate new signal streams. Decades of work on motor prostheses, sensory substitution, neurofeedback, and tool use have demonstrated that the nervous system can adapt to novel inputs and outputs that are not part of its evolved repertoire. These systems work not because the brain passively receives meaning, but because it actively learns to coordinate new patterns of neural activity with action, perception, and feedback. Over time, what begins as an alien signal can become functionally integrated into the organism’s sensorimotor economy.

Acknowledging this plasticity does not weaken skepticism about direct brain-to-brain thought transfer. On the contrary, it sharpens the distinction between what is genuinely possible and what remains a fantasy. Learning is one thing. Zero-shot “plug-and-play” communication is something else entirely. The speculative proposals advanced by Elon Musk, Christof Koch, and Rodolfo Llinás depend not merely on plasticity, but on the assumption that meaningful mental content can be transferred between brains without a learning history, without negotiation, and without interpretive work. That assumption is precisely what fails.

The Zero-Shot Assumption

The defining feature of most brain-to-brain communication fantasies is immediacy. Thoughts are imagined to pass directly from one person to another, bypassing language, culture, and development. Koch’s examples of ghostly visual overlays and mind fusion, as well as Musk’s talk of “uncompressed conceptual communication,” all presuppose that the recipient brain can immediately make sense of neural activity originating elsewhere. The temporal dimension of learning—the weeks, months, or years required to integrate new signal regimes—is simply ignored.

This is not a minor omission. It is the conceptual hinge on which the entire proposal turns. Without a learning trajectory, there is no mechanism by which foreign neural activity could acquire meaning for the receiving brain. A signal does not become meaningful by virtue of its richness or bandwidth. It becomes meaningful only through use, within a system that can test, revise, and stabilize interpretations through action.

Why Learning Cannot Proceed in a Brain-Bridge

One might reply that learning could occur even in a brain-to-brain link, given enough time. But this response overlooks the conditions under which learning is possible in the first place. Learning requires a closed perception–action loop. The organism must be able to act on the basis of a signal, observe the consequences of that action, and adjust its internal dynamics accordingly. In brain–machine interfaces, this loop is explicit: the user moves a cursor, grasps an object, or modulates a tone, and receives immediate feedback. The signal becomes meaningful because it is embedded in a task space with clear success and failure conditions.

A direct brain-to-brain link provides no such structure. The receiving brain cannot act into the other brain in any systematic way, nor can it test hypotheses about what a given pattern of activity “means.” The signal stream has no stable reference point in the shared environment, no agreed-upon goal, and no external criterion of correctness. Under such conditions, learning has nothing to converge on. What is sometimes described as “another person’s thought” arrives as undifferentiated neural activity, untethered from the bodily and environmental contexts that made it meaningful in the first place.

The Persistent Problem of Origin

Even if one were to imagine some form of slow co-adaptation, a deeper problem remains: the brain must be able to distinguish between activity it generates itself and activity it should treat as input. In ordinary perception and action, this distinction is grounded in efference copy, proprioception, and the tight coupling between movement and sensation. These mechanisms allow the brain to tag certain patterns as self-generated and others as world-generated.

A foreign brain provides none of these anchors. Neural spikes arriving from another person’s cortex are indistinguishable, in their physical characteristics, from spikes arising endogenously. Without a principled way to mark activity as coming from an other, the receiving brain has no basis for interpretation, let alone learning. The problem is not noise in the engineering sense, but indeterminacy in the biological sense. The system lacks the resources to sort the signal at all.

Meaning Is Not a Payload

Underlying the zero-shot fantasy is a deeper theoretical mistake: the treatment of meaning as something that exists prior to expression and can therefore be transmitted once bandwidth constraints are removed. This is the same mistake that underwrites the conduit metaphor of language. Words are imagined as containers for thoughts, and communication as the transfer of those containers from one mind to another. Neuralink-style speculation simply replaces words with spikes, while leaving the basic picture intact.

But meaning does not work that way. Whether one follows Vygotsky, contemporary enactivism, or predictive-processing accounts, the conclusion is the same: meaning is enacted, not transmitted. It arises through socially scaffolded activity, through interaction with the world and with others, and through the internalization of those interactions in inner speech. There is no pre-linguistic, pre-social format of “pure thought” waiting to be uploaded or shared.

Augmentation Without Communion

None of this casts doubt on the medical and augmentative goals of current BCI research. Restoring motor function, providing artificial sensory channels, and extending human capabilities through learned interfaces are all plausible and worthwhile. But these technologies work precisely because they respect the conditions under which brains learn: limited task spaces, stable feedback, and prolonged adaptation. They augment agency; they do not merge subjectivities.

Direct brain-to-brain thought transfer, by contrast, promises communion without development, understanding without negotiation, and immediacy without practice. It imagines semantic interoperability where none can exist. For that reason, it fails not because the technology is immature, but because the underlying conception of thought, meaning, and learning is mistaken.

The issue, in the end, is not whether brains can change. They can, and they do. The issue is whether meaning can be detached from the histories that make it possible. On that point, the answer remains no.

Friday, January 9, 2026

Variations on a Wild Image

I decided to run a bunch of variations on the image that I used for the cover of my working paper, Serendipity in the Wild: Three Cases, With remarks on what computers can’t do (link to the blog post where I introduce it).  I'm listing them in the order that I had ChatGPT create them. That will help explain why, for example, the shark in Mughal version has has a goofy look on its face and why that version has a mangled watch in the sand. 

For the most part the prompt was simple, the name of a style  or an artist. The manga style is my favorite. Some of the images are less successful than others.

The style of a Renaissance etching 

The style of a comic book devoted to science fiction stories. 

Manga style 

Japanese Ukiyo-e print.

I realize that the subject matter is, shall we say, anachronistic for ukiyo-e, but that goofy guy at the lower right is a freakin' bridge too far.

Chuck Jones style (Warner Brothers).