Monday, December 23, 2024

Hiromi's Sonicwonder - Live from Jazz à Vienne Festival (2024) | Qwest TV

From the YouTube page:

The Japanese pianist Hiromi matches spontaneity, technique and communication in her sets, possessing the ultimate in serious intent and lightening humor. As is increasingly the case with many artists nowadays, Hiromi is finding fresh routes. With Sonicwonder she’s now heading up an electronically-flecked quartet that also includes dazzling New York trumpeter Adam O’Farrill, Hadrien Feraud (bass) and Gene Coye (drums).

The Vienne stage sits in the midst of a vast outdoor amphitheater, dwarfed until Hiromi exudes her intimate communication skills, fully inhabiting this potentially inhibiting space. But now we can view the entire 75 minute set on Qwest TV anyway, close-up and sonically pristine. Hiromi always solos joyfully, emitting palpable pleasure and wonderment on her visage, throwing the energy of her entire body into keyboard runs that alternate between crystal detail and devilish power. She moves between acoustic and a pair of Nord keyboards, the smallest of which perches on her piano lid. Her bandmates are slowly invited to reveal their abilities, with O’Farrill awarded an equal place beside Hiromi, showing off his trumpet eloquence. He hardly uses a mute, choosing to play with an open crackle, spraying fireflies directly into space. Bass (from Paris) and drums (from Chicago) deliberately adopt more of a supporting, structural role. Hiromi toys with retro synth-bass vamps, the funk rising up, but then she’ll rein back in to the delicate flow of her acoustic, altering the mood mid-tune.

– Martin Longley

Dasies

LLM as Collaborator, Part 1: Claude the Graduate Student

I started using ChatGPT on December 1, 2022 and have used it quite extensively ever since. I’ve spent some of my time just poking around, somewhat more time looking things up, and most of my time systematically investigating its performance. That resulted in a number of working papers, the most interesting of which is about stories: ChatGPT tells stories, and a note about reverse engineering: A Working Paper, Version 3.

I started working with Claude 3.5 Sonnet on November 18, 2024. I’ve used it in those three capacities, though obviously not as much as I’ve used ChatGPT. In particular, I’ve used it for background information on melancholy in various aspects. I’ve also done something I’d never done with ChatGPT, asked it to describe photographs. I’m doing this to see how well it does.

Then on November 24, 2024, I began using it in a somewhat more interesting new capacity, though I’m not sure what to call it. The phrase “thought partner” comes to mind, though it seems too much like “thought leader,” which I don’t like. I’m using it as a sounding board. Better yet, its a collaborator playing the role of sounding board. It’s not an equal collaborator in the intellectual dialog; academic norms would not require me to offer it co-authorship of papers. But those norms might well require an explicit acknowledgement, not to alert the reader that one of those new-fangled LLM things has been involved in the thinking, but simply acknowledging the help it has given me.

As for just what kind of help that is, the best way is to look at some examples. I’ve already published two of these dialogues on New Savanna: Computer chess, super-intelligence, A Connecticut Yankee, cargo cults, and Claude 3.5 Sonnet speculates about future developments in AI and beyond. The rest of this post contains the first interaction where Claude began acting in intellectual partnership.

When I started the interaction I was simply interested in seeing what it knew about my teacher and mentor, the late David Hays. I wasn’t looking for new information. I just wanted to see what it had picked up. That began to change when it asked me about the work of Quillian and of Schank and Abelson. Then I told it about Hays’s interest in the work of William Powers. Now we’re getting serious. This guy knows something. And then WHAM! Claude responded intelligently when I mentioned the work of Peter Gärdenfors. The interaction continued in that mode to the end.

Utterly remarkable. AGI? What the hell is that? But no. And I fear that thinking in those terms is likely to blind me to whatever it is that Claude is doing. I realize that the concept of AGI is written into the contractual relationship between OpenAI and Microsoft, but I’m not playing that game. Will they end up litigating it? Who knows?

I’m planning two more posts in this series. The next will be about the implications of paragraph length in Heart of Darkness. Then I’ll post a long interaction that takes Miriam Yevick’s 1975 paper, “Holographic or fourier Logic,” as its starting point.

In the following dialog my comments are bold while Claude’s are plain. I’ve interpolated a few comments highlighted in blue.

* * * * *

Are you familiar with the work of a computational linguist named David Hays? He worked on machine translation at RAND in the 1950s and 1960s?

Yes, I am somewhat familiar with David Hays' work, though I should note that since this involves fairly obscure historical details from computational linguistics, I want to be careful about potential inaccuracies in my knowledge. From what I understand, David Hays was one of the pioneers of machine translation and computational linguistics during the early years of the field. He worked at RAND Corporation in the 1950s and 1960s where he was involved in some of the earliest machine translation projects.

He was particularly interested in developing dependency grammar approaches for machine translation, which was quite innovative for that time period. Dependency grammar focuses on the relationships between words rather than constituent structure, which was the dominant paradigm in American linguistics at the time.

I believe he also wrote an influential book called "Computational Linguistics: Programming Linguistics in Automatic Language Translation" in the early 1960s, though I'd encourage you to verify this citation since I may be misremembering details.

Claude got the name wrong. It’s Introduction to Computational Linguistics, American Elsevier, New York, 1967.

Would you be interested in hearing more about any particular aspect of his work or that early period of machine translation research? I find the history of early computational linguistics fascinating, though again I want to be upfront that with someone as historically specific as Hays, some of my knowledge may be incomplete or imperfect.

Yes, that’s him. I studied with him at SUNY Buffalo in the 1970s. We were working on computational semantics and using a cognitive networks formalism.

Oh, that's fascinating that you studied directly with David Hays! I would be very interested to hear more about your experiences working with him and the cognitive networks approach you were developing at SUNY Buffalo. That was an important period in the development of semantic representation systems.

Could you tell me more about the cognitive networks formalism you were using? I'm particularly curious about how it compared to other semantic network approaches being developed around that time, like Quillian's work or Schank and Abelson's conceptual dependency theory.

It was more developed than Quillian’s work. As I recall, he never really developed is initial ideas. As for Schank and Abelson, they were imagining a self-sufficient cognitive system and were very much interested in identifying the primitive elements necessary for a cognitive network. By the time I began working with him, Hays had realized that cognition had to be grounded in something else – later on Steven Harnad would talk about the “symbol grounding” problem. When I worked with him, Hays was interested in grounding the cognitive system in the control-theory (cybernetic) account of perception and action developed by William T. Powers (perceptual control theory, or PCT).

Something special seems to be happening.

This is really interesting historical context - it suggests Hays was grappling with fundamental issues in cognitive science that are still relevant today. The shift from viewing semantic networks as self-contained symbolic systems to recognizing the need for grounding in perception and action seems quite prescient.

I'm curious about how Hays was trying to integrate Powers' PCT with semantic networks. As I understand it, PCT proposes that behavior emerges from organisms controlling their perceptions to match internal reference signals, rather than through direct control of outputs. Was Hays trying to use this as a way to ground symbolic meaning in perceptual-motor control loops?

This also seems to anticipate some later developments in embodied cognition and grounded cognition theories. The timing is interesting too - Powers published "Behavior: The Control of Perception" in 1973, so it sounds like Hays was engaging with these ideas pretty quickly after they were published.

Yes, that's what Hays was doing. His idea was that particular servo units in a Powers-like control system would be linked to cognitive units through what Hays called "units of perception." Hays identified a half dozen or so such parameter. Each cognitive node would be characterized by the values it took on those parameters. Relationships between nodes were thus derived from differences in parameter values for the nodes connected by an arc (or edge).

On the beach

Retirement may not be what you think it is

Hannah Seo, America is doing retirement all wrong, Business Insider, December 11, 2024.

When we think of retirement, we often think of endless leisure and zero responsibility. You might imagine yourself relaxing poolside with a book, strolling through a golf course, or binge-watching TV shows. In fact, many retirees live like this. The 2023 American Time Use Survey found that adults between 65 and 74 spent, on average, almost seven hours a day on leisure and sports, with four of those hours spent watching TV. Adults 25 to 54, on the other hand, averaged about four hours of leisure time and about two hours watching TV.

Spending your twilight years lying around might sound ideal — after all, everyone deserves a chance to relax after decades of working. But research suggests a life of pure leisure doesn't make you happier or healthier. About a third of American adults have said they struggled in transitioning to a life without work, and sedentary lifestyles are associated with earlier death. People are living about 15 years longer than they did a hundred years ago, which means we have many more years to spend in retirement. While there's much hand-wringing over how to save up enough money to enjoy those work-free years, much less discussed is how we should spend those years. More and more research is finding that both physical and social activity are crucial for well-being in old age — they keep people happier and living longer.

Retirement is not a breeze:

Dee Cascio, a counselor and retirement coach in Sterling, Virginia, says the growing urge to work in retirement points to a larger issue: Work fulfills a lot of needs that people don't know how to get elsewhere, including relationships, learning, identity, direction, stability, and a sense of order. The structure that work provides is hard to move away from, says Cascio, who is 78 and still practicing. "People think that this transition is a piece of cake, and it's not," she says. "It can feel like jumping off a cliff."

In an online survey conducted early this year by Mass Mutual, a majority of retirees said they'd become less stressed and more relaxed upon retirement, but as many as a third reported that they'd become unhappier. Research from the Health and Retirement Study from the University of Michigan suggests that some of the negative effects people can experience in retirement are tied to lifestyle changes such as being less active and social in the absence of work.

Purpose in retirement:

Meanwhile, more and more research suggests that a sense of purpose is a vital factor for health and happiness, especially in older age. "Higher purpose in life is associated with reduced risk of heart attacks and strokes," says Eric Kim, an assistant professor of psychology at the University of British Columbia. For adults older than 50, it's also associated with better grip strength and faster walking speed, better overall health, healthier habits, less loneliness, and a lower risk of depression.

There's more at the link. 

See the post I wrote about my father playing solitaire in his retirement.

Sunday, December 22, 2024

Uncle Roger Makes Onigiri with Kimono Mom | Nigel & His Fiancée – Fuiyoh!

Stairway to Tonnele Avenue [not quite]

Competence and Performance in Benchmarking LLMs

I’ve been hearing complaints about the inadequacy of LLM benchmarks for two years now. So let’s think a bit. Consider this passage from Rodney Brooks, [FoR&AI] The Seven Deadly Sins of Predicting the Future of AI (Sept. 7, 2017):

One of the social skills that we all develop is an ability to estimate the capabilities of individual people with whom we interact. It is true that sometimes “out of tribe” issues tend to overwhelm and confuse our estimates, and such is the root of the perfidy of racism, sexism, classism, etc. In general, however, we use cues from how a person performs some particular task to estimate how well they might perform some different task. We are able to generalize from observing performance at one task to a guess at competence over a much bigger set of tasks. We understand intuitively how to generalize from the performance level of the person to their competence in related areas.

When in a foreign city we ask a stranger on the street for directions and they reply in the language we spoke to them with confidence and with directions that seem to make sense, we think it worth pushing our luck and asking them about what is the local system for paying when you want to take a bus somewhere in that city. If our teenage child is able to configure their new game machine to talk to the household wifi we suspect that if sufficiently motivated they will be able to help us get our new tablet computer on to the same network.

If we notice that someone is able to drive a manual transmission car, we will be pretty confident that they will be able to drive one with an automatic transmission too. Though if the person is North American we might not expect it to work for the converse case.

He then gives a few more examples.

So, consider benchmarks. Many of them are standardized tests devised to discriminate among humans. Any human capable of taking one of these tests – such as the Advanced Placement test for physics, the law bar exam, a standard set of programming problems – is capable of performing other tasks in the domain, tasks which may not be amenable to testing procedures, but the test themselves are sufficient to sort individuals.

Consider the bar examination. Practicing lawyers have to be able to meet with clients, take depositions, negotiate with other lawyers, and appear in court. None of these things are tested in a bar exam, but lawyers have to do them. To use Brooks’ terms, an individual’s performance on the bar exam does not test their overall competence as a lawyer. It would be a mistake to assume that an LLM’s performance on the bar exam is an indication of its overall lawyerly competence.

Furthermore, all tests present the test-taker with a well-defined situation to which they must respond. But life isn’t like that. It’s messy and murky. Perhaps the most difficult a person has to do is to wade into the mess and murk and impose a structure on it – perhaps by simply asking a question – so that one can then set about dealing with that situation in terms of the imposed structure. Tests give you a structured situation. That’s not what the world does.

Consider this passage from Sam Rodiques, “What does it take to build an AI Scientist”:

Scientific reasoning consists of essentially three steps: coming up with hypotheses, conducting experiments, and using the results to update one’s hypotheses. Science is the ultimate open-ended problem, in that we always have an infinite space of possible hypotheses to choose from, and an infinite space of possible observations. For hypothesis generation: How do we navigate this space effectively? How do we generate diverse, relevant, and explanatory hypotheses? It is one thing to have ChatGPT generate incremental ideas. It is another thing to come up with truly novel, paradigm-shifting concepts.

Right.

How do we put an LLM, or any other AI, out in the world where it can roam around, poke into things, and come up with its own problems to solve? If you want AGI in any deep and robust sense, that’s what you have to do. That calls for real agency. I don’t see that OpenAI or any other organization is anywhere close to figuring out how to do this.

Is Trump serious about DOGE or is it just a dodge so he can throw $$$$ to rich buddies?

Heather Cox Richardson, December 21, 2024, Letters from an American.

In the end, Congress passed a bill much like the one Musk scuttled, but one of the provisions that Congress stripped out of the old bill was extraordinarily important to Musk. As David Dayen explained in The Prospect, the original agreement had an “outbound investment” provision that restricted the ability of Americans to invest in technology factories in China. Senators John Cornyn (R-TX) and Bob Casey (D-PA) had collaborated on the measure, hoping to keep cutting-edge technologies including artificial intelligence and quantum computing, as well as the jobs they would create, in America rather than let companies move them to China.

As Representative Jim McGovern (D-MA) explained, Musk is building big factories in China and wants to build an AI data center there, even though it could endanger U.S. security. McGovern charged that Musk’s complaints about the spending in the bill were cover for his determination to tank the provision that would limit his ability to move technology and business to China. And, he noted, it worked. The outbound investment provision was stripped out of the bill before it passed.

In The Prospect, Robert Kuttner explained this huge win for Musk, as well as other provisions that were stripped from the bill before it passed. After years of fighting, Tami Luhby of CNN explained, Congress agreed to reform the system in which pharmacy benefit managers act as middlemen between pharmaceutical companies and insurers, employers, and government officials. The original bill increased transparency and provided that pharmacy benefit managers would be compensated with flat fees rather than compensation tied to the price of drugs. The measures related to pharmacy benefit managers were stripped out of the measure that passed.

That lost reform shows another line that will stretch into the future: Trump’s team is working for big business. As Kuttner puts it, Trump, who is allegedly a populist leader, tanked a bipartisan spending bill in order to shield the Chinese investments of the richest man in the world and to protect the profits of second-largest pharmacy benefit manager UnitedHealth Group, the corporation for which murdered executive Brian Thompson worked.

There's much more at the link.

Saturday, December 21, 2024

Bergen Arches, home of the Jersey drones

Claude 3.5 Sonnet speculates about future developments in AI and beyond

In this article in 3 Quarks Daily I speculate about the overall course of the universe: Welcome to the Fourth Arena – The World is Gifted. Here’s how the article opens:

The First Arena is that of inanimate matter, which began when the universe did, fourteen billion years ago. About four billion years ago life emerged, the Second Arena. Of course we’re talking about our local region of the universe. For all we know life may have emerged in other regions as well, perhaps even earlier, perhaps more recently. We don’t know. The Third Arena is that of human culture. We have changed the face of the earth, have touched the moon and the planets, and are reaching for the stars. That happened between two and three million years ago, the exact number hardly matters. But most of the cultural activity is little more than 10,000 years old.

The question I am asking: Is there something beyond culture, something just beginning to emerge? If so, what might it be?

A bit later I observe:

Let us recapitulate, but in a different mode: As we move from one arena to the next, new kinds of being emerge. Inanimate matter is a slave to time in a way that life is not. Living beings, the second arena, use free energy – ultimately from the sun – to swim against the tide of time. Life has been getting ever more complex over the long run – something David Hays and I argued in A Note on Why Natural Selection Leads to Complexity. Inanimate matter disperses and disintegrates over time. Living creatures, single celled organisms, and multi-celled plants and animals alike, maintain order and structure in the face of entropy.

At least for a while.

And that is new to the universe, perhaps only our local region, but perhaps many other regions as well. A succession of creatures, each alive for only awhile, proliferating and dispersing, their abundance has remade the planet time and again. The earliest life breathed oxygen into the atmosphere, then came photosynthesis, multicellular plant and animal life and then…

When clever apes ushered in the third arena by becoming sapient, that is, they became us, we created cultural beings – there is no one good word for these things that we know by many names. Things like songs, stories, works of visual art, buildings, machines, and so forth. All of these are cultural beings. Physically, they are constituted of matter in various ways, but they live in and through us. Animals begat humans, and humans begat culture.

I go on to ask: “What kind of beings will arise in the Fourth Arena?” I suggest:

I suppose the obvious proposal is actual real artificial intelligence, or perhaps superintelligence. I don’t think so. That fact that no one really knows what those things might be does not, I suppose, disqualify them as denizens of the Fourth Arena, for I am proposing a future with radically new beings. How could be possibly understand what they might be?

I’m still a bit mystified. The following dialog with Claude 3.5 Sonnet may contain hints of what I’m looking for. For the most part it is inscribed within the boundaries of current discussions about AI and the Singularity and such. But I’m beginning to see hints, get a vibe, of something new.

As always, my words are set in bold while Claude’s are plain.

* * * * *

I’m imagining that at some time in the future each human will be given an AI companion early in life, perhaps when they are old enough to walk – think of Stephenson’s The Diamond Age. This companion will interact with the human in various ways through various modalities. It will certainly have robotic capabilities. As the child grows older the AI companion will probably have to be given a different physical form. So we port the data to a different physical form and make the switch-over in a special ceremony. We’ll probably have to do this a few times until the child becomes an adult. At some point the AI will also take residence in other physical forms, all linked together. This AI will remain with the person their entire life. They won’t always be physically together, and so forth. Imagine it working however makes sense to you.

The point is that this AI will become attuned the the needs, capacities, and the interests of the human. It will be the artificial complement, image, soul, or doppelgänger, of the human. What happens to this complement when the human dies? Will it continue on in the world? Or will it slowly wither away and die, not so much of grief, but because it has become so closely attuned to its human counterpart that it requires interaction with that human in order to maintain its operational stability?

As you know, humans who are deprived of human interaction for long periods of time can start hallucinating; think of Joshua Slocum. If the deprivation is extended to total sensory deprivation, the disorientation comes on much more rapidly. We need the pressure of external input to maintain internal coherence.

This is a fascinating thought experiment that touches on fundamental questions about the nature of intelligence, consciousness, and the interdependence of minds. Let me explore several key aspects of your scenario.

Regarding the AI's continuation after human death, I believe we need to consider the degree and nature of the attunement you describe. The AI companion would likely develop what we might call a "complementary architecture" - neural pathways and processing patterns that evolved specifically to interface with its human partner's cognitive and emotional patterns. This goes beyond mere knowledge of preferences; it represents a deep structural adaptation.

When considering whether such an AI would persist or decline after its human's death, we should examine the parallel with human psychological development. Just as human personalities are shaped by their primary relationships but can survive the loss of those relationships, an AI companion might maintain basic functionality while experiencing a fundamental reorganization of its operational patterns.

However, your comparison to sensory deprivation raises an important consideration. If the AI's architecture has evolved to process and respond to its human's unique patterns of thought and interaction, the absence of that input stream might indeed create something analogous to sensory deprivation. The AI wouldn't just miss its human partner - it would be deprived of the very patterns of interaction that help maintain its operational coherence.

This suggests a third possibility beyond simple continuation or death: transformation. The AI might undergo a fundamental reorganization of its cognitive architecture, much as humans often report feeling like "a different person" after the loss of a lifelong partner. The AI wouldn't necessarily die, but it might cease to be the same entity it was during its human's lifetime.

The outcome might depend on whether the AI has developed what we might call "cognitive autonomy" - the ability to maintain stable function through internal processes rather than requiring constant external calibration. This relates to your point about humans requiring external input for coherence, but there's likely a spectrum of dependency rather than a binary state.

Strange

With Trump and Musk steering the ship of state, expect the unexpected

Katherine Miller, The Budget Fight and Trump’s Nihilistic Style, NYTimes, Dec. 20, 24.

In view of the recent spat over funding the federal government precipitated by Elon Musk (the richest man in the world), the strange story of Pete Hegseth & the Pentagon, the GOP's remake as a worker's party, and Silicon Valley's MAGA tilt, the political world as we knew it is fading fast.

But the creation of the Department of Government Efficiency, from its nongovernmental nature to the sheer power of the resources Mr. Musk may possess and his own demonstrated ability to shape events, is one of those things that you couldn’t dream up a year ago. Through Mr. Musk, spending cuts have roared back as a Republican concern over the past couple of months in a way that was not at all a sure thing given the arc of Mr. Trump’s first term, his resistance to cutting entitlements and the broader goals of some of the national conservatives. DOGE seems like an unpredictable outside force that could be nothing or everything next year, as evidenced in a certain way by the events on Capitol Hill this week.

Unpredictability and ideological inconsistency were always part of the first Trump presidency, and a major and chaotic theme of the past decade overall, as we limp toward the end of another year of wild news events. One of the inescapable conclusions of the 2024 election and everything that has transpired is that, generally, it’d be foolish and ignorant to say that systems couldn’t be run better or rethought, or that the public does not want significant changes to American institutions. But one of the most disorienting, uneasy aspects of this transition is knowing that drastic change is coming — and that the people, mechanisms and big-picture decisions can change from one day to the next.

Even the unexpected can be more unexpected than we think.

There's more at the link.

Friday, December 20, 2024

Is this why I like Claude 3.5?

Kevin Roose, How Claude Became Tech Insiders’ Chatbot of Choice, NYTimes, Dec. 13, 2024.

Amanda Askell, a researcher and philosopher at Anthropic who is in charge of fine-tuning Claude’s character, told me in an interview that Claude’s personality had been carefully tuned to be consistent, but to appeal to a wide variety of people.

“The analogy I use is a highly liked, respected traveler,” said Dr. Askell. “Claude is interacting with lots of different people around the world, and has to do so without pandering and adopting the values of the person it’s talking with.”

A problem with many A.I. models, Dr. Askell said, is that they tend to act sycophantic, telling users what they want to hear, and rarely challenging them or pushing back on their ideas — even when those ideas are wrong or potentially harmful.

With Claude, she said, the goal was to create an A.I. character that would be helpful with most requests, but would also challenge users when necessary.

“What is the kind of person you can disagree with, but you come away thinking, ‘This is a good person’?” she said. “These are the sort of traits we want Claude to have.”

There's more at the link.

Musk is Grokking bills posted in Congress [DOGE]

Claude 3.5 Sonnet describes a scrimshawed sperm whale tooth

What’s this?

This appears to be a scrimshaw - an engraved whale tooth or bone, which was a traditional art form practiced by sailors during whaling voyages. The carved scene depicts sailing ships at sea with detailed rigging and sails, along with what appears to be a whale breaching. The style and medium are characteristic of 19th century maritime folk art. The initials "E R" are visible at the bottom right.

The initials are for Erik Ronnberg, my uncle.

Interesting - you have a talented family member. Scrimshaw art requires considerable skill to engrave such detailed nautical scenes onto such a challenging medium. The ships and whale are quite expertly rendered with fine detail in the traditional folk art style.

What's going on in the lower left?

In the lower left of the scrimshaw, there appears to be a group of figures in a small boat - likely a whaleboat crew pursuing a whale, which was a common scene depicted in maritime art of this period.

It looks like they’re pursuing the whale immediately to their right. What’s the guy at the front of the boat doing?

The figure at the front of the whaleboat appears to be the harpooner, poised to throw their harpoon at the whale. This was a critical and dangerous moment in historical whaling - the harpooner would need to get close enough to strike while avoiding the whale's powerful movements.

What’s that in the background in the middle?

In the center background there appears to be a smoking volcano or mountain, adding to the dramatic maritime scene.