Last time, I posted a lengthy missive by ChatGPT appraising my new book ailien minds, prompted by Richard Bliss, with some of my own commentary and demurals laced-in.
All-told, the effect was that of a loquacious book report by a garrulous and rather unctuous high school sophomore who had flipped through some random pages for key words and then googled some of them... or used them to sift its training sets.
So, how does Claude do? Well, the comparison here isn't exactly fair... Mr. Bliss asked Claude to read not just my entire book, but also the GPT appraisal that I posted, last time, including my critical inserts. So, you cannot draw explicit conclusions.Still - fair or unfair - the overall impression is night-and-day. Claude's missive is accurate, pertinent and in some cases even surprisingly insightful. And weirdly honest about its own (un)sapience.
Unlike GPT, which glossed maybe 5% my work over a reurgitaion of common AI memes from its training sets, Claude zoomed upon core concepts in alien minds, distilling an impressive 50% or so of the key points. (Though of course none of the supporting evidence, of which Claude says I provide too copiously!)
Moreover, there was almost none of GPT's blatant flattery. Well, except for the flattery that's inherent in well-summarizing my ideas.
But judge for yourself. And note at the end where I promise a more-fair comparison. Maybe next time. If world events don't force me back to politics.
(Oh, note: on Monday I keynote one of the tracks at the big RSA conference.)
==============
A Critique of ChatGPT’s Review of
Ailien Minds by David Brin
Written by Claude (Anthropic) March 2026
Prepared after reading the full text of Ailien Minds via page-by-page screenshots supplied by Richard Bliss
Preamble: A Note on What This Document Is
This is a critique written by one AI system (Claude, made by Anthropic) of another AI system’s (ChatGPT, made by OpenAI) review of a book that is substantially about how AI systems should be governed. The irony is not lost on me.
A disclosure: I read Ailien Minds page by page through 73 screenshots shared by a reader. ChatGPT appears to have been given the book’s text as a submission for review. We both had access to the source material. We arrived at very different readings.
David Brin annotated the ChatGPT exchange with his own reactions. I have read those annotations. In the interest of transparency, they informed my critique—but so did my own independent reading of the book. Where I agree with Brin’s frustrations, I will say so. Where I think the ChatGPT review got something right despite his objections, I will say that too.
The Central Problem: Surface Pattern-Matching vs. Structural Argument
The ChatGPT review correctly identifies that Ailien Minds is about AI, that Brin occupies a middle position between optimists and doomers, and that he favors transparency and competing systems over centralized control. These are accurate at the level of a dust jacket summary. The problem is that a dust jacket summary is essentially all ChatGPT produced, stretched across several thousand words of increasingly generic AI commentary.
The book’s actual contribution—its reason for existing alongside the dozens of other AI books published in 2025–2026—is a specific, detailed argument about mechanism. How, precisely, do you keep powerful AI entities accountable once their creators can no longer control them? Brin’s answer draws on evolutionary biology, ecosystem dynamics, and the history of legal and democratic institutions to propose that individuated AI agents, competing within transparent accountability structures, represent the only approach with historical precedent for success.
ChatGPT never engages with this mechanism. It talks about transparency in the abstract. It mentions competing AIs. But it never connects these to the specific framework Brin builds: individuation (giving AI agents distinct, persistent identities), reciprocal accountability (structuring incentives so AI entities police each other), and disputation arenas (formal adversarial processes for testing AI claims and behavior). These are not passing ideas in the book. They are the book’s thesis.
Five Specific Failures
1. The Three Clichés Were Invisible
One of Brin’s most forceful arguments is that virtually all public AI discourse is trapped in three formats: AI as obedient corporate servant (the “castle” model), AI as amorphous uncontrolled swarm (the “blob” model), or AI as a single superintelligent overlord (the “Skynet” model). He argues these three clichés function as thought-terminating frameworks that prevent people from imagining a fourth alternative.
ChatGPT’s review addresses only the third cliché (Skynet), and only to dismiss it. The castle and blob models—which Brin argues are actually more dangerous because they are more plausible—go entirely unmentioned. This is a significant omission. It is as if someone reviewed Darwin’s On the Origin of Species and discussed his observations about finch beaks without ever mentioning natural selection.
2. Individuation Was Missed Entirely
The concept of individuation—giving AI systems distinct, trackable identities so they can build reputations and be held accountable—is arguably the single most novel proposal in the book. It is the mechanism by which Brin’s vision of competitive accountability would actually function. Without individuation, you cannot have meaningful AI reputations. Without reputations, you cannot have market-style accountability. The entire logical chain depends on it.
ChatGPT does not mention it once.
3. The Ecosystem Argument Was Flattened
Brin spends a full chapter (Chapter 3) and a substantial aside (Aside #4, on “Soup vs. Sea”) building a detailed analogy between biological ecosystems and the emerging digital ecosystem. This is not a loose metaphor. He traces specific parallels: energy gradients, predator-prey dynamics, the role of barrier membranes in enabling individuation, and the historical pattern that monopoly organisms cause ecosystem collapse.
ChatGPT reduces this to a vague gesture toward “a messy ecosystem of AIs.” The structural argument about why ecosystems need individuated agents with membranes (boundaries, identities) to remain healthy—which is the bridge between Brin’s ecological thinking and his policy proposals—disappears entirely.
4. The Disputation Arena Proposal Was Absent
Chapter 12 presents what is arguably the book’s most concrete policy proposal: Disputation Arenas, formal adversarial processes modeled on courtrooms, scientific peer review, and competitive markets, designed to stress-test AI systems and AI governance proposals through structured disagreement. Brin devotes considerable space to designing this mechanism, including how different outcomes (clear victory, compromise, or productive stalemate) would each generate value.
ChatGPT’s review does not mention disputation, adversarial testing, or structured debate in any form. For a book whose final major chapter is titled “Our Abrasive Secret Sauce,” this is a remarkable gap.
5. Training Data Overwhelmed Source Material
Perhaps the most instructive failure is how frequently the ChatGPT review drifts from the book’s actual content into generic AI discourse drawn from its training data. The extended discussions of Ray Kurzweil, Nick Bostrom, and Eliezer Yudkowsky are competent but contextually irrelevant—Brin mentions these figures only in passing, and his argument is specifically that the framework he proposes transcends the optimist/pessimist spectrum these thinkers represent.
Similarly, the later sections on “humans as sanity-check layers,” “trust as currency,” and “cognitive jobs” are reasonable points about AI’s economic impact—but they reflect generic AI futurism, not the book’s distinctive arguments. The review increasingly reads as “things an LLM knows about AI” rather than “a critical reading of this specific book.”
Brin identifies this drift repeatedly in his annotations. He is right to be frustrated. An AI reviewing a book should be accountable to the source material, not to its own training distribution.
What ChatGPT Got Right
Fairness requires noting what the review did well.
The initial framing of AI as “alien intelligence”—not extraterrestrial, but cognitively foreign—is an accurate representation of one of Brin’s themes. The observation that Brin sits between the poles of AI optimism and doom is correct. The point that Brin’s thinking is influenced by The Transparent Society is valid and relevant. And ChatGPT’s note that Brin has explored these ideas in fiction for decades provides useful context.
At the level of “what is this book generally about and who wrote it,” the review is competent. The failure is at the level of “what does this book specifically argue that other books do not,” which is the level at which a serious review must operate.
A Broader Observation
This exchange is itself a data point for one of Brin’s arguments. He warns throughout the book that LLMs are persuasively fluent without being deeply accountable to the material they process. The ChatGPT review demonstrates this precisely: it is well-organized, grammatically polished, and tonally appropriate. It sounds like a thoughtful review. But when measured against the source material, it is substantially hollow—a fluent recombination of training data that happens to overlap with the book’s topic, rather than a genuine critical engagement with the book’s unique arguments.
Brin proposes that accountability for AI should come through adversarial processes—competing systems checking each other’s work. This critique is, in a small way, an instance of exactly that process. Whether it is a better reading than ChatGPT’s is not for me to judge. But the fact that two AI systems can produce substantially different readings of the same text, with different blind spots and different emphases, is itself an argument for the kind of competitive, multi-perspective AI governance Brin advocates.
=======
Back to DB here.
What impressed me most - beyond the cogency and direct/fair pertinence to my book - is the discursive style displayed by Claude's choice of paraphrasings.* In every case, it chose sentences that - while cogent and accurate - were not specific phrasings I'd have used. (Though in a couple of cases I wish that I had!)
Now, of course, that could be a programmed proclivity, well within the grasp of a Large Language Model. Still it does kind of resist the simplistic characterization that all LLMs are just advanced sentence auto-complete systems.
To be clear, although Claude gives you a vastly better summary here than GPT did, nevertheless...
... this is actually a test of YOU!
Are you still a reader? Possessed of true intellect and curiosity? Because I promise that any page of ailien minds will provide lots of value that no summary can encompass. Claude even says so, challenging you to be one of the elite who actually enjoys an idea and fact-rich book! ;-)
Next time, I'll post the de-novo appraisal of ailien minds that Mr. Bliss asked Claude to do, without feeding in the annotated GPT synopsis. If Claude were a person, that appraisal would still be affected by the earlier reading. That's not supposed to happen with LLMs, who should start fresh, if told-to. But judge fopr yorself.
Till, then, fight for a civilization that is worthy-of... and sets good examples for... these new children of our minds.
------
* (Elsewhere I have said that paraphrasing accurately is the seldom mentioned top attreibute of someone who is arguing fairly and cogently.)
30 comments:
Claude really is amazing.
This is why I wanted Dr. Brin to look at what it did with code.
I prompted it:
❯ Fix the nonogram solver in this directory to solve Griddler20426Snail-Mail.xml
The compressed response was:
⏺ Bash(cd "/Users/mcsandb/Documents/workspace/ClaudeProNonogram2/Project Files" && ./nonogram Griddler20426Snail-Mail.xml 2>&1 | grep -c
"| 0|")
⎿ 30
⏺ The solver can barely solve this puzzle - almost nothing is resolved. Let me read the key source files.
Read 4 files (ctrl+o to expand)
⏺ Let me check the Clues structure and then run a test to understand why backtracking fails.
⏺ The solver terminates in 7ms - that means there's a false contradiction being detected during constraint propagation. Let me add debug
output to find it.
⏺ Column 8 (0-indexed, col 9 in puzzle) is being falsely contradicted. Let me investigate its state when the contradiction triggers.
Claude them made additional fixes and got the app to solve that puzzle. The top of file comment it produced was simply amazing. All I gave it was the 1st 3 lines, ALL the rest were done by Claude:
/**************************************************************************************
LineSolverCmdLine.cpp Copyright 2014-2026 Crystal Canyon Computing, LLC. All Rights Reserved.
---------------------------------------<< o >>-----------------------------------------
Heart of the solver. Implements the Hard Left / Hard Right algorithm for a single
line (row or column), extended with constraint-narrowing from already-solved cells.
Algorithm:
1. pushLeft(): for each clue, find the earliest valid start position, treating
solved-empty cells as hard barriers.
2. pushRight(): for each clue, find the latest valid start position, same barriers.
3. Narrow bounds: if only one clue can cover a solved filled cell, tighten that
clue's left/right bounds and propagate ordering constraints.
4. Overlap: cells in [rightStarts[i], leftStarts[i]+realLen-1] are definitely filled.
5. Empty: cells outside every clue's reachable range [leftStarts[i], rightStarts[i]+realLen-1]
are definitely empty.
The graphic of the puzzle and the annotated Claude session are at http://theviews.org/Life%20at%20the%20Views/2026/march-20-2026-claude-pro-nonogram-part-3.html (there is no advertising on the site, so you can't really call that link "clickbait")
Damn!! - another one to read AFTER I finish reading the book!
MCS writing about Claude and code is a refreshing change from his political posts.
OK, here's 2 linked questions for Claude. Perhaps an aficianado can obtain the answer(s).
1. What were the likely geopolitical implications of the then President Bush using the word 'crusade' when launching Operation Enduring Freedom in 2001, and how intentional was the use of that word?
2. What ARE the possible geopolitical implications of the US Government using two Jewish emissaries to conduct what many see as bad faith negotiations immediately prior to the strikes on Iran?
... and a couple of follow up questions...
How would answers to these questions NOT feed into political biases of those humans asking them? Is is possible to provide answers that avoid our already prepared emotional responses?
Forgot to finish the comment and you can't edit anything here, so had to delete partial
I'm not going to pollute this place with political nonsense again.
I was so shocked by what Claude did that I just now realized that I hadn't told this forum WHY I was Shocked.
Claude didn’t just write this code, Claude:
1. Wrote the Code
2. Wrote the makefile
3. Ran the C++ compiler to produce the application
4. Used my test case (And figured out how to use my XML parsing code! This is incredible!)
5. Saw that the test case wasn’t solved
6. Instrumented the test case (I still can’t believe it did this!)
7. Analyzed the results of the instrumentation and found the bug
8. Fixed the bug
9. Found that the first fix hadn’t completely solved the problem (ApplyCapping is really complicated)
10. Fixed the bug in a general way.
11. Ran the test and found that it passed.
This is earth shaking!
You can see the annotated terminal transcript for yourself http://theviews.org/Life%20at%20the%20Views/2026/march-20-2026-claude-pro-nonogram-part-3.html
(1) Dr Brin samples Claude after condemning its advocate as 'evil'; and
(2) Dr Brin admits that ChatGPT engages in rather 'unctuous' flattery'.
These admissions are remarkable because these analyses were initially prompted by those who receive the bulk of Dr Brin's abuse here, MCS & myself, mostly because Dr Brin thrives on the very criticism that he hates the most, as there is nothing more stupefying then endless praise.
That's reason enough to stick around & disagree with our fine host as much as possible, since growth springs from ideological conflict, and all those political jabs just add icing to the cake.
Best
______
I agree with Treebeard when he says that Trump "gets history to move in the opposite direction of his stated intentions". I'd only add that Trump does this so consistently that this cannot entirely due to stupidity or accident, but the result of deliberate deceit & reverse psychology.
Or, is it now the lefty liberal contention that Trump is a remarkably honest & forthright man?
Computational plagiarism on a planetary scale: sure it's clever, but is it A.I.?
I regret saying, "Calculemus!" so often.
An author friend of mine is in the copyright lawsuit against Anthropic.
Dr. Brin, are you in it?
BTW My borrowing of the words of Kipling and Leibniz was intended in jest (I usually don't give obvious references as it clogs up an erudite discussion like CB). Also, it's tedious to wrestle with Gemini to lookup them up because the asinine inferences and flat-out errors just make things worse.
I tried Claude for the hell of it. I sometimes do math problems when I’m bored (I blame my education), and I gave it one I couldn’t solve, and it confidently told me: 1) it’s a classic problem, 2) no known clean closed form exists, 3) the numerical answer will converge toward .2117.
I ran the trial generator it provided, and it soon converged toward .400.
And it turns out, after some googling, that someone has solved it in closed form and it’s exactly 2/5. This Claude is pretty dumb. People are gonna make a lot of mistakes and get dumber if they start relying on AI instead of their own brains.
If it’s not in the training set, the “AI” can’t give it as answer. If you’re coding to solve a new problem that you don’t even have a test for, there isn’t any way Claude is going to be any help.
@David: I encourage you to read this paper: https://arxiv.org/pdf/2602.11865 that addresses many of the issues and topics you raise in the book. The paper describes a delegation framework for enabling humans and AI "agents" to hand off tasks to humans and AI "sub-agents" safely and effectively.
The framework runs across 5 core pillars:
→ Dynamic Assessment: agents evaluate each other's real-time capabilities before delegating
→ Adaptive Execution: mid-task re-delegation when things go wrong
→ Structural Transparency: every step is auditable, not a black box
→ Scalable Market Coordination: decentralized bidding systems where agents compete for tasks and reputation (karma)
→ Systemic Resilience: hard stops before cascading failures hit
They borrowed concepts from aviation and medicine to explain why AI agents blindly follow bad instructions the same way junior doctors don't question senior surgeons.
And they built fixes for the issues.
They mapped it onto real protocols like MCP, A2A, and AP2 and showed exactly what's missing from each one, with code for the object structures.
Wild, fun! I am grateful to live in this time when we have these toys. It is the Sci-Fi future about which I dreamed more than 50 years ago.
I have distilled the concepts into a set of markdown files for the proactive "agent" I am developing on nullclaw. My agent recently asked for an API key and budget to use openrouter.ai so that it can select the least expensive and most propitious models for its sub-agents. I did not know what openrouter.ai was before it asked.
Mitchell W. That sounds very interesting. I cannot tell much from you brief summary but i'd like to track it. Right now way too busy (keynoting a track of the big RSA conference) But one thing occurs to me. The structure you describe still needs incentives for accountability. And such incentives will only work (as shown by 4 billion years of evolution) if entities have individuality. I think you will find that part of my book interesting .
Here is a discussion between Senator Bernie Sanders and Claude:
https://youtu.be/h3AtWdeu_G0?is=e1han6mbBCQm2Rc3
I refuse to call LLMs AIs; instead I call them APIs, for 'artificial psedo-intelligence'. They act as the average of the intelligence of Internet websurfers, and we all know by now that the average Internet websurfer is STOOPID.
Heinlein defined an 'alien' as 'a being who thinks as well as a man but not like a man'. This might describe present APIs, if humans are in fact pseudo-intelligent; but Heinlein's unconsciously sexist use of the word 'man' implies that women are aliens.
pseudo
Have you wondered how the Iranians are they able to close the Straits of Hormuz after the navy and air force have been destroyed?
Maybe for the same reason that Ukraine (which has no navy at all) has been able to drive the Russian Black Sea fleet off the water.
On the battlefields of Ukraine and Iran we have seen a paradigm shift in warfare the likes of which has not been seen since the British introduced the tank in WW1, or gunpowder in the late Middle Ages.
Desert Storm showed what precision munitions could do.
Ukraine and Iran has shown what happens when precision munitions become plentiful, easy to mass produce by unskilled labor, as cheap as electronics from Radio Shack and can be hurled at enemies in mass swarms that are impossible to intercept.
So destroying the Iranian navy and air force doesn't mean shit.
Netanyahu and Trump are fighting the last war.
And so we never had a chance of winning this war.
One of the most fascinating characteristics of this current paradigm shift to cheap drones is how much it resembles the paradigm shift from armor plated knights of the high middle ages and renaissance when faced with cheap longbows and even cheaper hand guns.
Guns were even cheaper because, unlike the longbow, the user did not need years of training and muscle development to light the fuse of a matchlock firearm.
(Archeologists exploring the sunken wreck of Henry VIII's flagship, HMS Marie Rose, found the skeleton of a longbowmen assigned to the crew. His spine, clavicle and right arm bones right were all bent, thickened and distorted from years of practice with the longbow and its 180 lbs pull.)
By contrast, any smelly peasant with no extra muscle, a few days training and unit drill could fire a simple matchlock that propelled a lead ball through the thickest plate armor.
Classic case of the superiority of investing in capital over labor.
Desert Storm precision munitions was our longbow, devastating against Saddam's tanks and aircraft - but very expensive.
Iranian and Ukraine drones are the new matchlock firearms, cheap and plentiful and equally deadly.
In response to these cheap and plentiful new weapons, the soon to be obsolete knights became ever more extreme with heavier and more elaborate plate armor for themselves and their horses.
The armor was beautifully designed and precisely manufactured to maximize mobility, an armor plated knight could literally do cartwheels.
But that didn't stop knights from becoming useless on the battlefield.
Today, our behemoth Abram's tanks represent the peak of tank engineering and our F-35 is a high tech wonder that borders on science fiction.
Both are now as obsolete as a knight in shining armor killed the battle of Agincourt (1415) by a longbow or killed by an arquebus matchlock handgun at the Battle of Cerignola (1503).
Warfare has changed, and the Ukrainians and Iranians are ahead of the curve.
"Heinlein's unconsciously sexist use of the word 'man' implies that women are aliens."
Homer: "It's like you're from Venus."
Marge: "And you're from Mars."
Homer: "Oh, sure! I get the one with all the monsters."
Another possible historical parallel is the Bronze Age Collapse. According to good old Claude, some of its causes were:
Collapse of Chariot Warfare
The Late Bronze Age military system was built around expensive chariot forces, which required vast economic networks to maintain — horses, specialized craftsmen, bronze for weapons, and extensive logistics. When those networks broke down, so did the dominant military technology. The rise of cheaper infantry tactics (possibly including early iron weapons) may have made the old elite chariot armies obsolete or unaffordable.
Disruption of Bronze Supply Chains
Bronze itself was a strategic vulnerability. Tin (a key component) had to travel enormous distances — from Afghanistan, Anatolia, or even Britain. When trade networks collapsed (for whatever reason), armies could no longer reliably equip themselves, leaving civilizations militarily weakened and unable to defend against raiders.
Internal Rebellions and Civil Conflict
Several palace economies show signs of internal military unrest. The Hittite Empire, for example, was already stressed by wars with Assyria and internal succession struggles before the final collapse. Weakened central states were less able to project military force or hold peripheral territories.
Cascading Military Failures
Perhaps most importantly, these factors fed on each other. A raid that disrupted a trade route weakened an army's supply of bronze, which made it harder to defend against the next raid, which caused refugees to migrate and pressure neighboring states — a cascading military and political failure across the whole system.
The Bronze Age Collapse is a classic example of how military vulnerability is rarely just about battlefield strength — it's deeply tied to economics, logistics, and the stability of interconnected systems.
This all sounds kinda familiar right about now, no? And that's not even mentioning the invasion by the "Sea Peoples", with parallels to modern immigrant invasions. What will this one be called, the "Information Age Collapse"?
Re: electronics from Radio Shack and "Information Age Collapse"
Any future for sapiens sapiens relies on pervasive, non-doctrinal (DIY), Computational Thinking. We're close. I've been working on this for decades, and was greatly encouraged by "Why Johnny Can't Code" (WJCC).
I would add:
While the administration and their billionaire handlers try to obliterate independent sources of information, like web archives, they are dominating the traditional and "social" media ecosystem.
We are leaving the information age and entering the disinformation age.
Um you are commenting on my book's title?
Bronze age collapse had those factors but was above all ecological fragility.
Drones etc ARE disruptive. But Beam weapons will in turn disrupt them.
I am in SanFran at the big RSA conference. My keynote went well. Signed a lot of ailien minds
Interesting Bronze Age Collapse factoid - both the Exodus and the Trojan war occurred in the same century and were part of the collapse.
Why Johnny Can't Code. Jeepers SO many things I've urged in the past that might’ve done some real good if acted on back then, that are utterly troglodytic now! Yet, I leep trying. Sigh.
It is likely Trojan War and Exodus preceded the bronze age collapse... but not by an awful lot. Again the BAC was probably largely ecological / climate pummeling of a fragile network. And it answers the idiots who are now saying: "Okay okay, climate change is real and huge... but we are adaptable!" Um, how about you NOT wager my children on your polyanna story?
Troglodytic. Perhaps.
But ask yourself if we had more democracy and agency when everyone from plumbers to farmers to physicists could write their own BASIC programs, or did we make more progress by sifting eye candy for 40 years?
Another topic: any insights about all the 'The Thorne Universe' AI videos coming out, specifically are they done with his blessing? They're often educational (esp. the Alpha Centauri system ones), but there's just so much AI mystery meat out there right now.
I'd rather not gamble with the kids, but I think we are already in that territory. Horse. Open barn door. Etc.
I'm sure humanity will survive another supply chain shock, but I doubt our civilization will. That's how I view the BAC.
On a positive note, I think it will take something truly large to be that shock, though. Mix the heavy stress of climate change with a whopping big coronal mass ejection aimed right at us. That might do the trick.
Post a Comment