Seeing Like a Language Model
<table><tr><td><img alt="Chain of Thought" src="https://d24ovhgu8s7341.cloudfront.net/uploads/publication/logo/59/small_chain_of_thought_logo.png" /></td><td></td><td><table><tr><td>by <a href="https://every.to/@danshipper" itemprop="name">Dan Shipper</a></td></tr><tr><td>in <a href="https://every.to/chain-of-thought">Chain of Thought</a></td></tr></table></td></tr></table><figure><img src="https://d24ovhgu8s7341.cloudfront.net/uploads/post/cover/3776/full_page_cover_seeing_liek_ai.png"><figcaption>Midjourney/Every illustration.</figcaption></figure><p><em>Last week <u><a href="https://every.to/on-every/every-s-master-plan-part-ii" rel="noopener noreferrer" target="_blank">I wrote</a></u> that we’d be publishing a few excerpts from a book I’m writing about the worldview I’ve developed by w
Last week I wrote that we’d be publishing a few excerpts from a book I’m writing about the worldview I’ve developed by writing, coding, and living with AI. Here’s the first piece, about the differences between the old (pre-GPT-3) worldview and the new.—Dan Shipper
When I say the word “intelligent,” you probably think of being rational. But language models show us that this assumption is wrong.
To us in the West, smarts are about being able to explicitly lay out what you know and why you know it. For us, the root of intelligence is logic and reason to ward off superstition and groupthink; it is clear and concise definitions to eradicate vague and wooly-headed thinking; it is formal theories that explain the hidden laws of the world around us—simple, falsifiable, and parsimonious yet general enough to tie together everything in the universe from atoms to asteroids.
Our romantic picture of ourselves as “rational animals,” as Aristotle said, has produced everything in the modern world—rockets, trains, medicines, computers, smartphones.
But this picture is incomplete. It contains a huge blind spot: It neglects the fundamental importance of intuition in all of intelligent behavior—intuition which is by nature ineffable; that is to say, not fully describable by rational thought or formal systems.
How to build a thinking machine
The best way to understand what I’m talking about is to imagine trying to build a thinking machine. How would you do it?
Let’s start with a task, something easy and basic that humans do every day. Maybe something like scheduling an appointment.
Let’s say we’re a busy executive who gets the following appointment request:
New request
From: Mona Leibnis
Hey,
I’m available Monday at 3 p.m., Tuesday at 4 p.m., Friday at 6 p.m.
When can you meet?
We want to build a machine to intelligently schedule an appointment. How would we go about it?
We’d probably start by giving our machine a couple of rules to follow:
-
First, check available time slots on my calendar.
-
Then, compare my open slots to the open slots on the invitee’s calendar.
-
If you find one, add the appointment to the calendar.
That all seems pretty reasonable. You could definitely write a computer program to follow those rules. But there’s a problem: The rules we’ve specified so far can’t handle urgency or importance.
For example, consider a case where you desperately want to meet with someone and you’d be willing to move another appointment in order to make the time work. Now we have to introduce a new rule:
- If it is urgent that I meet with the invitee, you can reschedule a less urgent appointment in order to make the meeting happen sooner.
But this rule is incomplete because it introduces the concept of urgency without defining it. How do we know what’s urgent? Well, there must be some rules for that too. So we need to delineate them.
In order to measure urgency, we have to have some conception of the different people in your life—who your clients and potential clients are, and which clients are important or not.
Now things are starting to get hairy. In order to determine the relative importance of clients, we have to know about your business aims—and about which clients are likely to close, and which clients are likely to pay a lot of money, and which clients are likely to stay on for a long time. And don’t forget—which clients were introduced by an important friend whom you need to impress, so while they may not be directly responsible for a lot of revenue, they’re still a priority.
This is only a taste of the rules we’d have to define in order to build an adequate automatic scheduling system. And that’s just for dealing with calendars!
The problem that we’re finding is that it’s very hard to lay everything out explicitly—because everything is interconnected. To paraphrase the late astronomer Carl Sagan: If you wish to schedule a meeting from scratch, you must first define the universe.
The old, Western worldview
This approach—which seemed most natural to us—is the exact one that the first generation of artificial intelligence researchers took to try to build AI.
Create a free account to continue reading
The Only SubscriptionYou Need to Stay at the Edge of AI
The essential toolkit for those shaping the future
"This might be the best value youcan get from an AI subscription."
- Jay S.
Every Content
AI&I Podcast
Monologue
Cora
Sparkle
Spiral
Join 100,000+ leaders, builders, and innovators
Email address
Already have an account? Sign in
What is included in a subscription?
Daily insights from AI pioneers + early access to powerful AI tools
Front-row access to the future of AI
Bundle of AI software
Thanks for rating this post—join the conversation by commenting below.
Chain of Thought (Every.to)
https://every.to/chain-of-thought/seeing-like-a-language-modelSign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modellanguage modelneural networkAnthropic Responsible Scaling Policy v3: A Matter of Trust
Anthropic has revised its Responsible Scaling Policy to v3. The changes involved include abandoning many previous commitments, including one not to move ahead if doing so would be dangerous, citing that given competition they feel blindly following such a principle would not make the world safer. Holden Karnofsky advocated for the changes. He maintains that the previous strategy of specific commitments was in error, and instead endorses the new strategy of having aspirational goals. He was not at Anthropic when the commitments were made. My response to this will be two parts. Today’s post talks about considerations around Anthropic going back on its previous commitments, including asking to what extent Anthropic broke promises or benefited from people reacting to those promises, and how we
InkSF, an Opening on Finding the Highest Impact in AI Safety and Moving to SF
How can we actually minimize the odds that AI leads to catastrophic outcomes for all of us humans? This question has been rattling around my head for the last two months. The world might be ending. Nobody seems to care. The incentives are steaming us ahead. When I ask strangers on the street: “ How likely is it that superhuman [1] AI could become too powerful for humans to control?”, 78% say either "very likely" (51.6%) or "somewhat likely" (26.3%) [2] . My guess is AI capabilities spending is at least 20x the spending on ensuring AI leads to the flourishing of humans [3] . Moloch [4] is winning. So what can actually be done? As a toy example [5] : let’s say I currently think there is a 40% chance AI eventually goes extraordinarily bad for humanity. I could either: Try really hard to get l
March 2026 Links
Why We Have Prison Gangs : Q&A whose ultimate answer is that gangs are a form of governance in a place that has little. Skarbek also talks about what being in a gang is like, rules they have in place (bedtime, taxes, no affiliating with sex offenders or former LEOs), similarities. Plane Crash : Delian gives a play-by-play of his plane engine cutting off mid-flight, culminating in him crash landing on a golf course. The lessons learned extended elsewhere, namely where else did he simply say "I'll do it later", when later wasn't guaranteed? Sirat is not about the end of the world : A great perspective on Sirat that contains spoilers. Everything you ever wanted to know about Roblox, but were afraid to ask a 12-year-old Maybe there's a pattern here? : Technology, no matter what its original pu
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models
InkSF, an Opening on Finding the Highest Impact in AI Safety and Moving to SF
How can we actually minimize the odds that AI leads to catastrophic outcomes for all of us humans? This question has been rattling around my head for the last two months. The world might be ending. Nobody seems to care. The incentives are steaming us ahead. When I ask strangers on the street: “ How likely is it that superhuman [1] AI could become too powerful for humans to control?”, 78% say either "very likely" (51.6%) or "somewhat likely" (26.3%) [2] . My guess is AI capabilities spending is at least 20x the spending on ensuring AI leads to the flourishing of humans [3] . Moloch [4] is winning. So what can actually be done? As a toy example [5] : let’s say I currently think there is a 40% chance AI eventually goes extraordinarily bad for humanity. I could either: Try really hard to get l
Apex Fintech Solutions and Wavvest Announce Strategic Partnership to Deliver AI-Powered Financial Planning - FF News | Fintech Finance
<a href="https://news.google.com/rss/articles/CBMi3wFBVV95cUxNUTN6SXpDWDVSYldwVllMVThlMG51TVFkakZ0ZGVUMmpMSUswZ3V5QzBTVHRMcGU4Y3hFSEl1cmN6WVpUWnRoOW9Ob1VsaXBzUFEtV0hrcWJhbE9DODBnUldfV3hyQ0tZb1prOXNSM0pUa3ZZeG1Gd3hzUThsMXdPZVhOR1RUUHY4V1d3dGhhZHp0NkVYZ1hxZzRUR1NpNWhkQ0RKSlIxem51M0hvYk5GcW1qTHlJYWhGOTdnSmJQY0tSSmtCU1NrTHVuZHJOcmd2TndRMVRVMlJraVBTdGd3?oc=5" target="_blank">Apex Fintech Solutions and Wavvest Announce Strategic Partnership to Deliver AI-Powered Financial Planning</a> <font color="#6f6f6f">FF News | Fintech Finance</font>
Double PyTorch Inference Speed for Diffusion Models Using Torch-TensorRT | NVIDIA Technical Blog - NVIDIA Developer
<a href="https://news.google.com/rss/articles/CBMiqwFBVV95cUxQZmN5cmtxaXgzbzUzVk1pT1JtY1dTOFh5ZUNZTmx2eDAwc2JUY29OczJSVXJfaDFxZkl5eWR2eUh6MlZyZGZDVG44b3RQdk1FS0ZLdC00cXRoUGFFVzR2Wng1YXFOZ0NWUE02ZFpidW1KNTYtYmppWk1zaHFjdGNqalFQTkdRaTdERl9HNmhrRTU5UmFPTGdSbEYzYjhhdGdvNnpEUnlKUWlFLWc?oc=5" target="_blank">Double PyTorch Inference Speed for Diffusion Models Using Torch-TensorRT | NVIDIA Technical Blog</a> <font color="#6f6f6f">NVIDIA Developer</font>
Pixel super-resolved virtual staining of label-free tissue using diffusion models - Nature
<a href="https://news.google.com/rss/articles/CBMiX0FVX3lxTE52Ry0wOHRWSzVXSG8wQkx2WUxzRFQzd2FQUHlqYmhFRVhEUDUwdTBVUlExR21GT0E0VG4wT191ZHZPUXVwWjFmTzh4Y3diTTMyd1NYWm9wajRLNUJlSDFv?oc=5" target="_blank">Pixel super-resolved virtual staining of label-free tissue using diffusion models</a> <font color="#6f6f6f">Nature</font>
Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!