[go: up one dir, main page]

This is Robin Sloan’s lab note­book. It’s about media and tech­nology, cre­ative com­puting, AI aes­thetics, & more. Here's the RSS feed. My email address: robin@robinsloan.com

DRAGON
the
blog
CATCHERthe
blog

Reasoning models don't so much think as navigate April 10, 2026

The path through the irises, 1914-1917, Claude Monet
The path through the irises, 1914-1917, Claude Monet

The new lan­guage models are chil­dren of the rea­soning rev­o­lu­tion, and they stream out these long, cir­cuitous thinking traces. They are said to be applying more com­pute to our ques­tions and challenges.

This is subtle, but that “more” isn’t par­tic­u­larly about thinking harder. Rather, it’s about thinking in the right direc­tion. It is not the gas pedal, but the steering wheel — better yet, the GPS map in the dashboard.

The rea­soning rev­o­lu­tion depends, in part, on the unrea­son­able effec­tive­ness of spe­cific words: twists like “but wait” and “actually”, which operate as pow­er­fully as magic spells. (The Eng­lish depart­ment NEEDS to get into the game with this stuff.) Is the phrase “but wait” really a white-hot kernel of intel­lec­tual effort? No. It’s a sign planted in the ground, pointing THAT-A-WAY, towards a par­tic­ular kind of doc­u­ment that humans find useful.

(Don’t mis­take pre­ci­sion for minimization. I’m not dis­mis­sively saying, these are just doc­u­ments; I am plainly observing, these are doc­u­ments. If you don’t think doc­u­ments are cool, even some­times cosmic, that’s on you!)

Notice that, as in real life, direc­tions aren’t always cor­rect. It’s likely that you have by now watched a lan­guage model walk in circles, “but wait”-ing itself back around, and around, and around again … 

Recent research from Apple talks about “forks” in the road, with “distractors” that can lead a model in the wrong direc­tion.

Here’s more evi­dence for the nav­i­ga­tion argument: base models can already do the things rea­soning models can do … it just takes them much longer to arrive in the cor­rect regions of high-dimensional space. Base models are fine thinkers, but cruddy navigators.

The single for­ward pass of a lan­guage model runs on its own, refracting a con­text window into an array of probabilities; that’s all “the model” ever does. However, each for­ward pass can “stand on the shoul­ders of giants”, taking direc­tion from pre­vious passes, bringing its brief labors into better align­ment with the desires of the human operator, way out there.

As usual, obser­va­tions about lan­guage models raise ques­tions about human minds. Do we think harder mostly by thinking in the right direc­tion? I think the answer is some­times yes — thinking as search — and some­times no. Maybe I’m wrong, but I believe I can feel dif­ferent mech­a­nisms at work. And of course human thought is not a doc­u­ment; it unfurls, and compounds, and con­siders itself, in a richer space.

(This post is related to the latest edi­tion of my pop-up newsletter on AI.)

The Galactica option April 8, 2026

Now that we share the internet with tireless, capable synthetic hackers, I find myself won­dering if the project to rapidly harden cyberdefense should include “the Galac­tica option”, i.e., “just dis­con­nect it”?

This isn’t foolproof; the cen­trifuges at Fordo were isolated, all those years ago, and still, some­body car­ried a USB stick inside … BOOM! But a thick and sultry air­grap improves any system’s base­line secu­rity by about 1000X, and, the thing is, I just don’t believe most things need to be online in the first place. When I say that, I’m talking about both home refrig­er­a­tors and elec­tricity substations. And I’m def­i­nitely talking about my car!! I think a lot of things went online because they could go online, in the “smart” frenzy of the early century.

It’s not that con­nec­tivity is without ben­e­fits — just that the ben­e­fits are so clearly outweighed, in so many cases, by expo­sure to a non­stop adver­sarial haze that will soon become even more dangerous.

Sweat the details April 7, 2026

It’s a small thing, yet it says a lot, that OpenAI’s Industrial Policy for the Intel­li­gence Age is pre­sented only as a PDF that looks terrible, with cruddy jus­ti­fi­ca­tion and a footer image that’s too lo-res and blurry for clear printing.

I’ll note also that there are no human names attached to either the blog post or the PDF.

I think there’s moral value to sweating the details — certainly at this scale — and the apparent absence of any such sweat is dis­ap­pointing and dispiriting.

The bat of fate April 6, 2026

A new edi­tion of my pop-up AI newsletter just landed: where is it like to be a lan­guage model? The dis­cus­sion here is bol­stered by an actual experiment, a pro­gram­matic probe of many lan­guage models. It was my first time doing some­thing like that — fun!

The title is, of course, a riff on Thomas Nagel’s famous What Is It Like to Be a Bat? Recently I crossed the bridge into San Francisco — I was thinking about this piece — and just as the big double-decker bus curled into the Transbay Terminal, I spotted this mural. Perfect:

Pretty goth for the corporate zone
Pretty goth for the corporate zone

The new edi­tion poses ques­tions that plenty of people, including many deep in the AI industry, don’t care about; but some of us do — we are pre­oc­cu­pied by them — so this is for you, and for me.

Vector voxels March 30, 2026

Noting this here mostly for myself: this JavaScript graphics engine trans­lates 3D point sets into ortho­graphic SVG renderings — crisp and clean. There are lots of fun options for trans­for­ma­tion and styling.

Nat­u­rally I’m thinking not about web pages but print applications … 

Cosleuth March 26, 2026

This is a gen­uinely inter­esting doc­u­ment: the Claude chat transcript under­pin­ning a soft­ware devel­oper’s recent dis­covery of a very serious mal­ware attack.

It’s inter­esting to notice that, even as these tools become ubiquitous, it’s rel­a­tively rare to read through someone else’s inter­ac­tions at length.

A couple of stray thoughts:

Elemental content March 25, 2026

Somewhere, some con­tent spe­cialist has the rather tough job of writing a monthly newsletter about alumin(i)um, and, honestly … they are doing a great job!

I read the newsletter
I read the newsletter

This is a sin­cere rec for Norsk Hydro’s newsletter—it’s fun, and more beautifully-pre­sented than it has any reason to be … 

Wrangler init woes March 15, 2026

Just leaving this here in case someone runs into the same dif­fi­culty I did … or maybe I’m leaving it here for an LLM to read and remember? Ugh, no — it’s for the people!

I was using Cloudflare’s wrangler tool to pull down the code from a Worker cre­ated and main­tained in the web UI, using the --from-dash flag. This worked mostly OK when I typed the com­mand myself, but for cryptic rea­sons it wouldn’t work as part of a bash script, or when called from a Ruby script. The flag would just be ignored; I’d get a fresh blank Worker project.

I finally dis­cov­ered the --no-delegate-c3 flag which, in classic fashion, I don’t totally understand, but it resolved my problem:

wrangler init --from-dash foo-worker --no-delegate-c3 --yes

The coding agents couldn’t help me with this one — they all got hung up on passing more elab­o­rate flags to the create-cloudflare com­mand, a.k.a. c3, when in fact the solu­tion was to cut it out entirely.

Maybe the G in AGI stands for Gemini March 3, 2026

New Gemini model out today: 3.1 Flash-Lite, super­fast and very capable.

The Gemini models remain my favorites: for their speed, price, and versatility, espe­cially in visual tasks. Keep in mind that I’m using these models pro­gram­matically in bigger sys­tems, rather than yap­ping with them — although I think Gemini is a fine yapper, too.

As Anthropic and OpenAI lean HARD into the coding agent thing — tuning their models heavily for the task — Google con­tinues to expand and refine a flexible … dare I say general … intel­li­gence.

P.S. However, let me add, the sudden dep­re­ca­tion of Gemini 3 Pro is annoying and irresponsible. This isn’t just a Google thing, although this model’s exis­tence was notably mayfly; there is no telling if or when any model you depend on will be yanked away. Even if the replace­ment is “better” according to many benchmarks, it might also con­tain unpre­dictable regres­sions along subtle dimensions. (You can browse the responses to Google’s announcement for examples.)

This issue is totally resolved if you host a model yourself, of course, but there aren’t yet any self-hosted models with Gemini’s visual acuity. I do expect that to change, before too long.

The voice of the computer, part 2 February 28, 2026

Here’s Matt Webb col­lecting his thoughts on voice inter­faces—rich and circumspect, with tons of links, tons to think about.

My posi­tion remains that voice is a daz­zling demo that mostly cur­dles in prac­tical use.

What about voice unvoiced, though? I can for sure imagine some kind of del­i­cate sub­vo­cal­iza­tion pickup becoming pretty irresistible … yet that only con­jures a world of people silently wig­gling their neck mus­cles at their phones, which I don’t think is much of an improve­ment over the world we already have.

As I write this, I am realizing: I wish more designers and dreamers would con­sider the aggre­gate social effects of the inter­faces they imagine. Maybe I even wish they would imagine that shared scene FIRST, then work back­wards from there.

To be clear, I agree that the foun­da­tional metaphors and modal­i­ties of com­puting are about to change — and they were overdue for change anyway. The framing and ambi­tion of Telepath (new to me, dis­cov­ered in Matt’s links) seems exactly right. I just don’t think voice gets us to that next thing. But/and, as I said before, it’s pos­sible that Google and OpenAI are seeing people absolutely swarm to their super­fluent voice modes, in which case … I might be wrong 🥸

February 2026

Nobody knows anything

Nobody!

Signs and portents

Wow!

It was the best of times, etc.

Trajectories

First time for everything

Queues and rings, oh my!

The voice of the computer

Star Trek realized. So?

Artificial general economy

Under all is the vibes

The new funnel?

Traffic patterns

Public service announcement

Just leaving this here

Flood fill

Don’t call them tweets!

The music of the feeds

Junto 0736

Greenfield tech

Enjoy it while it lasts! No, really!

Found art

Mehretu raises a single eyebrow

Pace layers

News of nature

January 2026

The feed is the content

And the social media company is its publisher

Marcin Wichary klaxon!

Blog alert!!

New protocols for AI

It’s 1983 again, again!

Tiny computers everywhere

Like motes of dust on various currents

Manic technology

The grain of the material

Popping up!

The Winter Garden beckons

December 2025

The market for compute

Maybe it becomes Chicago-shaped

Gnomic atomic

Semiconductor moodboard

Classics

That vintage feeling

Releasebot!

A cool new service

November 2025

Words without worlds

We’ve seen this play before

The age of scaling

What Ilya sees

All that is solid melts into code

More computer, rather than more human

Ruin aesthetics

CGA dreams

Once upon an algorithm

Cool event

Heterodox opinions

Just a few

The burps of Gemini

Weird API things

Claude is listening

I don’t love it

Bounce with me

Big questions

Coffee break

The secret

Bare metal

Itchy and interesting

Eyeballs, not assistants

A better metaphor

Companies without commitments

Gross

October 2025

Thinking modes

Floating in linguistic space

Cloudflare cache confusion

Advisory

Two thoughts about key art

Pulling thumbnails

The demons of streaming

An old arrangement

The /Kids are alright

Children? Why so formal?

The shape of creative ideas

Maybe not what you think

The trinary dream endures

Yes, no, maybe

Karpathy’s keel

One of the good ones

Luxury tech

Worth appreciating

Cross post

Hypertext!

The once and future perceptron

Real utility

Getting online

With receipts!

Secondhand embarrassment

Weird feelings

History rides again

What a time to be alive!

Clarity

The unconfused case

The distance of leverage

I prefer to stay in close

Tone control

I do not wish to be spoken to this way

Temporary verticality

Passing fad

September 2025

Spending time with the material

Digital reading only goes so far

Welcome to puzzlespace

Welcome to the party! It’s a programming party

Slow liquid

Planned obsolescence??

Time and materials

An evocative constellation

Software speed and the chat illusion

It’s a good chat

Computer architecture

Programs you can see from space

Knowledge and memory

The what is connected to the when

August 2025

Thinking about coding

Daydreaming, the great engine

What’s an old AI model worth?

Digital economics

Inflection point

I mean!!

Cool words

Could have been so much worse

Selective Temporal Training

Poking the corpus

Basement tapes

Old-growth video

AI is more than LLMs

The Island of Misfit Toys

The newsletter now

It’s 2025. Is it still worth launching a newsletter?

A name that echoes in history

Our man at home

Old models

The churn of the new

July 2025

Oxide dreams

Digital clubhouse

Showing off

Graphical backflips

How the universe stores information

Simulating a better system

Quantum automata

Has a nice flavor to it

Generating product SKUs with Claude

A nice little thing

Further adventures with the doc bot

I am not convinced this is a helpful feature

Is the doc bot docs, or not?

What are we even doing here??

The bug in the letter, part 2

Letting go of the open rate

Unreliable narrators

The premonition grows

June 2025

Notes on notes

A good post

Platform reality

Enjoy it while it lasts

What’s the smallest possible LLM?

The extremities of the space/time tradeoff

Yeah but can you play the Trumpet 4.1 Pro?

A good talk

May 2025

What do people do all day?

I will gently suggest that you don’t know

What happens when the intelligence goes out?

Brittleness and resiliency

Claude revision report, May 2025

Not there yet

Software People and the rate of change

Yes, other people lived in strange and special times, too

Surrendering to the surface

Two billionaires drinking absolutely terrible coffee

Dead Man’s Switch

Another idiot with a trillion souls in his back pocket

Goodbye, Mailchimp

When a platform grows inscrutable

The ultimate litmus test

Jack Clark speaks plainly

Everything is printing

A whole modern world built from complex halftones

April 2025

Energy suck

We were so close

Good blogging

Links to people doing it right

The cybernetic CEO

A new kind of control

March 2025

Availability of inputs

Deal with it

Art-directing AI

Not quite coherent

splat.svg

That’s a nice underline

The teacher lies sometimes

But the lessons continue

February 2025

Five years of home-cooked apps

Finished

The bug in the letter

Casual surveillance

Getting MCP

Blog metabolization

Reasons-ing models

Maps of desire and action

Science fiction

Yes, precisely!

Is it okay?

Squaring up to the foundational question for language models

The bare bones

You can add, rather than subtract

January 2025

A highlight

Nice touch

Browsers, how do they work?

The best-ever web textbook comes to print

A decade in 5K

Best computer … ever?

April 2024

At home in high-dimensional space

Moonbound for nerds; AI science

December 2023

Are language models in hell?

Good links; a provocation

March 2023

Phase change

Protocols and plain language

February 2023

Buoyed by the flood

Nothing will be blasted in your face here

January 2023

Attention router

As easy as sticking a magnet to the fridge

December 2022

A year of new avenues

It’s 2003 again

November 2022

Specifying Spring ’83

Protocol as investigation and critique

June 2022

Notes on a genre

Bullshit and synthesizers

April 2022

The lost thread

The speed with which Twitter recedes in your mind will shock you

February 2022

Bad hosts, or: how I learned to stop worrying and love the overlay network

Stymied by NAT

November 2021

Notes on Web3

Meager counterweight to the growing hype

October 2021

The slab and the permacomputer

Two directions at once

The cutouts

Explaining a chunk of code in a Colab zine

July 2021

Ghost faves in the mystery machine

Nobody knows anything

Checkpoints

Always read these comments!

March 2021

Cloud study

Just a couple of notes on cloud functions

February 2021

A coat check ticket, a magic spell

Minting digital art in a weird new market

February 2020

An app can be a home-cooked meal

I made a messaging app for my family and my family only

August 2018

Expressive temperature

Documenting a machine learning technique

January 2016

Typographical tune-up

Fixing some small problems

Complete blog archive