This is Robin Sloan’s lab notebook. It’s about media and technology, creative computing, AI aesthetics, & more. Here's the RSS feed. My email address: [email protected]
DRAGON
the
blogCATCHERthe
blog
Here is Derek Thompson on AI and where it’s going:
[ … ] I am lucky to have participated in conversations about the future of AI with executives and builders at frontier labs, economists at AI conferences, AI investors, and other bigwigs at off-the-record dinners where important truths can theoretically be bandied about without risk. And if I had to pick three words to summarize this collective expert view of the future, I could not in a million years, or with a trillion tokens, find three words more suitable than these: Nobody knows anything.
This was my assertion in the first edition of my pop-up newsletter, but of course, one always wonders, is it just me who doesn’t know anything … ? Hearing Derek say it, too, makes me feel more confident.
I’m sure it won’t last long, but I just want to memorialize a strange event: on Tuesday, February 24, at approximately 8:10 p.m. Pacific Time … Google search … was DOWN!
I guess they Gemini’d too hard!
On his fabulous new blog, Marcin Wichary links to a post of mine, the one where I wrote:
What makes the AI chatbots and agents feel light and clean, here and now in 2026? Is it an innate architectural resistance to advertising, to attention hacks, to adversarial crud? No — it’s that they are simply new! The language models in 2026 are Google in 1999, Twitter in 2009. Their vast conjoined industry of influence hasn’t yet arisen … though it is stirring.
Marcin’s view is characteristically clear:
The AI community tends to say “this is the worst this will ever be” in response to criticism, but in a very learned sense, in many aspects it is also the best it will ever be.
This is useful: a sort of counter-mantra. The enthusiasts really do invoke “this is the worst AI will ever be” almost like a magic spell. Along the narrow technical dimension, the claim is certainly true … but the technical dimension is not the only one, nor even the most important. There are commercial and ecological dimensions, too, and we can look at every other digital system — literally every other digital system! — for clues about trajectories from worst to best, best to worst.
Your claw does okay on the internet of 2026. How will it fare on the internet of 2029, which will surely be crowded with adversaries — hacker claws and trickster claws? Recall that the hackers and tricksters are, as always, MUCH more motivated than you to field this software. Even if they are denied access to the tip-top models, their next-best instruments can compensate with volume and malevolence.
My strong hunch is that claw and counterclaw will net out into a sort of cruddy detente. See exhibit A: email. Or exhibit B: the World Wide Web.
That’s the thing about technology: everybody gets it!
My path as a programmer has mostly been guided by avoidance: of complexity, fragility, excessive moving parts. I like my programs simple, stateless, cool to the touch: mortar and pestle, not food processor.
So, it took thirty years … but I finally used a queue, and I finally used a ring. I am replicating data across three different buckets, and the environment imposes constraints that prevent me from doing what I’d normally do, which is load the data into memory and slam it into all three locations at once.
I can’t say I love the feeling of that queue out there … looming … growing … but it seems to be working. I’m proud of myself for figuring it out, even if I’m mostly just using Cloudflare’s ergonomic off-the-shelf offerings.
I’ve still never written a test, and I don’t plan to start!
Here is the most spectacular demo of present AI systems: pull out your phone, initiate Gemini’s live voice mode, say “please translate this conversation between English and Japanese”, and allow the system to act as a responsive and competent interpreter.
ChatGPT offers a mode like this, too; it’s clear Google and OpenAI have both invested a ton in these features, and that both believe they will represent a significant — THE significant? — interface to their models. Indeed, OpenAI’s upcoming devices are all premised on this.
Meanwhile, it seems odd to imagine “the voice of Claude”. I’m sure Anthropic could buy itself a voice mode, but this interface shouldn’t be mistaken as a language model sandwiched between voice recognition and TTS — it’s subtler and more fluent than that. I believe (though of course I could be wrong) these voice modes are the result of supplemental end-to-end training.
I’ve demoed Gemini’s live voice mode several times, always to great reception, yet never actually used it for a practical purpose. I mean: I was just in Japan for two weeks! And I don’t speak Japanese! And I never used it!
I still imagine I might — it is really sharp — but, for me, it has remained a dazzling demo, not a useful tool. The infinite bummer of “let’s both talk into my stupid phone” remains an insuperable barrier.
I’d love to know the usage statistics that Google and OpenAI are seeing; I’d love to know if and how people are really engaging with the voice of the computer, straight out of Star Trek.
Here’s a gimlet post from Benn Stancil, on this weird manic moment:
Vertigo is the sensation of spinning. It is feeling dizzy, even when the world is standing still. If every Anthropic press release is all we talk about, have the robots not already taken over? If every company is urgently rearranging itself around a workforce of agents, does it matter how well they score on the tests? Are we learning, by posting to our Substacks and reading the discourse, or are we becoming obsessed? Is this takeoff, or just takes?
Here’s an interesting and surprising nugget in a recent post by Justin Duke, he of Buttondown:
From both onboarding surveys and overall quantitative traffic: half of our customers come through word of mouth, one-fourth through search, and one-fourth through LLM-based traffic. (I’ll talk more about that last bucket in a later post.) When you divide that not by overall customer count but by revenue, it’s two-thirds word of mouth and one-third everything else.
I will be very curious to read that later post about the LLM traffic!
Justin’s 50% word of mouth, 25% search, 25% robo-brain feels both (1) very healthy, and (2) very futuristic — a characteristically good balance.
I’ll be away from the blog for a few weeks, so let me just pin this to the top here:
I still write my own code, because I enjoy it.
I have never been a normal programmer, and maybe you aren’t, either. So, let’s just do things the way we want — the way we like.
😌
A new edition of my pop-up newsletter has arrived: Flood fill vs. the magic circle, about AI automation and its limits.
Also, the actual email I transmit contains a compendium of don’t-call-them-tweets, i.e., stray thoughts adjacent to AI. These are fun to pull out of my notebook; consider them EXCLUSIVE CONTENT, and an enticement to subscribe.
Four more editions, then poof!
The latest Disquiet Junto assignment is out, and this blog provides a link in its creative chain. The assignment is:
Write a piece of music emulating the dopamine engine that is social media.
Click through for Marc’s instructions, which are, as usual, provocative and fun. I myself have responded to a grand total of one (1) Junto assignment; if I wasn’t headed out of town, I’d give this one a try, and double my output.
Enjoy it while it lasts! No, really!
Mehretu raises a single eyebrow
News of nature
January 2026
And the social media company is its publisher
Blog alert!!
It’s 1983 again, again!
Like motes of dust on various currents
The grain of the material
The Winter Garden beckons
December 2025
Maybe it becomes Chicago-shaped
Semiconductor moodboard
That vintage feeling
A cool new service
November 2025
We’ve seen this play before
What Ilya sees
More computer, rather than more human
Cool event
CGA dreams
Just a few
Weird API things
I don’t love it
Big questions
The secret
Itchy and interesting
A better metaphor
Gross
October 2025
Floating in linguistic space
Advisory
Pulling thumbnails
An old arrangement
Children? Why so formal?
Maybe not what you think
Yes, no, maybe
One of the good ones
Worth appreciating
Hypertext!
Real utility
With receipts!
Weird feelings
What a time to be alive!
The unconfused case
I prefer to stay in close
I do not wish to be spoken to this way
Passing fad
September 2025
Digital reading only goes so far
Welcome to the party! It’s a programming party
Planned obsolescence??
An evocative constellation
It’s a good chat
Programs you can see from space
The what is connected to the when
August 2025
Daydreaming, the great engine
Digital economics
I mean!!
Could have been so much worse
Poking the corpus
Old-growth video
The Island of Misfit Toys
It’s 2025. Is it still worth launching a newsletter?
Our man at home
The churn of the new
July 2025
Digital clubhouse
Graphical backflips
Simulating a better system
Has a nice flavor to it
A nice little thing
I am not convinced this is a helpful feature
What are we even doing here??
Letting go of the open rate
The premonition grows
June 2025
A good post
Enjoy it while it lasts
The extremities of the space/time tradeoff
A good talk
An interesting limitation for AI
May 2025
I will gently suggest that you don’t know
Brittleness and resiliency
Not there yet
Yes, other people lived in strange and special times, too
Two billionaires drinking absolutely terrible coffee
Another idiot with a trillion souls in his back pocket
When a platform grows inscrutable
Jack Clark speaks plainly
A whole modern world built from complex halftones
April 2025
We were so close
Links to people doing it right
A new kind of control
March 2025
Deal with it
Not quite coherent
That’s a nice underline
But the lessons continue
February 2025
Finished
Casual surveillance
Blog metabolization
Maps of desire and action
Yes, precisely!
Squaring up to the foundational question for language models
You can add, rather than subtract
January 2025
Nice touch
The best-ever web textbook comes to print
Best computer … ever?
April 2024
Moonbound for nerds; AI science
December 2023
Good links; a provocation
March 2023
Protocols and plain language
February 2023
Nothing will be blasted in your face here
January 2023
As easy as sticking a magnet to the fridge
December 2022
It’s 2003 again
November 2022
Protocol as investigation and critique
June 2022
Bullshit and synthesizers
April 2022
The speed with which Twitter recedes in your mind will shock you
February 2022
Stymied by NAT
November 2021
Meager counterweight to the growing hype
October 2021
Two directions at once
Explaining a chunk of code in a Colab zine
July 2021
Nobody knows anything
Always read these comments!
March 2021
Just a couple of notes on cloud functions
February 2021
Minting digital art in a weird new market
February 2020
I made a messaging app for my family and my family only
August 2018
Documenting a machine learning technique
January 2016
Fixing some small problems
Complete blog archive