We're Not Platonists, We've Just Learned The Bitter Lesson
Seen on Twitter: Intelligence explosion arguments don’t require Platonism. They just require intelligence to exist in the normal fuzzy way that all concepts exist. First, I’ll describe what the normal way concepts exist is. I’ll have succeeded if I convince you that claims using the word “intelligence” are coherent and potentially true. Second, I’ll argue from humans and animals for why I think these coherent-and-potentially-true things are actually true. Third, I’ll argue that so far this has been the most fruitful way to think about AI, and people who try to think about it differently make worse AIs. Finally, I’ll argue this is sufficient for ideas of “intelligence explosion” to be coherent. 1: What’s The Normal Way That Concepts Exist?Concepts are bundles of useful correlations. Consider the claim “Mike Tyson is stronger than my grandmother”. This doesn’t necessarily rely on a Platonic essence of Strength. It just means things like:
Each of these might decompose into even more sub-sub-claims. For example, the last one might decompose into:
We don’t really distinguish all of these claims in ordinary speech because they’re so closely correlated that they’ll probably all stand or fall together. These subclaims within “strength” are well-correlated, but not perfectly correlated. Is Mike Tyson stronger or weaker than some other very strong person like Arnold Schwarzenegger? It might be that Tyson could win at boxing but Schwarzenegger could lift more weights, or that Tyson has better bicep strength but Scharzenegger has better grip strength. Still, the correlations are high enough that “strength” is a useful shorthand that saves time / energy / cognitive load over always discussing every subclaim individually. In fact, there are a potentially infinite number of subclaims (could Mike Tyson lift an alligator more quickly than my grandmother while standing on one foot in the Congo rainforest?) so we have to use shorthands like “strength” to have discussions in finite time. If somebody learned that actually arm strength was totally uncorrelated with grip strength, and neither was correlated with ability to win fights, then they could fairly argue that “strength” was a worthless concept that should be abandoned. On the other hand, if somebody tried to argue that Mike Tyson was objectively stronger than Arnold Schwarzenegger, they would be reifying the concept of “strength” too hard, taking it further than it could go. But absent this kind of mistake, “strength” is useful and we should keep talking about it. “Intelligence” is another useful concept. When I say “Albert Einstein is more intelligent than a toddler”, I mean things like:
…and so on. Just as we can’t objectively answer “who is stronger, Mike Tyson or Arnold Schwarzenegger?”, we can’t necessarily answer “who is smarter, Einstein or Beethoven?”. Einstein is better at physics, Beethoven at composing music. But just as we can answer questions like “Is Mike Tyson stronger than my grandmother”, we can also answer questions like “Is Albert Einstein smarter than a toddler?” 1.1: Why Is A Concept Like Strength Useful?Why is someone with more arm strength also likely to have more leg strength? There are lots of specific answers, for example:
A more general answer might be that arm muscles are similar enough to leg muscles, and linked enough by being in the same body, that overall we expect their performance to be pretty correlated. 2: Do The Assumptions That Make “Intelligence” A Coherent Concept Hold?All human intellectual abilities are correlated. This is the famous g, closely related to IQ. People who are good at math are more likely to be good at writing, and vice versa. Just to give an example, SAT verbal scores are correlated 0.72 with SAT math scores. These links don’t always hold. Some people are brilliant writers, but can’t do math to save their lives. Some people are idiot savants who have low intelligence in most areas but extremely high skill in one. But that’s what it means to have a correlation of 0.72 instead of 1.00. In different situations, it can be surprising how much everything is correlated and how little everything is correlated. For example, Garry Kasparov, former chess champion, took an IQ test and got 135. You can think of this two different ways:
I think both of these are good takeaways. Compare the 0.72 verbal/math correlation with the 0.76 dominant-hand/non-dominant hand grip strength correlation and I think intelligence is a useful concept in the same way strength is. But also, humans are better at both the SAT verbal and the SAT math than chimps, cows, or fish. And GPT-4 is better at both those tests than GPT-3 or GPT-2. It seems to be a general principle that people, animals, or artifacts who are better at the SAT math are also better at the SAT verbal. 2.1: Why Is A Concept Like Intelligence Useful?Across different humans, skill at different kinds of intellectual tasks are correlated. Partly this is for prosaic reasons, like:
But these skills are also correlated for more fundamental reasons. Variation in IQ during adulthood is about 70% genetic. A lot of this seems to have to do with literal brain size, which is correlated with intelligence at about 0.2, and with myelination of neurons, which is hard to measure but apparently important. These considerations become even more relevant when you start comparing different species. Humans are smarter than chimps in a wide variety of ways (although not every way!) Likewise, chimps are smarter than cows, cows are smarter than frogs, et cetera. Research by Suzana Herculano-Houzel and others suggests this is pretty directly a function of how many neurons each animal has, with maybe some contribution from how long a childhood they have to learn things. Some animals are very carefully specialized for certain tasks - for example, birds can navigate very well by instinct. But when this isn’t true, their skill at a wide range of cognitive abilities depends on how big their brain is, how much of a chance they have to learn things, and how efficient the interconnections between brain regions are. Just as there are biological reasons why arm strength should be related to leg strength, there are biological reasons why different kinds of intellectual abilities should be related to each other. 3: Intelligence Has Been A Very Fruitful Way To Think About AIAll through the 00s and 10s, belief in “intelligence” was a whipping boy for every sophisticated AI scientist. The way to build hype was to accuse every other AI paradigm except yours of believing in “cognition” or “intelligence”, a mystical substance that could do anything, whereas your new AI paradigm realized that what was really important was emotions / embodiment / metaphor / action / curiosity / ethics . Therefore, it was time to reject the bad old AI paradigms and switch to yours. This announcement was usually accompanied by discussion of how belief in “intelligence” was politically suspect, or born of the believer’s obsession with signaling that they were Very Intelligent themselves. In Rodney Brooks’ famous term, they were “computational bigots”. When they weren’t proposing new anti-intelligence paradigms, the Responsible People were emphasizing how rather than magical intelligence, AI would need slow, boring work on the nitty gritty of computation - things like linguists applying their expert domain knowledge to natural language processing, or neuroscientists who really understood how the visual cortex worked slowly designing algorithms for robot vision. The idea of “intelligence” was a mirage that you could skip all of this hard work, just magic in a lump of “intelligence” and solve problems without understanding them. It was like thinking a car was made of “horsepower”, and as long as you crammed enough horsepower (whatever that was) into the front hood, you didn’t need to worry about complicated stuff like pistons or spark plugs or catalytic converters. In the middle of a million companies pursuing their revolutionary new paradigms, OpenAI decided to just shrug and try the “giant blob of intelligence” strategy, and it worked. They’re not above gloating a little; when they wanted to prove GPT-4 could understand comics, this was the comic they chose: Computer scientist Richard Sutton calls this the Bitter Lesson - that extremely clever plans to program “true understanding” into AI always do worse than just adding more compute and training data to your giant compute+training data blob. It’s related to Jelinek’s Law, named after language-processing AI pioneer Frederick Jelinek: “Every time I fire a linguist, the performance of the speech recognizer goes up.” The joke is that having linguists on your team means you’re still trying to hand-code in deep principles of linguistics, instead of just figuring out how to get as big a blob of intelligence as possible and throw language at it. The limit of Jelinek’s Law is OpenAI, who AFAIK didn’t use insights from linguistics at all, and so made an AI that uses language near-perfectly. Why does this work so well? Because animal intelligence (including human intelligence) is a blob of neural network arranged in a mildly clever way that lets it learn whatever it wants efficiently. The bigger the blob and the cleverer the arrangement, the faster and more thoroughly it learns. Trying to second-guess your blob by hand-coding stuff in doesn’t work as well (it might work for evolution sometimes, like the bird migration thing, but it probably won’t work for you). The bigger your blob, the cleverer its arrangement, and the more training data you give it, the better it’s likely to perform on a very wide variety of cognitive tasks. This explains why chimps are smarter than cows, why Einstein is smarter than you, and why GPT-4 is smarter than GPT-2. The correlations won’t be perfect, any more than strength correlations are perfect. But they’ll be useful enough to talk about. I think if you get a very big blob, arrange it very cleverly, and give it lots and lots of training data, you’ll get something that’s smarter than humans in a lot of different ways. In every way? Maybe not: humans aren’t even smarter than chimps in every way. But smarter in enough ways that the human:chimp comparison will feel appropriate. 3.1: This Is Enough For An “Intelligence Explosion”Suppose that someone put steroids in a box with a really tight lid. You need to be very strong to get the lid off the box. But once you do, you can take the steroids. This is enough to cause a “strength explosion”, in the sense that there’s some amount of strength that lets you become even stronger. It’s not a very interesting example. But that’s an advantage! People tend to get all mystical and philosophical about this stuff. I think the best way to think about it is with the commonest of common sense. There have already been intelligence explosions. Long ago, humans got smart enough to invent reading and writing, which let them write down training data for each other and become even smarter (in a practical sense; this might or might not have raised their literal IQ, depending on how you think about the Flynn Effect). Later on, we invented iodine supplementation, which let us treat goiter and gain a few IQ points on a population-wide level. There’s nothing mystical about any of this. Once you get smart enough, you can do things that make you even smarter. AI will be one of those things. We already know that bigger blobs of compute with more training data can do more things in correlated ways - frogs are outclassed by cows, chimps, and humans; village idiots are outclassed by Einstein; GPT-2 is outclassed by GPT-4. At some point we might get a blob which is better than humans at designing chips, and then we can make even bigger blobs of compute, even faster than before. Ten years ago, I asked people to Beware Isolated Demands For Rigor. When you’re dealing with topics in Far Mode, it’s tempting to get all philosophical - what if matter doesn’t exist? What if everything’s an illusion? Instead, I recommend thinking about future intelligence explosions in Near Mode, in which superintelligent machines are no philosophically different than machines that are very very big. You're currently a free subscriber to Astral Codex Ten. For the full experience, upgrade your subscription. |
Older messages
Open Thread 286
Sunday, July 23, 2023
...
Your Book Review: The Laws of Trading
Friday, July 21, 2023
Finalist #10 in the Book Review Contest
Highlights From The Comments On British Economic Decline
Thursday, July 20, 2023
...
The Extinction Tournament
Thursday, July 20, 2023
...
Berkeley Meetup On Sunday, Special Guest Philip Tetlock
Wednesday, July 19, 2023
...
You Might Also Like
GeekWire's Most-Read Stories of the Week
Sunday, November 24, 2024
Catch up on the top tech stories from this past week. Here are the headlines that people have been reading on GeekWire. ADVERTISEMENT GeekWire SPONSOR MESSAGE: Get your ticket for AWS re:Invent,
13 Things That Delighted Us Last Week: From Daschund Bags to Sparkly Toilet Seats
Sunday, November 24, 2024
Plus, the Gucci poker set that Jennifer Tilly packs in her carry-on. The Strategist Logo Every product is independently selected by editors. If you buy something through our links, New York may earn an
LEVER WEEKLY: Trump's Cabinet Of Curiosities
Sunday, November 24, 2024
Opening up Trump's corruption-riddled cabinet and more from The Lever this week. LEVER WEEKLY: Trump's Cabinet Of Curiosities By The Lever • 24 Nov 2024 View in browser View in browser This is
What our travel expert brings on every trip
Sunday, November 24, 2024
M&Ms? View in browser Ad The Recommendation Ad Traveling is stressful for everyone, even travel writers Various travel gear items laid out on a yellow background. Michael Hession/NYT Wirecutter
☕ The Brew’s Holiday Gift Guide
Sunday, November 24, 2024
What to get everyone in your family... Presented By Bose November 24, 2024 | View Online | Sign Up | Shop Sunny Eckerle NOTE FROM THE WRITERS Good morning! Cassandra and Matty here, Morning Brew's
How Friendsgiving became America's favorite made-up holiday
Sunday, November 24, 2024
Plus: The real story behind FX's "Say Nothing," the horrifying effects of air pollution in South Asia, and more. November 25, 2024 View in browser Friendsgiving is just what America
'The most serious telecom hack in our history'
Saturday, November 23, 2024
Elon Musk's problem with Microsoft | Can you lie to an AI chatbot? ADVERTISEMENT GeekWire SPONSOR MESSAGE: Get your ticket for AWS re:Invent, happening Dec. 2–6 in Las Vegas: Register now for AWS
Bitcoin Nears $100,000 | Ledger’s Big Break
Saturday, November 23, 2024
A historic rally fueled by Trump's crypto agenda pushes bitcoin to new heights. Forbes START INVESTING • Newsletters • MyForbes Nina Bambysheva Staff Writer, Forbes Money & Markets Follow me on
The New MASTER PLAN
Saturday, November 23, 2024
Our second season will expose another hidden plot that has brought our world to the brink of collapse.
Guest Newsletter: Five Books
Saturday, November 23, 2024
Five Books features in-depth author interviews recommending five books on a theme Guest Newsletter: Five Books By Sylvia Bishop • 23 Nov 2024 View in browser View in browser Five Books features in-