We're Not Platonists, We've Just Learned The Bitter Lesson
Seen on Twitter: Intelligence explosion arguments don’t require Platonism. They just require intelligence to exist in the normal fuzzy way that all concepts exist. First, I’ll describe what the normal way concepts exist is. I’ll have succeeded if I convince you that claims using the word “intelligence” are coherent and potentially true. Second, I’ll argue from humans and animals for why I think these coherent-and-potentially-true things are actually true. Third, I’ll argue that so far this has been the most fruitful way to think about AI, and people who try to think about it differently make worse AIs. Finally, I’ll argue this is sufficient for ideas of “intelligence explosion” to be coherent. 1: What’s The Normal Way That Concepts Exist?Concepts are bundles of useful correlations. Consider the claim “Mike Tyson is stronger than my grandmother”. This doesn’t necessarily rely on a Platonic essence of Strength. It just means things like:
Each of these might decompose into even more sub-sub-claims. For example, the last one might decompose into:
We don’t really distinguish all of these claims in ordinary speech because they’re so closely correlated that they’ll probably all stand or fall together. These subclaims within “strength” are well-correlated, but not perfectly correlated. Is Mike Tyson stronger or weaker than some other very strong person like Arnold Schwarzenegger? It might be that Tyson could win at boxing but Schwarzenegger could lift more weights, or that Tyson has better bicep strength but Scharzenegger has better grip strength. Still, the correlations are high enough that “strength” is a useful shorthand that saves time / energy / cognitive load over always discussing every subclaim individually. In fact, there are a potentially infinite number of subclaims (could Mike Tyson lift an alligator more quickly than my grandmother while standing on one foot in the Congo rainforest?) so we have to use shorthands like “strength” to have discussions in finite time. If somebody learned that actually arm strength was totally uncorrelated with grip strength, and neither was correlated with ability to win fights, then they could fairly argue that “strength” was a worthless concept that should be abandoned. On the other hand, if somebody tried to argue that Mike Tyson was objectively stronger than Arnold Schwarzenegger, they would be reifying the concept of “strength” too hard, taking it further than it could go. But absent this kind of mistake, “strength” is useful and we should keep talking about it. “Intelligence” is another useful concept. When I say “Albert Einstein is more intelligent than a toddler”, I mean things like:
…and so on. Just as we can’t objectively answer “who is stronger, Mike Tyson or Arnold Schwarzenegger?”, we can’t necessarily answer “who is smarter, Einstein or Beethoven?”. Einstein is better at physics, Beethoven at composing music. But just as we can answer questions like “Is Mike Tyson stronger than my grandmother”, we can also answer questions like “Is Albert Einstein smarter than a toddler?” 1.1: Why Is A Concept Like Strength Useful?Why is someone with more arm strength also likely to have more leg strength? There are lots of specific answers, for example:
A more general answer might be that arm muscles are similar enough to leg muscles, and linked enough by being in the same body, that overall we expect their performance to be pretty correlated. 2: Do The Assumptions That Make “Intelligence” A Coherent Concept Hold?All human intellectual abilities are correlated. This is the famous g, closely related to IQ. People who are good at math are more likely to be good at writing, and vice versa. Just to give an example, SAT verbal scores are correlated 0.72 with SAT math scores. These links don’t always hold. Some people are brilliant writers, but can’t do math to save their lives. Some people are idiot savants who have low intelligence in most areas but extremely high skill in one. But that’s what it means to have a correlation of 0.72 instead of 1.00. In different situations, it can be surprising how much everything is correlated and how little everything is correlated. For example, Garry Kasparov, former chess champion, took an IQ test and got 135. You can think of this two different ways:
I think both of these are good takeaways. Compare the 0.72 verbal/math correlation with the 0.76 dominant-hand/non-dominant hand grip strength correlation and I think intelligence is a useful concept in the same way strength is. But also, humans are better at both the SAT verbal and the SAT math than chimps, cows, or fish. And GPT-4 is better at both those tests than GPT-3 or GPT-2. It seems to be a general principle that people, animals, or artifacts who are better at the SAT math are also better at the SAT verbal. 2.1: Why Is A Concept Like Intelligence Useful?Across different humans, skill at different kinds of intellectual tasks are correlated. Partly this is for prosaic reasons, like:
But these skills are also correlated for more fundamental reasons. Variation in IQ during adulthood is about 70% genetic. A lot of this seems to have to do with literal brain size, which is correlated with intelligence at about 0.2, and with myelination of neurons, which is hard to measure but apparently important. These considerations become even more relevant when you start comparing different species. Humans are smarter than chimps in a wide variety of ways (although not every way!) Likewise, chimps are smarter than cows, cows are smarter than frogs, et cetera. Research by Suzana Herculano-Houzel and others suggests this is pretty directly a function of how many neurons each animal has, with maybe some contribution from how long a childhood they have to learn things. Some animals are very carefully specialized for certain tasks - for example, birds can navigate very well by instinct. But when this isn’t true, their skill at a wide range of cognitive abilities depends on how big their brain is, how much of a chance they have to learn things, and how efficient the interconnections between brain regions are. Just as there are biological reasons why arm strength should be related to leg strength, there are biological reasons why different kinds of intellectual abilities should be related to each other. 3: Intelligence Has Been A Very Fruitful Way To Think About AIAll through the 00s and 10s, belief in “intelligence” was a whipping boy for every sophisticated AI scientist. The way to build hype was to accuse every other AI paradigm except yours of believing in “cognition” or “intelligence”, a mystical substance that could do anything, whereas your new AI paradigm realized that what was really important was emotions / embodiment / metaphor / action / curiosity / ethics . Therefore, it was time to reject the bad old AI paradigms and switch to yours. This announcement was usually accompanied by discussion of how belief in “intelligence” was politically suspect, or born of the believer’s obsession with signaling that they were Very Intelligent themselves. In Rodney Brooks’ famous term, they were “computational bigots”. When they weren’t proposing new anti-intelligence paradigms, the Responsible People were emphasizing how rather than magical intelligence, AI would need slow, boring work on the nitty gritty of computation - things like linguists applying their expert domain knowledge to natural language processing, or neuroscientists who really understood how the visual cortex worked slowly designing algorithms for robot vision. The idea of “intelligence” was a mirage that you could skip all of this hard work, just magic in a lump of “intelligence” and solve problems without understanding them. It was like thinking a car was made of “horsepower”, and as long as you crammed enough horsepower (whatever that was) into the front hood, you didn’t need to worry about complicated stuff like pistons or spark plugs or catalytic converters. In the middle of a million companies pursuing their revolutionary new paradigms, OpenAI decided to just shrug and try the “giant blob of intelligence” strategy, and it worked. They’re not above gloating a little; when they wanted to prove GPT-4 could understand comics, this was the comic they chose: Computer scientist Richard Sutton calls this the Bitter Lesson - that extremely clever plans to program “true understanding” into AI always do worse than just adding more compute and training data to your giant compute+training data blob. It’s related to Jelinek’s Law, named after language-processing AI pioneer Frederick Jelinek: “Every time I fire a linguist, the performance of the speech recognizer goes up.” The joke is that having linguists on your team means you’re still trying to hand-code in deep principles of linguistics, instead of just figuring out how to get as big a blob of intelligence as possible and throw language at it. The limit of Jelinek’s Law is OpenAI, who AFAIK didn’t use insights from linguistics at all, and so made an AI that uses language near-perfectly. Why does this work so well? Because animal intelligence (including human intelligence) is a blob of neural network arranged in a mildly clever way that lets it learn whatever it wants efficiently. The bigger the blob and the cleverer the arrangement, the faster and more thoroughly it learns. Trying to second-guess your blob by hand-coding stuff in doesn’t work as well (it might work for evolution sometimes, like the bird migration thing, but it probably won’t work for you). The bigger your blob, the cleverer its arrangement, and the more training data you give it, the better it’s likely to perform on a very wide variety of cognitive tasks. This explains why chimps are smarter than cows, why Einstein is smarter than you, and why GPT-4 is smarter than GPT-2. The correlations won’t be perfect, any more than strength correlations are perfect. But they’ll be useful enough to talk about. I think if you get a very big blob, arrange it very cleverly, and give it lots and lots of training data, you’ll get something that’s smarter than humans in a lot of different ways. In every way? Maybe not: humans aren’t even smarter than chimps in every way. But smarter in enough ways that the human:chimp comparison will feel appropriate. 3.1: This Is Enough For An “Intelligence Explosion”Suppose that someone put steroids in a box with a really tight lid. You need to be very strong to get the lid off the box. But once you do, you can take the steroids. This is enough to cause a “strength explosion”, in the sense that there’s some amount of strength that lets you become even stronger. It’s not a very interesting example. But that’s an advantage! People tend to get all mystical and philosophical about this stuff. I think the best way to think about it is with the commonest of common sense. There have already been intelligence explosions. Long ago, humans got smart enough to invent reading and writing, which let them write down training data for each other and become even smarter (in a practical sense; this might or might not have raised their literal IQ, depending on how you think about the Flynn Effect). Later on, we invented iodine supplementation, which let us treat goiter and gain a few IQ points on a population-wide level. There’s nothing mystical about any of this. Once you get smart enough, you can do things that make you even smarter. AI will be one of those things. We already know that bigger blobs of compute with more training data can do more things in correlated ways - frogs are outclassed by cows, chimps, and humans; village idiots are outclassed by Einstein; GPT-2 is outclassed by GPT-4. At some point we might get a blob which is better than humans at designing chips, and then we can make even bigger blobs of compute, even faster than before. Ten years ago, I asked people to Beware Isolated Demands For Rigor. When you’re dealing with topics in Far Mode, it’s tempting to get all philosophical - what if matter doesn’t exist? What if everything’s an illusion? Instead, I recommend thinking about future intelligence explosions in Near Mode, in which superintelligent machines are no philosophically different than machines that are very very big. You're currently a free subscriber to Astral Codex Ten. For the full experience, upgrade your subscription. |
Older messages
Open Thread 286
Sunday, July 23, 2023
...
Your Book Review: The Laws of Trading
Friday, July 21, 2023
Finalist #10 in the Book Review Contest
Highlights From The Comments On British Economic Decline
Thursday, July 20, 2023
...
The Extinction Tournament
Thursday, July 20, 2023
...
Berkeley Meetup On Sunday, Special Guest Philip Tetlock
Wednesday, July 19, 2023
...
You Might Also Like
UW and computer science student reach truce in ‘HuskySwap’ spat
Saturday, January 11, 2025
Blue Origin set for first orbital launch | Zillow layoffs | Pandion shutdown | AI in 2025 ADVERTISEMENT GeekWire SPONSOR MESSAGE: GeekWire's special series marks Microsoft's 50th anniversary by
Cryptos Surrender Recent Gains | DOJ's $6.5 Billion Bitcoin Sale
Saturday, January 11, 2025
Bitcoin and other tokens retreated as Fed signaled caution on rate cuts. Forbes START INVESTING • Newsletters • MyForbes Presented by Nina Bambysheva Staff Writer, Forbes Money & Markets Follow me
Just Buy a Balaclava
Saturday, January 11, 2025
Plus: What Raphael Saadiq can't live without. The Strategist Every product is independently selected by editors. If you buy something through our links, New York may earn an affiliate commission.
Up in Flames
Saturday, January 11, 2025
January 11, 2025 The Weekend Reader Required Reading for Political Compulsives 1. Trump Won't Get the Inauguration Day He Wanted The president-elect is annoyed that flags will be half-staff for
YOU LOVE TO SEE IT: Biden’s Grand Finale
Saturday, January 11, 2025
Biden drills down on offshore drilling, credit scores get healthier, social security gets a hand, and sketchy mortgage lenders are locked out. YOU LOVE TO SEE IT: Biden's Grand Finale By Sam Pollak
11 unexpected things you can put in the dishwasher
Saturday, January 11, 2025
(And 7 things you should keep far away from there) View in browser Ad The Recommendation January 11, 2025 Ad 11 things that are surprisingly dishwasher-safe An open dishwasher with a variety of dishes
Weekend Briefing No. 570
Saturday, January 11, 2025
Black Swan Threats in 2025 -- Why Boys Don't Go To College -- US Government's Nuclear Power Play ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
Your new crossword for Saturday Jan 11 ✏️
Saturday, January 11, 2025
View this email in your browser Take a mental break with this week's crosswords: We have six new puzzles teed up for you this week. Play the latest Vox crossword right here, and find all of our new
Firefighters Make Progress, Water Rankings, and Ohio St. Wins
Saturday, January 11, 2025
Multiple wildfires continued to burn in Southern California yesterday, with officials reporting at least 10 deaths. Over 10000 homes across 27000 acres have burned, and 20 suspected looters have been
☕ So many jobs
Saturday, January 11, 2025
So why did stocks fall? January 11, 2025 View Online | Sign Up | Shop Morning Brew Presented By Indacloud Good morning. It's National Milk Day, the one day of the year you're allowed to skim