Astral Codex Ten - Mantic Monday 4/18/22
WarcastingChanges in Ukraine prediction markets since my last post March 21:
If you like getting your news in this format, subscribe to the Metaculus Alert bot for more (and thanks to ACX Grants winner Nikos Bosse for creating it!) Nuclear Risk UpdateLast month superforecaster group Samotsvety Forecasts published their estimate of the near-term risk of nuclear war, with a headline number of 24 micromorts per week. A few weeks later, J. Peter Scoblic, a nuclear security expert with the International Security Program, shared his thoughts. His editor wrote:
In other words: the Samotsvety analysis was the best that domain-general forecasting had to offer. This is the best that domain-specific expertise has to offer. Let’s see if they line up: Superficially not really! In contrast to Samotsvety’s 24 micromorts, Scoblic says 370 micromorts, an order of magnitude higher. Most of the difference comes from two steps. First, conditional on some kind of nuclear exchange, will London (their index city for where some specific person worrying about nuclear risk might be) get attacked? Samotsvety says only 18% chance. Scoblic raises this to 65%, saying:
Second, what is the probability that an “informed and unbiased” person could escape a city before it gets nuked? Samotsvety said 75%; Scoblic said 30%. I think this is a fake disagreement. Some people I know were so careful that they had already left their cities by the time this essay was posted; the odds of this person escaping a nuclear attack are 100%. Other people are homebound, never watch the news, and don’t know there’s a war at all; the odds of these people escaping a nuclear attack are 0%. In between are a lot of different levels of caution; do you start leaving when the war starts to heat up? Or do you want until you hear that nukes are already in the air? Do you have a car? A motorcycle for weaving through traffic? Do you plan to use public transit? My guess is that the EAs who Samotsvety were writing for are better-informed, more cautious, and better-resourced than average, and the 75% chance they’d escape was right for them. Scoblic seems to interpret this question as saying that people have to escape after the nuclear war has already started, and his 30% estimate seems fine for that situation. If we halve Scoblic’s estimate (or double Samotsvety’s) to adjust for this “fake disagreement” factor - then it’s still 24 vs. 185 micromorts, a difference of 8x. What do we want - and what do we have the right to expect - from forecasting? If it’s order-of-magnitude estimates, it looks like we have one: we’ve bounded nuclear risk in the order of magnitude between 24 and 185 (at least until some third group comes around with something totally different than either of these two). Or maybe it’s a better understanding of our “cruxes” - where the real disagreement is that accounts for almost all of the downstream uncertainty. In this case, this exercise is pretty successful - everyone is pretty close to each other on the risk of small-scale nuclear war, and the big disagreement is over whether small-scale nuclear war would inevitably escalate. The Samotsvety team says they plan to meet, discuss Scoblic’s critiques, and see if they want to update any of their estimates. And they made what I consider some pretty strong points in the comments that maybe Scoblic will want to adjust on. Both sides seem to be treating this as a potential adversarial collaboration, and I’d be interested in seeing if this can bound the risk even further. AI Risk “Update”Everyone’s been talking about this Metaculus question: ”Weakly general AI” in the question means a single system that can perform a bunch of impressive tasks - passing a “Turing test”, scoring well on the SAT, playing video games, etc. Read the link for the full operationalization, but the short version is that this is advanced stuff AI can’t do yet, but still doesn’t necessarily mean “totally equivalent to humans in any way”, let alone superintelligence. For the past year or so, this had been drifting around the 2040s. Then last week it plummeted to 2033. I don’t want to exaggerate the importance of this move: it was also on 2033 back in 2020, before drifting up a bit. But this is certainly the sharpest correction in the market’s two year history. The drop corresponded to three big AI milestones. First, DALL-E2, a new and very impressive art AI. Second, PALM, a new and very impressive language AI: Third, Chinchilla, a paper and associated model suggesting that people have been training AIs inefficiently all this time, and that probably a small tweak to the process could produce better results with the same computational power. (there’s also this Socratic Models paper that I haven’t even gotten a chance to look at, but which looks potentially impressive) This raises the eternal question of “exciting game-changer” vs. “incremental progress at the same rate as always”. These certainly don’t seem to me to be bigger game changers than the original DALL-E or GPT-3, but I’m not an expert and maybe they should be. It’s just weird that they used up half our remaining AI timeline (ie moved the date when we should expect AGI by this definition from 20 years out to 10 years out) when I feel like there have been four or five things this exciting in the past decade. Or is there another explanation? A lot of AI forecasters on Metaculus are Less Wrong readers; we know that the Less Wrong Yudkowsky/Christiano debate on takeoff speeds moved the relevant Metaculus question a few percent: Early this month on Less Wrong, Eliezer Yudkowsky posted MIRI Announces New Death With Dignity Strategy, where he said that after a career of trying to prevent unfriendly AI, he had become extremely pessimistic, and now expects it to happen in a few years and probably kill everyone. This caused the Less Wrong community, already pretty dedicated to panicking about AI, to redouble its panic. Although the new announcement doesn’t really say anything about timelines that hasn’t been said before, the emotional framing has hit people a lot harder. I will admit that I’m one of the people who is kind of panicky. But I also worry about an information cascade: we’re an insular group, and Eliezer is a convincing person. Other communities of AI alignment researchers are more optimistic. I continue to plan to cover the attempts at debate and convergence between optimistic and pessimistic factions, and to try to figure out my own mind on the topic. But for now the most relevant point is that a lot of people who were only medium panicked a few months ago are now very panicked. Is that the kind of thing that moves forecasting tournaments? I don’t know. Shorts1: Will Elon Musk acquire over 50% of Twitter by the end of 2022? Why are these two so different? Do lots of people expect Musk to acquire Twitter after June 1 but still in 2022? 2: Will Marine Le Pen win the 2022 French presidential election? Beautiful correspondence, beautiful volume numbers. 3: Will cumulative reported deaths from COVID-19 in China exceed 50,000 by the end of 2022? You’re a free subscriber to Astral Codex Ten. For the full experience, become a paid subscriber. |
Older messages
Open Thread 220
Sunday, April 17, 2022
...
Deceptively Aligned Mesa-Optimizers: It's Not Funny If I Have To Explain It
Monday, April 11, 2022
A Machine Learning Monday post, 4/11/22
Open Thread 219
Sunday, April 10, 2022
...
Spring Meetups In Seventy Cities
Sunday, April 10, 2022
...
Dictator Book Club: Xi Jinping
Wednesday, April 6, 2022
...
You Might Also Like
Silicon Aristotle
Thursday, November 28, 2024
Who Can Claim Aristotle? // Private Chefs For Silicon Valley's Elite Silicon Aristotle By Caroline Crampton • 28 Nov 2024 View in browser View in browser Who Can Claim Aristotle? Edith Hall | Aeon
How the Pilgrims differed from the Puritans
Thursday, November 28, 2024
+ how to avoid awkwardness at Thanksgiving table
♻️ Gratitude & Joy flow in a cycle
Thursday, November 28, 2024
Fun stuff for you to click on curated with joy by CreativeMornings HQ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
How “Y.O.L.O. Joe” Can Beat The Lame Duck
Thursday, November 28, 2024
Here is what Democrats could actually achieve in the months before Trump takes office. Need a productive political topic to discuss at the Thanksgiving table? Want to impart key facts as you pass the
Trump Cabinet Bomb Threats, Ancient Sandwiches, and a Popsicle Caper
Thursday, November 28, 2024
Several of President-elect Donald Trump's Cabinet nominees and administration appointees faced bomb threats and "swatting" attacks on Tuesday and Wednesday. ͏ ͏ ͏ ͏ ͏ ͏ ͏ ͏
First-ever UEFI bootkit for Linux in the works, experts say [Thu Nov 28 2024]
Thursday, November 28, 2024
Hi The Register Subscriber | Log in The Register Daily Headlines 28 November 2024 KITTY LOOKS AT SCREEN AI GENERATED First-ever UEFI bootkit for Linux in the works, experts say Bootkitty doesn't
On My Mind: Fig Ornaments and Striped Bath Mats
Thursday, November 28, 2024
Plus: Eensy-weensy, teeny-tiny gifts. The Strategist Every product is independently selected by editors. If you buy something through our links, New York may earn an affiliate commission. November 27,
What It’s Like to Be on Trump’s Enemies List
Wednesday, November 27, 2024
Columns and commentary on news, politics, business, and technology from the Intelligencer team. Intelligencer power What It's Like to Be on Trump's Enemies List “Revenge does take time.” Photo-
GeekWire Mid-Week Update
Wednesday, November 27, 2024
Read the top tech stories so far this week from GeekWire Top stories so far this week Microsoft credited with spotting sophisticated Chinese hack that hit telecoms including T-Mobile US officials say a
Thursday Briefing: A fragile cease-fire in Lebanon
Wednesday, November 27, 2024
Plus, a post-election Thanksgiving. View in browser|nytimes.com Ad Morning Briefing: Asia Pacific Edition November 28, 2024 Author Headshot By Gaya Gupta Good morning. We're covering the first day