Will at least 16 AI x-risk researchers move to doomsday bunkers in the Southern Hemisphere by 2026-01-01?
21
1kṀ8152026
13%
chance
1D
1W
1M
ALL
This question is managed and resolved by Manifold.
Get
1,000 to start trading!
Sort by:
predictedYES
@mariopasquato The pressure on humanity to eventually build aligned ASI doesn't stop with a nuclear war. It will still be essential to have as many safety-oriented researchers working on the problem as possible.
And I suspect there is so little AI x-risk-oriented intellectual capital in locations likely to survive a nuclear war today that even $10M on a properly located and stocked compound + AI lab with a couple dozen high-calibre rationalists (e.g. Nate Soares) could increase the probability of existential security by a factor over 1/20, given a nuclear war.
Related questions
Related questions
Contingent on AI being perceived as a threat, will humans deliberately cause an AI winter before 2030?
33% chance
Will one of these AI researchers claim we're in an AI winter before 2026?
8% chance
Will an AI Doomer turn to violence by the end of 2026?
31% chance
Will someone kill an AI researcher to save the world by the end of 2025?
6% chance
Will anyone commit terrorism in order to slow the progression of AI before 2026?
29% chance
Will there be an AI Winter by the end of 2025?
10% chance
What will be the average P(doom) of AI researchers in 2025?
20% chance
Are we about to hit another AI winter in 2025?
7% chance
Are we about to hit another AI winter in 2026?
13% chance
Does an AI disaster kill at least 100 people before 2029?
74% chance