If the human population decreases by at least 10% in any period of 5 or fewer consecutive calendar years, which we will define here as a "global catastrophe", resolves to the principal cause. Otherwise, resolves to "no global catastrophe".
See https://www.metaculus.com/questions/1493/global-population-decline-10-by-2100/ and https://www.metaculus.com/questions/2568/ragnar%25C3%25B6k-seriesresults-so-far/ for more precise definitions of these categories and for related forecasts.
If multiple global catastrophes occur, resolves based on the first.
See also:
How does the climate one resolve in the event of comorbidity where both boneheaded government policies and climate change contribute to reducing agricultural output and without either one of those people would not have starved
@JonathanRay If there is a 10% decline, it resolves to whatever the "principal cause" is. In a case like you described, generally there should be one that is more important than the other. I will resolve the same as Metaculus if possible.
@JonathanRay You seem to have bet them to the same price, so you're saying you think it's ~0% likely? I would put it at >20%.
@jack I only bought this down to 20% versus 16% for the human extinction one https://manifold.markets/JamesDillard/will-ai-wipe-out-humanity-before-th
@evergreenemily I think for a disaster killing 10% of humans, non-AI risks are substantial. But climate, nuclear, etc are very unlikely to cause extinction, whereas AI is substantially more likely. The forecasts at https://www.metaculus.com/questions/2568/ragnar%25C3%25B6k-seriesresults-so-far/ align with this reasoning.
@jack Outright extinction, I agree. But runaway climate change could, in a worst-case scenario, kill 10% of humans in a relatively short period of time. In particular, climate-caused ecological/agricultural collapse followed by famines, especially in densely populated areas, could kill hundreds of millions or even billions of people.
Is it likely? No, I think it's pretty unlikely. But there's way more than a 3% chance of it happening by 2100, and that's what tne percentage was at when I checked this market this morning.
(Edit: FWIW, I think Metaculus underestimates the risks of a 10%-of-humans-die catastrophe from climate change, but is accurate in giving human extinction as a result of climate change an extremely low chance by 2100.)
@evergreenemily I think reasonable people can disagree about the probability, I personally think the ~3% predicted by Metaculus is about right. I think climate impacts are most likely to be spread out over time, not concentrated to reach 10% population loss in 5 years - it's possible but very unlikely. As https://astralcodexten.substack.com/p/the-extinction-tournament points out,
It’s unclear whether anything in recorded history would qualify; Genghis Khan’s hordes and the Black Plague each killed about 20% of the global population, but both events were spread out over a few decades.
@jack I agree that reasonable people can disagree! I'd definitely put the odds of "climate change kills 10% of humans over the course of 10-20 years" significantly higher than the odds of "climate change kills 10% of humans over the course of five years;" but I'd still put the latter at around 5-10%.
Note that this can be arbitraged with a ton of the markets in https://manifold.markets/group/ai-doom
Specifically, chance that AI wipes out humanity is < chance that AI kills 10% of humanity.
Yes, this is a pretty typical result I have observed in many other prediction contexts, and I think there are several reasons for it:
There's a classic https://en.wikipedia.org/wiki/Conjunction_fallacy where asking "Will AI kill X% of humans" sometimes gets a higher prediction than "Will anything kill X% of humans".
Prediction markets can be arbitraged to correct this fallacy, but the ROI for arbitraging a market that resolves in years or decades is so low that it's not strongly enforced by the market actors.
The people predicting/trading on AI x-risk are different than the people predicting/trading on non-x-risk questions. And, even when it's the same people, the questions tend to elicit different modes of thought that may result in irrational inconsistent predictions. It's particularly noteworthy that people betting YES on x-risk cannot collect their winnings on resolution, so the market incentives are fundamentally different than "normal" prediction markets.
How is the fragility of industrial ag. not in the list (I'm betting 'Other')...
I don't think it falls directly under "climate" but it depends on whether you consider direct or indirect causes, either way in that scenario we will blame climate when it does contribute to gigafamines.
@parhizj I think food crises that we believe stemmed from climate change - if we believe they would not have occurred in a counterfactual world with little or no global warming - would count as "Climate". But if we believe the food crises would have occurred regardless, that would count as "Other"
I think this is a reasonable definition for what qualifies as "Climate" in this question:
The question results positively if a human-made climate change or geoengineering catastrophe occurs that claims at least 10% in any period of 5 years or less before 2100. This catastrophe must be due to the direct effects of climate change that are generally believed very unlikely in a counterfactual world with little or no global warming but otherwise similar to ours. These include (but are not limited to) floods, extreme weather, the spreading of infectious disease, and the health effects of extreme heat. Finally, the effects due to the effects of the use of geoengineering that has been principally motivated to mitigate climate change risks, also count towards the population decline.
@jack It might be hard to differentiate lower yields from climate and how we are treating soil like dirt. (Mainly I think because there won't be a control, as regenerative agriculture is such a small minority to be useless for statistics).