Before 2028, will there be enough inference capacity to generate 30T frontier model tokens per day?
Plus
6
Ṁ4442028
39%
chance
1D
1W
1M
ALL
In "Situational Awareness: The Decade Ahead", Leopold Aschenbrenner claims:
Another way of thinking about it is that given inference fleets
in 2027, we should be able to generate an entire internet’s worth of
tokens, every single day.
Resolves YES if by the end of 2027, there is enough deployed inference capacity to generate 30 trillion tokens in a 24-hour period using a combination of frontier models. "Frontier models" in the sense that GPT-4 is a frontier model today in mid-2024.
This is one of a series of markets on claims made in Leopold Aschenbrenner's Situational Awareness report(s):
Other markets about Leopold's predictions:
This question is managed and resolved by Manifold.
Get
1,000
and3.00
Related questions
Related questions
Will a OpenAI model have over 500k token capacity by the end of 2024.
50% chance
Will a new lab create a top-performing AI frontier model before 2028?
57% chance
Will anyone train a TokenFormer model at scale before 2026?
25% chance
100GW AI training run before 2031?
53% chance
Will OpenAI inference costs fall by 100x over the next 18 months?
33% chance
AI: Will someone train a $1B model by 2025?
67% chance
By March 14, 2025, will there be an AI model with over 10 trillion parameters?
63% chance
Will models be able to do the work of an AI researcher/engineer before 2027?
40% chance
Will a model costing >$30M be intentionally trained to be more mechanistically interpretable by end of 2027? (see desc)
57% chance
$1T AI training cluster before 2031?
73% chance