Will AI accelerators improve in FLOPs/watt by 100x of an NVidia H100 by 2033?
Basic
17
Ṁ33632033
90%
chance
1D
1W
1M
ALL
Compared to an H100, will tensor TFLOPs/Watt improve by 100x by 2033? AI accelerators in scope for this question must be deployed significantly - with at least 100k units or $100M (in 2024 dollars) in production, and have published perf/watt numbers.
This market will count peak FLOPs/watt at k bits of precision, adjusted by a factor of 2^(1 - 32/k). That is, 16 bit precision counts 1/4 as much as 32 bit, which counts 1/4 as much as 64 bit precision.
This question is managed and resolved by Manifold.
Get
1,000
and3.00
Sort by:
Related questions
Related questions
Will a machine learning training run exceed 10^25 FLOP in China before 2025?
77% chance
Will software-side AI scaling appear to be suddenly discontinuous before 2025?
24% chance
When will a US government AI run overtake private AI compute by FLOP?
Will 2024 be the year when AI capabilities progress from AI hardware scaling hits a wall?
26% chance
Will OpenAI report that it's no longer heavily GPU limited in 2024?
16% chance
Will a machine learning training run exceed 10^26 FLOP in China before 2025?
15% chance
What will be the maximum achievable flop utilization on the next generation of Nvidia server chips?
Will the Groq chip inspire Nvidia/AMD to produce radically new AI chips before 2025?
15% chance
100GW AI training run before 2031?
37% chance
Will AGI be powered by Nvidia GPUs?
61% chance