Will GPT-4 be trained on more than 10T text tokens?
36
1kṀ6012Jun 2
36%
chance
1H
6H
1D
1W
1M
ALL
If GPT-4 is multimodal, I will only include the subset of text tokens in this estimate.
Oct 10, 4:24pm: Will GPT-3 be trained on more than 10T text tokens? → Will GPT-4 be trained on more than 10T text tokens?
Added detail:
For the purposes of this question, only original tokens will be counted. That is, two passes do not double the token count.
This question is managed and resolved by Manifold.
Get
1,000 to start trading!
People are also trading
Related questions
Will there be an LLM (as good as GPT-4) that was trained with 1/100th the energy consumed to train GPT-4, by 2026?
83% chance
In yottaFLOPs (10^24), how much compute will GPT-4 be trained with?
22
GPT-5 trained with >=24k GPUs?
82% chance
Will GPT-5 have over 10 trillion parameters?
56% chance
Will it be possible to disentangle most of the features learned by a model comparable to GPT-4 this decade?
37% chance
How much compute will be used to train GPT-5?
Was GPT-4 trained in 4 months or less?
59% chance
Will GPT-5 have over 100 trillion parameters?
14% chance
Will a GPT-4 level system be trained for <$1mm by 2028?
89% chance
GPT-4 #5: Will GPT-4 be a dense model?
1% chance