By 2028, will there be a language model of less than 10B parameters that is superior to GPT-4?
13
65
310
2028
81%
chance

Pretty impressed by this:

https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/

I will not bet in this market. I will look primarily at whatever benchmarks are most prominent at the time to determine which model is better.

The model does not have to be clearly superior to gpt-4, if it's just a bit better I will resolve YES. If anyone has suggestions for less subjective resolution criteria that won't be goodharted I am open.

Get Ṁ600 play money
Sort by:

Related

if the model can be stored on 10b params but takes 100000x more time at inference time, because it searches through lots of possibilities or reprompts itself a bunch of times or similar tricks to improve its output, would that still count? In 2023 it would still not be practical to run, but it would still be smaller in the ‘# of parameters‘ sense

@TheBayesian I'll allow it, parameters are the only variable of interest

More related questions