Will GPT-4 have at least 100 trillion parameters?
Will GPT-4 have at least 100 trillion parameters?
81
1.6kṀ25k
resolved Nov 17
Resolved
N/A

This will resolve TRUE if GPT-4's announcement states that it has at least 100 trillion parameters.

For context, GPT-3 has 175 billion parameters.

See also:

Get
Ṁ1,000
to start trading!


Sort by:
predictedNO 1y

Sorry for the delay here, and for writing the market poorly. The title "Will GPT-4 have at least 100 trillion parameters?" is quite different from my description "GPT-4's announcement states that it has at least 100 trillion parameters."

OpenAI has not revealed the number of parameters, though George Hotz has claimed it's 1.7 trillion. So according to the title, this market may resolve N/A by the time it closes, if it remains without an authoritative source.

According to my description, it would resolve NO, since OpenAI didn't state the number of parameters when they announced GPT-4. But traders may not have reasonably assumed that the resolution criteria in the description followed from the title. Conditional on GPT-4 having over 100 trillion parameters, the probability that OpenAI would announce that when introducing GPT-4 is plausibly small.

Given the ambiguity, I am resolving this N/A.

1y

@MaxGhenis Can this resolve?

predictedNO 2y

@MaxGhenis This market is supposed to be conditional on the value stated in the anouncment. Unfortunatly it seems the announcment does not contain those architecture details. Resolve N/A?

predictedNO 2y

@tinytitan Technically correct, the best kind of correct.

1y

@tinytitan Why would it be N/A?

if GPT-4's announcement states that it has at least 100 trillion parameters

There was an announcement, right? Did it state it or not? (As best I'm aware, it did not.)

2y

Should be lower than 10%...should be 1% or less. The cost for 100T in 2023 would likely be equivalent to the total US military spending in Ukraine to date. I don't see a company ponying up that money. Now, if a OpenAI comes out and redefines what a parameter is for marketing purposes, or if the say some bullshit like, "precision parameters," which is not the same thing and then everyone's mind is blown because they called a cat a dog...it ain't happening. Now...go forward another 4 years and the cost might be closer to $200M instead of $20B so I could see it happening in 2027 or after since Megatron-Turing NLG 530B cost Microsoft $100M's.

Manifold in the wild: A Tweet by Max Ghenis

@AlexHormozi https://manifold.markets/MaxGhenis/will-gpt4-have-at-least-100-trillio?referrer=MaxGhenis

2y

Buying Yes because the potential upside is so big

2y

I think this is mostly a question about if GPT-4 will be a MoE
(https://manifold.markets/vluzko/will-gpt4-be-a-dense-model), and then how many experts a GPT-4 model would have.

The largest MoE model in the GShard paper has 2048 experts per expert-layer. https://download.arxiv.org/pdf/2006.16668v1

If a 300B parameter model had 2048 experts per layer, that would be ~600T parameters
If a 175B parameter model had 2048 experts per layer, that would be ~360T parameters

GPT-3's batch size was 500K (https://arxiv.org/pdf/2005.14165.pdf), so 2K experts wouldn't be mad, esp. with better hardware.

Given that the odds GPT-4 will be a MoE seem ~30% to me, and I think that the odds are that it would cross the 100T threshold 35% of the time, I think the odds that GPT-4 will have >100T parameters are ~11%.

predictedYES 2y

@NoaNabeshima I feel most uncertain about the probability it crosses the 100T threshold conditional on GPT-4 being a MoE, so I think this estimate could be improved if someone thought about that

predictedYES 2y

@NoaNabeshima in particular it seems plausible my probability should be higher

predictedYES 2y

@NoaNabeshima but also lower, eh

2y

@NoaNabeshima oh also I'm incorrectly assuming that the number of parameters scales linearly with the number of experts but actually probably only the feedforward layers would be duplicated, making the parameter scaling factor be (1+E)/2 instead of E, where E is the number of experts.

Additionally, not every feedforward layer needs to be a mixture-of-experts layer. EG in GShard only half of them are mixture-of-experts layers. So if P of the feedforward layers are mixture of experts, the parameter scaling factor would be

(1+(P*E+(1-P))/2

For P = 0.5, and 2048 experts, that's a scaling factor of 512x, so only 89T parameters for a 175B parameter base model.

predictedNO 2y

@NoaNabeshima

If E is large,

(1+(P*E + (1-P)))/2 ~= P*E/2

2y

Started one that uses a numeric answer:

2y

In an August 2021 Wired article, Andrew Feldman, founder and CEO of Cerebras, said, "From talking to OpenAI, GPT-4 will be about 100 trillion parameters."

According to Alberto Romero, though, "not much later, Sam Altman, OpenAI’s CEO, denied the 100T GPT-4 rumor in a private Q&A."

Will GPT-4 have at least 100 trillion parameters?, 8k, beautiful, illustration, trending on art station, picture of the day, epic composition

What is this?

What is Manifold?
Manifold is the world's largest social prediction market.
Get accurate real-time odds on politics, tech, sports, and more.
Or create your own play-money betting market on any question you care about.
Are our predictions accurate?
Yes! Manifold is very well calibrated, with forecasts on average within 4 percentage points of the true probability. Our probabilities are created by users buying and selling shares of a market.
In the 2022 US midterm elections, we outperformed all other prediction market platforms and were in line with FiveThirtyEight’s performance. Many people who don't like betting still use Manifold to get reliable news.
ṀWhy use play money?
Mana (Ṁ) is the play-money currency used to bet on Manifold. It cannot be converted to cash. All users start with Ṁ1,000 for free.
Play money means it's much easier for anyone anywhere in the world to get started and try out forecasting without any risk. It also means there's more freedom to create and bet on any type of question.
© Manifold Markets, Inc.TermsPrivacy