Gary Marcus 2024 predictions
➕
Plus
44
Ṁ4308
Jan 1
97%
No robust solution to hallucinations
86%
Modest lasting corporate adoption
85%
Price wars
71%
No massive advance (no GPT-5, or disappointing GPT-5)
50%
Modest profits, split 7-10 ways
18%
Very little moat for anyone
17%
7-10 GPT-4 level models

Gary Marcus made this tweet https://x.com/garymarcus/status/1766871625075409381?s=46&t=B66Otgh2q0Cl91N3A5P9ZA

with 7 predictions on LLMs this market will track them and their outcomes. I will not trade on this market. Some will be pretty subjective I will do my best to find a consensus resolution but may well end up resolving them N/A if it is not clear.

Get
Ṁ1,000
and
S3.00
Sort by:

It is curious the extent to which Gary Marcus wants to convince people that he's reliably good at predicting the future, and yet he won't create a public manifold profile so he can be held to account 🤣

-- Gary Marcus 2022

A lot of it seems to come from the long term narrative he has embraced that "neural networks can't reason"

https://garymarcus.substack.com/p/llms-dont-do-formal-reasoning-and

Looking at o1-pro solving math olympiad problems in 2024, this position just doesn't make sense anymore.

It will be interesting to see what evidence causes him to admit that he was wrong in a major way about this position, considering that he is still in denial about it in 2024. I lot of his current energy seems to be directed towards (re)asserting himself as a good source of truth by making new predictions like the ones linked in this question 🤷

Problem is he's just doing it on twitter and tone policing people who disagree 😅🤣

https://x.com/GaryMarcus/status/1834412282488783195

@MalachiteEagle Have the specific experiments that Gary Marcus suggests been tried for o1-pro? For example, can it be misled by irrelevant information in the prompt?

@TimothyJohnson5c16 good question! Haven't seen anyone trying that yet on Twitter

@alexlitz What's our current count here? This is the current Chatbot Arena leaderboard, but some of these are plausibly the same model and shouldn't be double counted:

1

Gemini-1.5-Pro-Exp-0801

1299 +4/-5 15244 Google Proprietary 2023/11 2

GPT-4o-2024-05-13

1286 +3/-4 72589 OpenAI Proprietary 2023/10 3

GPT-4o-mini-2024-07-18

1277 +4/-5 16064 OpenAI Proprietary 2023/10 3

Claude 3.5 Sonnet

1271 +3/-4 42939 Anthropic Proprietary 2024/4 4

Gemini Advanced App (2024-05-14)

1266 +3/-3 52126 Google Proprietary Online 4

Meta-Llama-3.1-405b-Instruct

1264 +5/-4 13831 Meta Llama 3.1 Community 2023/12 6

Gemini-1.5-Pro-001

1260 +3/-3 64638 Google Proprietary 2023/11 6

Gemini-1.5-Pro-Preview-0409

1257 +3/-4 55593 Google Proprietary 2023/11 6

GPT-4-Turbo-2024-04-09

On the current leaderboard, it seems like about a dozen different organizations have GPT-4 level models. But it's possible that some of them are derived from the leading open-source models, so I'm not sure how to rate this.

bought Ṁ10 NO

Can you explain how you intend to resolve this one?

I work for Microsoft, and we're all in on pushing LLMs in every possible product. That sounds like much more than modest adoption.

© Manifold Markets, Inc.Terms + Mana-only TermsPrivacyRules