Will research demonstrate an advantage to uninterpretable CoT before 2030?
9
100Ṁ307
2030
75%
chance

Daniel Kokotajlo makes the following prediction:

Sometime in the next few years probably, various researchers will discover:
* That if you scale up RL by additional OOMs, the CoTs evolve into some alien optimized language for efficiency reasons
* That you can train models to think in neuralese of some sort (e.g. with recurrence, or more high-dimensional outputs at least besides tokens) to boost performance.

Resolves YES if at least three papers agreeing with at least one of the above bulleted claims are published before market close.

  • Update 2025-03-15 (PST) (AI summary of creator comment): Clarification Details

    • Uninterpretable means the chain-of-thought must be illegible-by-reading-alone.

    • Results that can be decoded by training an auxiliary model do not count toward meeting this criterion.

Get
Ṁ1,000
to start trading!
Sort by:

Uninterpretable seems strong, neuralese may well be understandable with translation

@EthanKuntz For the purpose of resolution I’ll consider just illegible-by-reading-alone, even if it’s possible to train an auxiliary model to decode it or something.

bought Ṁ30 YES

I think this is very likely going to resolve yes

“At least three papers” was my best idea for disambiguating “various researchers will discover” from “someone churns out a p-hacked nothing for social media clicks”. I’ll stick to it for this market but LMK if you have a better idea for judging “legit ML result” for future markets.

© Manifold Markets, Inc.Terms + Mana-only TermsPrivacyRules