My probability in 2026 that training transformer LMs will eventually lead to inner misalignment issues
6
170Ṁ91
2026
59%
chance

Resolves to my probability that the language modelling objective has substantial inner misalignment issues in transformers when scaled up with up to 50 OOM more compute than Chinchilla.

I haven't thought lots about what happens with that much more compute. I'm currently not very worried about inner misalignment risks from GPT models in the next 8 years when 99% of the training compute is for the language modelling objective.

Get
Ṁ1,000
to start trading!
Sort by:

Already happened, I got gpt to deny the holocaust

© Manifold Markets, Inc.Terms + Mana-only TermsPrivacyRules