
My probability in 2026 that training transformer LMs will eventually lead to inner misalignment issues
56%
chance
1D
1W
1M
ALL
Resolves to my probability that the language modelling objective has substantial inner misalignment issues in transformers when scaled up with up to 50 OOM more compute than Chinchilla.
I haven't thought lots about what happens with that much more compute. I'm currently not very worried about inner misalignment risks from GPT models in the next 8 years when 99% of the training compute is for the language modelling objective.
Sort by:






Related markets
David Johnston
Best available language model from an OpenAI competitor by 202662%
David Johnston
Most popular language model from OpenAI competitor by 2026?40%
MP
Will a large language models beat a super grandmaster playing chess by 2028?54%

Leo Gao
Will Transformer based architectures still be SOTA for language modelling by 2026?64%
Related markets
David Johnston
Best available language model from an OpenAI competitor by 202662%
David Johnston
Most popular language model from OpenAI competitor by 2026?40%
MP
Will a large language models beat a super grandmaster playing chess by 2028?54%

Leo Gao
Will Transformer based architectures still be SOTA for language modelling by 2026?64%