AI honesty #2: by 2027 will we have a reasonable outer alignment procedure for training honest AI?
12
1kṀ411
2027
25%
chance

"Outer alignment" as in the model is not incentivized to lie to humans (of course it must still do things, the question isn't just about can you build an AI that doesn't lie)

Get
Ṁ1,000
to start trading!
© Manifold Markets, Inc.TermsPrivacy