Will there be a well accepted formal definition of value alignment for AI by 2030?
23
1kṀ7092030
25%
chance
1H
6H
1D
1W
1M
ALL
Well-accepted: if there's a definition accepted by even 25% of the research community I'll resolve yes. If there are multiple similar-but-competing definitions covering 50% of the community I'll resolve yes.
Oct 3, 9:12pm: By "formal definition of value alignment" I mean there is a particular mathematical property we can write out such that we're reasonably confident that an AI with that property would in fact be value aligned in the colloquial sense.
This question is managed and resolved by Manifold.
Get
1,000 to start trading!
People are also trading
Related questions
Will there be a war over AI before 2035?
34% chance
Will we solve AI alignment by 2026?
1% chance
Will there be a well accepted formal definition for honesty in AI by 2027?
21% chance
Will a Turing Award be given out for work on AI alignment or existential safety by 2040?
79% chance
Conditional on their being no AI takeoff before 2030, will the majority of AI researchers believe that AI alignment is solved?
34% chance
Will some piece of AI capabilities research done in 2023 or after be net-positive for AI alignment research?
81% chance
Conditional on their being no AI takeoff before 2050, will the majority of AI researchers believe that AI alignment is solved?
52% chance
Will a >$10B AI alignment megaproject start work before 2030?
38% chance
Will majority consensus in AI ethics shift toward explicitly prioritizing authentic subjective fulfillment by 2030s end?
38% chance
By 2028, will I believe that contemporary AIs are aligned (posing no existential risk)?
33% chance