Will "AI Control: Improving Safety Despite Intentio..." make the top fifty posts in LessWrong's 2023 Annual Review?
Plus
12
Ṁ502Feb 1
86%
chance
1D
1W
1M
ALL
As part of LessWrong's Annual Review, the community nominates, writes reviews, and votes on the most valuable posts. Posts are reviewable once they have been up for at least 12 months, and the 2023 Review resolves in February 2025.
This market will resolve to 100% if the post AI Control: Improving Safety Despite Intentional Subversion is one of the top fifty posts of the 2023 Review, and 0% otherwise. The market was initialized to 14%.
This question is managed and resolved by Manifold.
Get
1,000
and3.00
Related questions
Related questions
Will "The Checklist: What Succeeding at AI Safety W..." make the top fifty posts in LessWrong's 2024 Annual Review?
36% chance
Will "Predictable updating about AI risk" make the top fifty posts in LessWrong's 2023 Annual Review?
53% chance
Will "There should be more AI safety orgs" make the top fifty posts in LessWrong's 2023 Annual Review?
34% chance
Will "There should be more AI safety orgs" make the top fifty posts in LessWrong's 2023 Annual Review?
37% chance
Will "Speaking to Congressional staffers about AI risk" make the top fifty posts in LessWrong's 2023 Annual Review?
32% chance
Will "My May 2023 priorities for AI x-safety: more ..." make the top fifty posts in LessWrong's 2023 Annual Review?
25% chance
Will "Cognitive Emulation: A Naive AI Safety Proposal" make the top fifty posts in LessWrong's 2023 Annual Review?
37% chance
Will "AI catastrophes and rogue deployments" make the top fifty posts in LessWrong's 2024 Annual Review?
41% chance
Will "Talking publicly about AI risk" make the top fifty posts in LessWrong's 2023 Annual Review?
21% chance
Will "Deep atheism and AI risk" make the top fifty posts in LessWrong's 2024 Annual Review?
42% chance