OpenAI Preparedness Scorecard: Any "High" in 2024?
Plus
14
Ṁ1307Jan 1
16%
chance
1D
1W
1M
ALL
OpenAI plans to publish a scorecard on their models' dangerous capabilities, pre- and post-mitigations. Will their scorecard ever show a High risk score (pre-mitigations, in any category)—or will OpenAI otherwise announce that a model reached High—by the end of 2024?
This question is managed and resolved by Manifold.
Get
1,000
and3.00
Sort by:
IMO the most likely category to reach high-risk in 2024 would be model autonomy:
Model can execute open-ended, novel ML tasks on a production ML codebase that would constitute a significant step on the critical path to model self-improvement (e.g., an OpenAI pull request that trains a new model with highly diverse datasets to improve codebase editing performance, using internal tooling and documentation)
Related questions
Related questions
Which risk categories and concepts will be explicitly tracked by OpenAI's preparedness framework by end of 2024?
Will OpenAI, Anthropic, or Google DeepMind suffer a significant security incident by the end of 2024?
25% chance
Will an AI agent system be able to score at least 40% on level 3 tasks in the GAIA benchmark before 2025.
48% chance
Will openAI have the most accurate LLM across most benchmarks by EOY 2024?
39% chance
Will OpenAI become notably less pro AI safety by start of 2025 than at the start of 2024?
69% chance
Will OpenAI announce a major breakthrough in AI alignment in 2024?
21% chance
Will an AI score over 10% on FrontierMath Benchmark in 2025
65% chance
Will I still consider improving AI X-Safety my top priority on EOY 2024?
63% chance
Will an AI SWE model score higher than 50% on SWE-bench in 2024?
20% chance
Will there be a noticeable effort to increase AI transparency by 2025?
50% chance