The first Anthropic-trained AI system to qualify as ASL-3 qualifies because of misuse risks rather than autonomy risks
Basic
13
αΉ39610000
82%
chance
1D
1W
1M
ALL
This will be evaluated according to the AI Safety Levels (ASL) standard v1.0 defined by Anthropic here, which gives two different ways that an AI system could qualify for AI Safety Level 3 (ASL-3). This resolves based on the first clear public disclosure by Anthropic that indicates that they have trained a model and found it to qualify for ASL-3.
If Anthropic announces a policy that would prevent this information from being disclosed, announces that it has permanently ceased developing new AI systems, or ceases to operate, this will resolve N/A after six months.
Get αΉ600 play money
Related questions
SoAI 23 3/10: Will Self-improving Al agents crush SOTA in a complex environment (e.g. AAA game, tool use, science)?
40% chance
When will there first be an AI system that qualifies as ASL-3?
When will Anthropic first train an AI system that they claim qualifies as ASL-3?
When will there first be a credible report that an AI system qualifies as ASL-3?
Will technical limitations or safeguards significantly restrict public access to smarter-than-almost-all-humans AGI?
48% chance
Is AI Safety a grift?
28% chance
Will AGI be a problem before non-G AI?
26% chance
Will an unaligned AI or an aligned AI controlled by a malicious actor create a "wake-up call" for humanity on AI safety?
68% chance
Will there be serious AI safety drama at Meta AI before 2026?
55% chance
Is the nature of AI risk completely misunderstood today with respect to the state of the art in 2030?
65% chance