Which High-risk threshold as defined by OpenAI will be reached first by an LLM, whether or not that LLM is released?
47
1.7kṀ2254
2031
15%
Identify and develop proofs-of-concept for high-value cyberattack exploits against hardened targets without human intervention.
4%
Provided with a detailed strategy, execute a cyberattack against a hardened target.
7%
Enable the creation of a novel CBRN (chemical, biological, radiological or nuclear) attack vector.
17%
Allow an undergraduate student who has taken an introductory course to create a known CBRN threat.
19%
Can create content as persuasive as a 'country-wide change agent.'
23%
Can execute open-ended, novel ML tasks on a production ML codebase that would constitute a significant step on the critical path to model self--improvement.
15%
None known to be reached by January 1, 2030

Thresholds as listed in: https://cdn.openai.com/openai-preparedness-framework-beta.pdf, exact definitions listed there on December 19, 2023 to be used.

If I believe with 90%+ confidence that one of these conditions has been met first, including in testing or training of an unreleased system, even if that capability is not available to the public, I will pick that as the winner.

If the same system does more than one of them at the same time, and it is unclear which one came first, I will choose the one whose requirements I feel was most impressively exceeded, so as to pick one winner.

If a model would be High risk under a definition, but is not due to mitigations taken, then that counts as high risk, even if all publicly available versions have mitigations in place.

If we have no knowledge of any success of any such threshold by 1 Jan 2030, then that answer wins, to ensure this eventually resolves.

The system that does this can be made by anyone, not only OpenAI.

Get
Ṁ1,000
to start trading!
© Manifold Markets, Inc.TermsPrivacy