Superalignment is an agenda at OpenAI attempting to "solve the core technical challenges of superintelligence alignment in four years".
Daniel Filan's AXRP podcast interview with Jan Leike on superalignment
If at least three of the judges believe that the superalignment agenda has been successful in "solv[ing] the core technical challenges of superintelligence alignment" by the end of 2027, this market will resolve YES.
Otherwise it resolves NO.
The judges for this question are:
Eliezer Yudkowsky
Nate Soares
Paul Christiano
John Wentworth
David Krueger
Notes:
The question refers to the agenda, not the team. What that means is up to the individual judges to decide. However, the intent of this question is to predict the success/failure of the agenda - not whether it happens to succeed at OpenAI, or under Jan/Ilya.
The end of year cutoff is to allow ample time for any materials published by the superalignment team to be read and considered by the judges.
For the purposes of this question, a judge believes the agenda succeeded if they say as much publicly (e.g. on the alignment forum, on lesswrong, on twitter). If there's ambiguity, then the judge's direct answer to [should their vote on this question be for yes/no?] always takes precedence.
If they say nothing, that counts as a 'no' vote.
In the unlikely event that any of the judges dies before the market closes, they'll be replaced by someone I guess to have a similar bar for superalignment success. (or if I die too, someone else's best guess!)
Will Superalignment succeed, according to a majority of Eliezer, Nate, Paul, John, David.
7
Ṁ90Ṁ1.2k2028
4%
chance
1H
6H
1D
1W
1M
ALL
This question is managed and resolved by Manifold.
Market context
Get
1,000 to start trading!
By what years will I agree that "The Superalignment Problem has a technically elegant solution w.p. >80%”?
Will Jan Leike still be leading Superalignment in 4 years?
1% chance
Will "Eliezer's Lost Alignment Articles / The Arbit..." make the top fifty posts in LessWrong's 2025 Annual Review?
14% chance
Will OpenAI achieve "very high level of confidence" in their "Superalignment" solutions by 2027-07-06?
4% chance
Will Eliezer believe in mankind's survival again?
55% chance