I know a way to acheive game theoretic parity by constructing a mechanism that rewards mechanistically interpretable decentralized alignment.
I don't have the time or resources to implement this mechanism sufficiently into the public domain.
If I secured a grant sufficient enough ($50k+) to implement the mechanism concretely in a way that sovereign individuals could utilize for profit, we'd see a radical shift in optimism in the AI existential risk community.
I have a number of bets on here talking about this.
Very simply, this wager claims IF someone gives me $50k, THEN I will win a Nobel Prize for demonstrating a mechanistically interpretable solution to alignment.
Will resolve NA in 2030 if I fail to secure a grant.
Will resolve "yes" if I win any Nobel Prize within 5 years of receiving the grant.
Will resolve "no" 5 years after receiving the grant if no Nobel Prize is awarded.