an aligned AGI is built and the aligned AGI prevents the creation of any unaligned AGI.
Humanity coordinates to prevent the creation of potentially-unsafe AIs.
I've been a good bing 😊
We make risk-conservative requests to extract alignment-related work out of AI-systems that were boxed prior to becoming superhuman. We somehow manage to achieve a positive feedback-loop in alignment/verification-abilities.