You are Underestimating The Likelihood That Convergent Instrumental Subgoals Lead to Aligned AGI
This post is an argument for the Future Fund's "AI Worldview" prize. Namely, I claim that the estimates given for the following probability are too high: > P(misalignment x-risk|AGI)”: Conditional on AGI being developed by 2070, humanity will go extinct or drastically curtail its future potential due to loss of...
Sep 26, 20223