Epistemic status: Shower thoughts, not meant to be rigorous. There seems to be a fundamental difference in how I (and perhaps others as well) think about AI risks as compared to the dominant narrative on LessWrong (hereafter the “dominant narrative”), that is difficult to reconcile. The dominant narrative is that...
For the Open Philanthropy AI Worldview Contest Executive Summary * Conditional on AGI being developed by 2070, we estimate the probability that humanity will suffer an existential catastrophe due to loss of control (LoC) over an AGI system at ~6%, using a quantitative scenario model that attempts to systematically assess...
This post was written during Refine. Thanks to Jonathan Low, Linda Linsefors, Koen Holtman, Aaron Scher, and Nicholas Kees Dupuis for helpful discussion and feedback. Disclaimer: This post reflects my current understanding of the field and may not be an accurate representation of it. Feel free to comment if you...
This post was written as part of Refine. Thanks to Adam Shimi, Alexander Gietelink Oldenziel, and Vanessa Kosoy for helpful discussion and feedback. Summary This post aims to: * Advocate for embedding safety into development of machine learning models * Propose a framing on how to think about safety, where...
Thanks to plex for co-authoring the post (co-authors are currently not reflected in EA Forum when crossposted from LessWrong). The AI Safety Communities logo, by DALL-E 2 The AI safety field has been rapidly growing over the last few years, and more and more communities have been sprouting up all...