The Long-Term Future Fund has an AMA up on the Effective Altruism Forum. There’s no real deadline for questions, but let’s say we have a soft commitment to focus on questions asked on or before September 8th. I'd prefer centralizing the questions to one place. If you don't want to...
Epistemic status: confident that the underlying idea is useful; less confident about the details, though they're straightforward enough that I expect they're mostly in the right direction. TLDR: This post describes a pre-mortem-like exercise that I find useful for thinking about AGI risk. It is the only way I know...
Subtitle: A partial defense of high-confidence AGI doom predictions. Introduction Consider these two kinds of accident scenarios: 1. In a default-success scenario, accidents are rare. For example, modern aviation is very safe thanks to decades of engineering efforts and a safety culture (e.g. the widespread use of checklists). When something...
Be the AGI you want to see in the world. Epistemic status: highly speculative, authors are not neuroscientists. Summary * It may be possible to enhance human intelligence via a brain-computer interface (BCI). We could put electrodes into a human brain, connect those electrodes to an artificial neural network, and...
Status: working notes Here's an exercise I've found very useful for intuition-building around alignment: 1. Propose a solution to the alignment problem. 2. Dig into the details until you understand why the proposal fails. 3. If there is an obvious fix, go back to step 2 and iterate. In this...
TLDR If you work in AI alignment / safety research, please fill out this form on how useful access to extra compute would be for your research. This should take under 10 minutes, and you don't need to read the rest of this post beforehand—in fact it would be great...