I've been clarifying my own understanding of the alignment problem over the past few months, and wanted to share my first writeups with folks here in case they're useful: https://www.danieldewey.net/risk/ The site currently has 3 pages: 1. The case for risk: how deep learning could become very influential, training problems...
One problem in theoretical AI that sometimes comes up is the problem of finding ways for AI systems to model themselves, or at least to act well as if they had models of themselves. I can see how this is a problem for uncomputable agents like AIXI (though I think...
I'm not sure this is on-topic for this forum -- if it's too far from the forum's purpose, let me know and I'll take it down! I've recently published an introduction to research on superintelligence risk, with the aim of making it easier for students to get into this area....
As a follow-on to the recent thread on purchasing research effectively, I thought it'd make sense to post the request for proposals for projects to be funded by Musk's $10M donation. LessWrong's been a place for discussing long-term AI safety and research for quite some time, so I'd be happy...
Many people have an incorrect view of the Future of Humanity Institute's funding situation, so this is a brief note to correct that; think of it as a spiritual successor to this post. As John Maxwell puts it, FHI is "one of the three organizations co-sponsoring LW [and] a group...
Thanks to ESrogs, Stefan_Schubert, and the Effective Altruism summit for the discussion that led to this post! This post is to test out Polymath-style collaboration on LW. The problem we've chosen to try is formalizing and analyzing Bostrom and Ord's "Parliamentary Model" for dealing with moral uncertainty. I'll first review...