LessWrong is currently in the process of migrating from AWS to Vercel, as part of a project to migrate our codebase to NextJS[1]. This post should go live shortly after we cut over traffic to the new host (and updated codebase). This should hopefully be a pretty low-risk operation. If...
Eliezer's AI doom arguments have had me convinced since the ancient days of 2007, back when AGI felt like it was many decades away, and we didn't have an intelligence scaling law (except to the Kurzweilians who considered Moore's Law to be that, and were, in retrospect, arguably correct). Back...
LessWrong has been receiving an increasing number of posts and comments that look like they might be LLM-written or partially-LLM-written, so we're adopting a policy. This could be changed based on feedback. Note: first-time writers are not permitted to use any AI text output in their submissions. The guidance below...
Arbital was envisioned as a successor to Wikipedia. The project was discontinued in 2017, but not before many new features had been built and a substantial amount of writing about AI alignment and mathematics had been published on the website. If you've tried using Arbital.com the last few years, you...
This open thread introduces an experimental extension of LessWrong's voting system: reactions. Unlike votes, reactions are public; hovering over the reactions will show a list of users who reacted. For now, this feature is only for comments on this post in particular; after collecting feedback, we might roll out more...
A lot of AI-risk-concerned people are struggling with how to relate to dual-use research, and relatedly, to doing alignment research inside of AI orgs. There's a pretty simple concept that seems, to me, to be key to thinking about this coherently: the dual-useness ratio. Most prosaic alignment techniques are going...
I think actual infohazardous information is fairly rare. Far more common is a fork: you have some idea or statement, you don't know whether it's true or false (typically leaning false), and you kow that either it's false or it's infohazardous. Examples include unvalidated insights about how to build dangerous...