[Note: This post is an excerpt from a longer paper, written during the first half of the Philosophy Fellowship at the Center for AI Safety. This post is something of a companion piece to my deontological AI post; both were originally written as parts of a single paper. (There's a small amount of overlap between the two.)]
1. Introduction
Two goals for the future development of AI stand out as desirable:
- First, advanced AI should behave morally, in the sense that its decisions are governed by appropriately chosen ethical norms.
- Second, advanced AI should behave safely, in the sense that its decisions shouldn’t unduly harm or endanger humans.
These two goals are often viewed as closely related.... (read 2996 more words →)
Another academic philosopher, directed here by @Simon Goldstein. Hello Wei!
- It's not common to switch entirely to metaphilosophy, but I think lots of us get more interested in the foundations and methodology of at least our chosen subfields as we gain experience, see where progress is(n't) being made, start noticing deep disagreements about the quality of different kinds of work, and so on. It seems fair to describe this as awakening to a need for better tools and a greater understanding of methods. I recently wrote a paper about the methodology of one of my research areas, philosophy of mathematical practice, for pretty much these reasons.
- Current LLMs are pretty awful at discussing the
... (read more)