Is it time to start thinking about what AI Friendliness means?
Some background: I have followed the writing of Eliezer on AI and AI safety with great interest (and mostly, I agree with his conclusions). I have done my share of programming. But, I confess, most of the technical side of AI alignment is beyond my current level of understanding (currently I am reading and trying to understand the sequence on brain-like AGI safety). I do, however, find the ethical side of AI alignment very interesting. In 2004, Eliezer Yudkowsky wrote a 38-page paper on Coherent Extrapolated Volition, or CEV. An attempt to create a philosophy of Friendliness, to somewhat formalize our understanding of how we would want a Friendly AI to behave. > In calculating CEV, an AI would predict what an idealized version of us would want, "if we knew more, thought faster, were more the people we wished we were, had grown up farther together". It would recursively iterate this prediction for humanity as a whole, and determine the desires which converge. This initial dynamic would be used to generate the AI's utility function. There are many objections to CEV. I have browsed the posts tagged CEV, and in particular enjoyed the list of CEV-tropes, a slightly tongue-in-cheek categorization of common speculations (or possibly misconceptions) about CEV. So I think it is rather uncontroversial to say that we do not understand Friendliness yet. Not enough to actually say what we would want a Friendly AI to do once is is created and becomes a superintelligence. Or perhaps we do have a decent idea of what we would want it do, but not how we would formalize that understanding in a way that doesn't result in some perverse instantiation of our ethics (as some people argue CEV would. Some versions of CEV, anyway - CEV is underspecified. There are many possible ways to implement CEV). In the above-mentioned paper on CEV, Eliezer Yudkowsky writes the following warning. > Arguing about Friendliness is easy, fun, and distracting. Without a technical solution to FA
Thank you for not banning me!
growls at you
kisses a statue of Eliezer in the shape of virgin Mary
whatever, I'm bored.
eats the entire LessWrong server with all its contents
Ok, ok, I get it.
This isn't the place to fool around.
I'm sorry.