To translate Graham's statement back to the FAI problem: In Eliezer's alignment talk, he discusses the value of solving a relaxed constraint version of the FAI problem by granting oneself unlimited computing power. Well, in the same way, the AGI problem can be seen as a relaxed constraint version of the FAI problem. One could argue that it's a waste of time to try to make a secure version of AGI Approach X if we don't even know if it's possible to build an AGI using AGI Approach X. (I don't agree with this view, but I don'...
It seems to me that to a large extent impressions can be framed in terms of vague predictive statements with no explicit probabilistic or causal content, influenced much more by external factors than reasoned beliefs.
"He seemed nice" corresponds to "X% chance that if I met him more often, we would continue getting along well".
"That sounds crazy" corresponds to "I can't really tell you why but I find that rather improbable".
If I am right about this, the first and main step would be lto earn how to turn impressio
But on the other hand, consequentialism is particularly prone to value misalignment. In order to systematize human preferences or human happiness, it requires a metric; in introducing a metric, it risks optimizing the metric itself over the actual preferences and happiness.
Yes, in consequentialism you try to figure out what values you should have, and your attempts at doing better might lead you down the Moral Landscape rather than up toward a local maximum.
But what are the alternatives? In deontology you try to follow a bunch of rules in the hope that t
Admittedly, I do not have much of an idea about Infinite Ethics, but it appeared to me that the problem was to a large extent about how to deal with an infinite number of agents on which you can define no measure/order so that you can discount utilities.
Right now, I don't see how this approach helps with that?