Comment author: thomblake 18 July 2012 09:10:47PM 2 points [-]

I've realized that my sibling comment is logically rude, because I've left out some relevant detail. Most relevantly, I tend to self-describe as a virtue ethicist.

I've noticed at least 3 things called 'virtue ethics' in the wild, which are generally mashed together willy-nilly:

  1. an empirical claim, that humans generally act according to habits of action and doing good things makes one more likely to do good things in the future, even in other domains
  2. the notion that ethics is about being a good person and living a good life, instead of whether a particular action is permissible or leads to a good outcome
  3. virtue as an achievement; a string of good actions can be characterized after the fact as virtuous, and that demonstrates the goodness of character.

There are virtue ethicists who buy into only some of these, but most often folks slip between them without noticing. One fellow I know will often say that #1 being false would not damage virtue ethics, because it's really about #2 and #3 - and yet he goes on arguing in favor of virtue ethics by citing #1.

Comment author: jacoblyles 18 July 2012 09:21:21PM 1 point [-]

This is a great framework - very clear! Thanks!

Comment author: thomblake 18 July 2012 09:00:30PM 2 points [-]

I'm still not sold on the idea that an intelligent being would slavishly follow its utility function.

If it's really your utility function, you're not following it "slavishly" - it is just what you want to do.

For AI, there are no questions about the meaning of life then? Just keep on U maximizing?

If "questions about the meaning of life" maximize utility, then yes, there are those. Can you unpack what "questions about the meaning of life" are supposed to be, and why you think they're important? ('meaning of "life"' is fairly easy, and 'meaning of life' seems like a category error).

Comment author: jacoblyles 18 July 2012 09:18:39PM -1 points [-]

Sorry, "meaning of life" is sloppy phrasing. "What is the meaning of life?" is popular shorthand for "what is worth doing? what is worth pursuing?". It is asking about what is ultimately valuable, and how it relates to how I choose to live.

It's interesting that we are imagining AIs to be immune from this. It is a common human obsession (though maybe only among unhappy humans?). An AI isn't distracted by contradictory values like a human is then, it never has to make hard choices? No choices at all really, just the output of the argmax expected utility function?

Comment author: jacoblyles 18 July 2012 08:59:02PM 1 point [-]

I follow the virtue-ethics approach, I do actions that make me like the person that I want to be. The acquisition of any virtue requires practice, and holding open the door for old ladies is practice for being altruistic. If I weren't altruistic, then I wouldn't be making myself into the person I want to be.

It's a very different framework from util maximization, but I find it's much more satisfying and useful.

Comment author: wedrifid 18 July 2012 10:42:36AM *  3 points [-]

What makes us think that AI would stick with the utility function they're given?

There are very few situations in which an agent can most effectively maximise expected utility according to their current utility function by modifying themselves to have a different utility function. Unless the AI is defective or put in a specially contrived scenario it will maintain its current utility function because that is an instrumentally useful thing to do.

If you are a paperclip maximiser then becoming a staples maximiser is a terribly inefficient strategy for maximising paperclips unless Omega is around making weird bargains.

I change my utility function all the time, sometimes on purpose.

No you don't. That is, to the extent that you "change your utility function" at all you do not have a utility function in sense meant when discussing AI. It only makes sense to model humans as having 'utility functions' when they are behaving in a manner that can be vaguely approximated as expected utility maximisers with a particular preference function.

Sure, it is possible to implement AIs that aren't expected utility maximisers either and those AIs could be made to do all sorts of arbitrary things including fundamentally change their goals and behavioral strategies. But if you implement an AI that tries to maximise a utility function then it will (almost always) keep trying to maximise that same utility function.

Comment author: jacoblyles 18 July 2012 08:37:52PM *  -1 points [-]

Let me see if I understand what you're saying.

For humans, the value of some outcome is a point in multidimensional value space, whose axes include things like pleasure, love, freedom, anti-suffering, and etc. There is no easy way to compare points at different coordinates. Human values are complex.

For a being with a utility function, it has a way to take any outcome and put a scalar value on it, such that different outcomes can be compared.

We don't have anything like that. We can adjust how much we value any one dimension in value space, even discover new dimensions! But we aren't utility maximizers.

Which raises the question - if we want to create AI that respect human values, then why would we make utility maximizer AI in the first place?

I'm still not sold on the idea that an intelligent being would slavishly follow its utility function. For AI, there are no questions about the meaning of life then? Just keep on U maximizing?

Comment author: thomblake 18 July 2012 04:14:44PM 0 points [-]

The point was not necessarily to advocate torture. It's to take the math seriously.

In fact, the mental discomfort caused by people who heard of the torture would swamp the disutility from the dust specks.

Just how many people do you expect to hear about the torture? Have you taken seriously how big a number 3^^^3 is? By how many utilons do you expect their disutility to exceed the disutility from the dust specks?

Comment author: jacoblyles 18 July 2012 07:06:17PM *  1 point [-]

First, I don't buy the process of summing utilons across people as a valid one. Lots of philosophers have objected to it. This is a bullet-biting club, and I get that. I'm just not biting those bullets. I don't think 400 years of criticism of Utilitarianism can be solved by biting all the bullets. And in Eliezer's recent writings, it appears he is beginning to understand this. Which is great. It is reducing the odds he becomes a moral monster.

Second, I value things other than maximizing utilons. I got the impression that Eliezer/Less Wrong agreed with me on that from the Complex Values post and posts about the evils of paperclip maximizers. So great evils are qualitatively different to me from small evils, even small evils done to a great number of people!

I get what you're trying to do here. You're trying to demonstrate that ordinary people are innumerate, and you all are getting a utility spike from imagining you're more rational than them by choosing the "right" (naive hyper-rational utilitarian-algebraist) answer. But I don't think it's that simple when we're talking about morality. If it were, the philosophical project that's lasted 2500 years would finally be over!

Comment author: jacoblyles 18 July 2012 09:08:28AM *  4 points [-]

I was very surprised to find that a supporter of the Complexity of Value hypothesis and the author who warns against simple utility functions advocates torture using simple pseudo-scientific utility calculus.

My utility function has constraints that prevent me from doing awful things to people, unless it would prevent equally awful things done to other people. That this is a widely shared moral intuition is demonstrated by the reaction in the comments section. Since you recognize the complexity of human value, my widely-shared preferences are presumably valid.

In fact, the mental discomfort caused by people who heard of the torture would swamp the disutility from the dust specks. Which brings us to an interesting question - is morality carried by events or by information about events? If nobody else knew of my choice, would that make it better?

For a utilitarian, the answer is clearly that the information about morally significant events is what matters. I imagine so-called friendly AI bots built on utilitarian principles doing lots of awful things in secret to achieve its ends.

Also, I'm interested to hear how many torturers would change their mind if we kill the guy instead of just torturing him. How far does your "utility is all that matters" philosophy go?

Comment author: jacoblyles 18 July 2012 08:04:55AM *  0 points [-]

Certain self-consistent metaphysics and epistemologies lead you to belief in God. And a lot of human emotions do too. If you eliminated all the religions in the world, you would soon have new religions with 1) smart people accepting some form of philosophy that leads them to theism 2) lots of less smart people forming into mutually supporting congregations. Hopefully you get all the "religion of love" stuff from Christianity (historically a rarity) and the congregations produce public goods and charity.

Comment author: jacoblyles 18 July 2012 07:38:36AM 0 points [-]

What makes us think that AI would stick with the utility function they're given? I change my utility function all the time, sometimes on purpose.

Comment author: jacoblyles 05 February 2012 09:23:54AM 4 points [-]

"Long-term monogamy should not be done on the pretense that attraction and arousal for one's partner won't fade. It will."

This is precisely the point of monogamy. Polyamory/sleeping around is a young man's game. Long-term monogamy is meant to maintain strong social units throughout life, long after the thrill is gone.

Comment author: conchis 20 March 2009 07:07:40PM *  0 points [-]

"Eliezer is basically lamenting that when people behave rationally, they refuse to act against their self-interest, and damn it, it's hurting the rational tribe. That's informative, and sort of my point."

So if that's Eliezer's point, and it's also your point, what is it that you actually disagree about?

I take Eliezer to be saying that sometimes rational individuals fail to co-operate, but that things needn't be so. In response, you seem to be asking him to prove that rational individuals must co-operate - when he already appears to have accepted that this isn't true.

Isn't the relevant issue whether it is possible for rational individuals to co-operate? Provided we don't make silly mistakes like equating rationality with self-interest, I don't see why not - but maybe this whole thread is evidence to the contrary. ;)

Comment author: jacoblyles 20 March 2009 07:14:52PM *  1 point [-]

My point isn't exactly clear for a few reasons. First, I was using this post opportunistically to explore a topic that has been on my mind for awhile. Secondly, Eliezer makes statements that sometimes seem to support the "truth = moral good = prudent" assumption, and sometimes not.

He's provided me with links to some of his past writing, I've talked enough, it is time to read and reflect (after I finish a paper for finals).

View more: Prev | Next