Comment author: Vaniver 20 July 2011 12:22:54AM *  3 points [-]

But since one common FAI proposal is "find out human preferences, and then do those", if it turns out human preferences don't really exist in a coherent way, that sounds like an important thing to know.

I would think that knowing evo psych is enough to realize this is a dodgy approach at best.

Comment author: TimFreeman 12 August 2011 08:53:02PM *  1 point [-]

I would think that knowing evo psych is enough to realize [having an FAI find out human preferences, and then do them] is a dodgy approach at best.

I don't see the connection, but I do care about the issue. Can you attempt to state an argument for that?

Human preferences are an imperfect abstraction. People talk about them all the time and reason usefully about them, so either an AI could do the same, or you found a counterexample to the Church-Turing thesis. "Human preferences" is a useful concept no matter where those preferences come from, so evo psych doesn't matter.

Similarly, my left hand is an imperfect abstraction. Blood flows in, blood flows out, flakes of skin fall off, it gets randomly contaminated from the environment, and the boundaries aren't exactly defined, but nevertheless it generally does make sense to think in terms of my left hand.

If you're going to argue that FAI defined in terms of inferring human preferences can't work, I hope that isn't also going to be an argument that an AI can't possibly use the concept of my left hand, since the latter conclusion would be absurd.

Comment author: kragensitaker 11 August 2011 08:58:42PM *  4 points [-]

If you can opt out of it, it's not mandatory! You could get the best of both worlds, though: vitrify your head and donate the rest of your body. The only loss is, I think, your corneas.

Comment author: TimFreeman 11 August 2011 09:07:53PM 10 points [-]

The process of vitrifying the head makes the rest of the body unsuitable for organ donations. If the organs are extracted first, then the large resulting leaks in the circulatory system make perfusing the brain difficult. If the organs are extracted after the brain is properly perfused, they've been perfused too, and with the wrong substances for the purposes of organ donation.

Comment author: TimFreeman 11 August 2011 01:04:49PM 2 points [-]

If "humility" can be used to justify both activities and their opposites so easily, perhaps it's a useless concept and should be tabooed.

Comment author: potato 08 August 2011 07:25:52AM *  2 points [-]

That's suggestion five on the list:

Make sure that each CSA above the lowest level actually has "could", "should", and "would" labels on the nodes in its problem space, and make sure that those labels, their values, and the problem space itself can be reduced to the managing of the CSAs on the level below.

Figuring out exactly how it is that our preferences, i.e., our utility function, emerge from the managing of our subagents is my main motivation for suggesting the construction of a parallel decision theory, as well as understanding how our problem space emerges from the managing of other CSAs.

Comment author: TimFreeman 09 August 2011 12:13:32PM 0 points [-]

Make sure that each CSA above the lowest level actually has "could", "should", and "would" labels on the nodes in its problem space, and make sure that those labels, their values, and the problem space itself can be reduced to the managing of the CSAs on the level below.

That statement would be much more useful if you gave a specific example. I don't see how labels on the nodes are supposed to influence the final result.

There's a general principle here that I wish I could state well. It's something like "general ideas are easy, specific workable proposals are hard, and you're probably wasting people's time if you're only describing a solution to the easy parts of the problem".

One cause of this is that anyone who can solve the hard part of the problem can probably already guess the easy part, so they don't benefit much from you saying it. Another cause is that the solutions to the hard parts of the problem tend to have awkward aspects to them that are best dealt with by modifying the easy part, so a solution to just the easy part is sure to be unworkable in ways that can't be seen if that's all you have.

I have this issue with your original post, and most of the FAI work that's out there.

Comment author: TimFreeman 08 August 2011 04:48:04AM 9 points [-]

Well, one story is that humans and brains are irrational, and then you don't need a utility function or any other specific description of how it works. Just figure out what's really there and model it.

The other story is that we're hoping to make a Friendly AI that might make rational decisions to help people get what they want in some sense. The only way I can see to do that is to model people as though they actually want something, which seems to imply having a utility function that says what they want more and what they want less. Yes, it's not true, people aren't that rational, but if a FAI or anyone else is going to help you get what you want, it has to model you as wanting something (and as making mistakes when you don't behave as though you want something).

So it comes down to this question: If I model you as using some parallel decision theory, and I want to help you get what you want, how do I extract "what you want" from the model without first somehow converting that model to one that has a utility function?

Comment author: Raw_Power 13 July 2011 08:43:28PM 0 points [-]

It's not pseudo profound, but, like The Matrix, it has a lot of window-dressing and pomous wanking around absolutely legitimate questions. It's also frustrating in that many of the questions are asked, but few are resolved. And they're mainly a framework for the character arcs to develop. Eva has a very simple plot, which it doses very carefully in order to make it more interesting, so that it comes off as a Jigsaw puzzle. The most interesting thing is how the characters evolve and... really, I don't want to spoil anything, but you should definitely give it a try: it's a character story where the characters are extremely human, layered, and rich, and their stories are extremely poignantes.

If you don't want to watch the original, all I can tell you is, the "inexplicable" turns out to be "not explained yet". Everything will be revealed in due time. As for why it is interesting... well, if you watch EVA, and especially the final movie, The End Of Evangelion, you might identify a lot with Shinji, put a lot of yourself into him. This is especially true if you watch it as a teenager of the same age. And then... well, stuff happens to him, and to you, by proxy. Seeing him well-adjusted, happy, strong, while still having the same fundamental character traits... it's a very intense feeling.

Comment author: TimFreeman 04 August 2011 04:59:29AM 2 points [-]

Okay, I watched End of Evangelion and a variety of the materials leading up to it. I want my time back. I don't recommend it.

Comment author: CarlShulman 09 December 2010 09:03:49AM *  18 points [-]

I find this whole line of conversation fairly ludicrous, but here goes:

Number 1. Time-inconsistency: we have different reactions about an immediate certainty of some bad than a future probability of it. So many people might be willing to go be a health worker in a poor country where aid workers are commonly (1 in 10,000) raped or killed, even though they would not be willing to be certainly attacked in exchange for 10,000 times the benefits to others. In the actual instant of being tortured anyone would break, but people do choose courses of action that carry risk (every action does, to some extent), so the latter is more meaningful for such hypotheticals.

Number 2. I have driven and flown thousands of kilometers in relation to existential risk, increasing my chance of untimely death in a car accident or plane crash, so obviously I am willing to take some increased probability of death. I think I would prefer a given chance of being tortured to a given chance of death, so obviously I care enough to take at least some tiny risk from what I said above. As I also said above, I'm not willing to make very big sacrifices (big probabilities of such nasty personal outcomes) for tiny shifts in probabilities of big impersonal payoffs (like existential risk reduction). In realistic scenarios, that's what "the cause" would refer to. I haven't made any verbal or explicit "precommitment" or promises or anything like that.

In sufficiently extreme (and ludicrously improbable) trolley-problem style examples, e.g. "if you push this button you'll be tortured for a week, but if you don't then the Earth will be destroyed (including all your loved ones) if this fair coin comes up heads, and you have incredibly (impossibly?) good evidence that this really is the setup" I hope I would push the button, but in a real world of profound uncertainty, limited evidence, limited personal power (I am not Barack Obama or Bill Gates), and cognitive biases, I don't expect that to ever happen. I also haven't made any promises or oaths about that.

I am willing to give of my time and effort, and forgo the financial rewards of a more lucrative career, in exchange for a chance for efficient do-gooding, interaction with interesting people who share my values, and a meaningful project. Given diminishing returns to money in rich countries today, and the ease of obtaining money for folk with high human capital, those aren't big sacrifices, if they are sacrifices at all.

Number 3. SIAIers love to be precise and analytical and consider philosophical thought experiments, including ethical ones. I think most have views pretty similar to mine, with somewhat varying margins. Certainly Michael Vassar, the head of the organization, is also keen on recognizing one's various motives and living a balanced life, and avoiding fanatics. Like me, he actively advocates Bostrom-like parliamentary model approaches to combining self-concern with parochial and universalist altruistic feelings.

I have never heard anyone making oaths or promises to make severe sacrifices.

Number 4. This is a pretty ridiculous question. I think that's fine and normal, and I feel more comfortable with such folk than the alternative. I think people should not exaggerate that do-gooding is the most important thing in their life lest they deceive themselves and others about their willingness to make such choices, which I criticized Roko for.

Comment author: TimFreeman 13 July 2011 05:17:45PM 2 points [-]

So many people might be willing to go be a health worker in a poor country where aid workers are commonly (1 in 10,000) raped or killed, even though they would not be willing to be certainly attacked in exchange for 10,000 times the benefits to others.

I agree with your main point, but the thought experiment seems to be based on the false assumption that the risk of being raped or murdered are smaller than 1 in 10K if you stay at home. Wikipedia guesstimates that 1 in 6 women in the US are on the receiving end of attempted rape at some point, so someone who goes to a place with a 1 in 10K chance of being raped or murdered has probably improved their personal safety. To make a better thought experiment, I suppose you have to talk about the marginal increase in rape or murder rate when working in the poor country when compared to staying home, and perhaps you should stick to murder since the rape rate is so high.

Comment author: Raw_Power 09 July 2011 11:12:26AM 0 points [-]

Hm, reading the original EVA is not compulsory, the story stands very well on its own... but since you ask, I heartily recommend you watch EVA and Gurren Lagann. They are both flawed, but they are still very good, and very memorable.

Comment author: TimFreeman 13 July 2011 04:53:45PM *  1 point [-]

The story isn't working for me. A boy or novice soldier, depending on how you define it, is inexplicably given the job of running a huge and difficult-to-use robot to fight with a sequence of powerful similarly huge aliens while trying not to do too much collateral damage to Tokyo in the process. In the original, I gather he was an unhappy boy. In this story, he's a relatively well-adjusted boy who hallucinates conversations with his Warhammer figurines. I don't see why I should care about this scenario or any similar scenarios, but maybe I'm missing something.

Can someone who read this or watched the original say something interesting that happens in it? Wikipedia mentions profound philosophical questions about the nature of reality, but it also mentions that the ending is widely regarded as incomprehensible. The quote about how every possible statement sounds profound if you get the rhetoric right seems to apply here. I don't want to invest multiple hours to end up reading (or watching) some pseudo-profound nonsense.

Comment author: TimFreeman 10 July 2011 06:16:47PM 0 points [-]

Your strength as a rationalist is your ability to be more confused by fiction than by reality.

Does that lead to the conclusion that Newcomb's problem is irrelevant? Mind-reading aliens are pretty clearly fiction. Anyone who says otherwise is much more likely to be schizophrenic than to have actual information about mind-reading aliens.

Comment author: Raw_Power 09 July 2011 11:51:55AM *  1 point [-]

an inability to properly argue your point should have you questioning your point instead.

When dealing with trolls, whether on the Internet or in Real Life, no matter how absolutely damn sure you are of your point, you have no time to unravel their bullshit for what it is, and if you try it you will only bore your audience and exhaust their patience. Debates aren't battles of truth: there's publishing papers and articles for that. Debates are battles of status. If you manage to come off as the one with higher status, people will listen more to what you said during the debate, and, more importantly, to what you said afterwards.

A very interesting way of taking advantage of this and neutralizing the effects of the dirty fighting would be to immediately afterwards publish a play-by-play analysis of the discussion, using the opportunity as an occasion to teach those who were impressed by you and went to see your work how debate really works. You could even go so far as actually listing the arguments you and your opponents use, and openly admit it if your opponent's arguments are good enough that they have caused you to actually undertake a Bayesian update. That way, you show that:

*You're smart, witty, and charismatic enough to win the debate.

*You're rational, honest, and moral enough to admit to the truth afterwards.

Comment author: TimFreeman 10 July 2011 06:13:00PM 0 points [-]

When dealing with trolls, whether on the Internet or in Real Life, no matter how absolutely damn sure you are of your point, you have no time to unravel their bullshit for what it is, and if you try it you will only bore your audience and exhaust their patience. Debates aren't battles of truth: there's publishing papers and articles for that. Debates are battles of status.

I agree. There's also the scenario where you're talking to a reasonable person for the purpose of figuring out the truth better than either of you could do alone. That's useful, and it's important to be able to distinguish that from debating with trolls for the purpose of gaining status. Trolls can be recognized by how often they use rhetoric that obviously isn't truth-seeking, and Schopenhauer is very good for that.

Well, actually, on the Internet you never gain status by debating with trolls. Even if I win an argument, I lose status to the extent my behavior justifies the conclusion "Tim wastes time posting to (LessWrong|SlashDot|whatever) instead of doing anything useful."

My ability to identify and stonewall trolls varies. Sometimes I catch them saying something silly and refuse to continue unless they correct themselves, and that stops the time-waste pretty quickly. Sometimes I do three-strikes-and-your-out, and the time-waste stops reasonably soon. Sometimes it takes me a long time to figure out if they're a troll, especially if they're hinting that they know something worthwhile. I wish I had a more stable rule of thumb for doing this right. Any suggestions?

View more: Prev | Next