Comment author: CarlShulman 02 February 2010 03:08:48PM 3 points [-]

The basement is the biggest, and matters more for goals that benefit strongly from more resources/security.

Comment author: arbimote 03 February 2010 06:04:46AM 0 points [-]

The basement is the biggest

I like that turn of phrase.

Comment author: cousin_it 02 February 2010 08:28:33PM *  5 points [-]

I think you misunderstood the question. Suppose the AI wants to prevent just 100 dustspeckings, but has reason enough to believe Dave will yield to the threat so no one will get tortured. Does this make the AI's behavior acceptable? Should we file this under "following reason off a cliff"?

Comment author: arbimote 03 February 2010 03:53:18AM 4 points [-]

I was about to point out that the fascinating and horrible dynamics of over-the-top threats are covered in length in Strategy of Conflict. But then I realised you're the one who made that post in the first place. Thanks, I enjoyed that book.

Comment author: MichaelVassar 03 February 2010 12:51:25AM 4 points [-]

Bad idea.

Comment author: arbimote 03 February 2010 03:39:00AM *  4 points [-]

It's much easier to limit output than input, since the source code of the AI itself provide it with some patchy "input" about what the external world is like. So there is always some input, even if you do not allow human input at run-time.

ETA: I think I misinterpreted your comment. I agree that input should not be unrestricted.

Comment author: jhuffman 03 February 2010 02:16:19AM 1 point [-]

You sir, have won this thread.

Comment author: arbimote 03 February 2010 03:28:09AM 5 points [-]

You sir, have made a gender assumption.

Comment author: orthonormal 03 February 2010 03:03:53AM -3 points [-]

Should have been an Open Thread comment, IMO.

Comment author: arbimote 03 February 2010 03:26:30AM 0 points [-]

Similar topics were discussed in an Open Thread.

Comment author: wedrifid 03 February 2010 01:35:37AM 0 points [-]

You just need to hope the room was made by an infallible carpenter and that you never gave the AI access to MacGyver.

Comment author: arbimote 03 February 2010 01:38:49AM 0 points [-]

Luckily digital constructs are easier to perfect that wooden ones. Although you wouldn't think so with the current state of most software.

Comment author: JamesAndrix 01 February 2010 07:59:50PM 5 points [-]

I gave up on trying to make a human-blind/sandboxed AI when I realized that even if you put it in a very simple world nothing like ours, it still has access to it own source code, or even just the ability to observe and think about it's own behavior.

Presumably any AI we write is going to be a huge program. That gives it lots of potential information about how smart we are and how we think. I can't figure out how to use that information, but I can't rule out that it could, and I can't constrain it's access to that information. (Or rather, if I know how to do that, I should go ahead and make it not-hostile in the first place.)

If we were really smart, we could wake up alone in a room and infer how we evolved.

Comment author: arbimote 03 February 2010 01:31:49AM 0 points [-]

It is difficult to constrain the input we give to the AI, but the output can be constrained severely. A smart guy could wake up alone in a room and infer how he evolved, but so long as his only link to the outside world is a light switch that can only be switched once, there is no risk that he will escape.

Comment author: Jayson_Virissimo 02 February 2010 07:44:16PM 3 points [-]

This is why you should make sure Dave holds a deontological ethical theory and not a consequentialist one.

Comment author: arbimote 03 February 2010 01:21:47AM 2 points [-]

If Dave holds a consequentialist ethical theory that only values his own life, then yes we are screwed.

If Dave's consequentialism is about maximizing something external to himself (like the probable state of the universe in the future, regardless of whether he is in it), then his decision has little or no weight if he is a simulation, but massive weight if he is the real Dave. So the expected value of his decision is dominated by the possibility of him being real.

Comment author: Kaj_Sotala 02 February 2010 04:39:52PM *  24 points [-]

Defeating Dr. Evil with self-locating belief is a paper relating to this subject.

Abstract: Dr. Evil learns that a duplicate of Dr. Evil has been created. Upon learning this, how seriously should he take the hypothesis that he himself is that duplicate? I answer: very seriously. I defend a principle of indifference for self-locating belief which entails that after Dr. Evil learns that a duplicate has been created, he ought to have exactly the same degree of belief that he is Dr. Evil as that he is the duplicate. More generally, the principle shows that there is a sharp distinction between ordinary skeptical hypotheses, and self-locating skeptical hypotheses.

(It specifically uses the example of creating copies of someone and then threatening to torture all of the copies unless the original co-operates.)

The conclusion:

Dr. Evil, recall, received a message that Dr. Evil had been duplicated and that the duplicate ("Dup") would be tortured unless Dup surrendered. INDIFFERENCE entails that Dr. Evil ought to have the same degree of belief that he is Dr. Evil as that he is Dup. I conclude that Dr. Evil ought to surrender to avoid the risk of torture.

I am not entirely comfortable with that conclusion. For if INDIFFERENCE is right, then Dr. Evil could have protected himself against the PDF's plan by (in advance) installing hundreds of brains in vats in his battlestation - each brain in a subjective state matching his own, and each subject to torture if it should ever surrender. (If he had done so, then upon receiving PDF's message he ought to be confident that he is one of those brains, and hence ought not to surrender.) Of course the PDF could have preempted this protection by creating thousands of such brains in vats, each subject to torture if it failed to surrender at the appropriate time. But Dr. Evil could have created millions...

It makes me uncomfortable to think that the fate of the Earth should depend on this kind of brain race.

Comment author: arbimote 03 February 2010 01:06:51AM *  2 points [-]

If we accept the simulation hypothesis, then there are already gzillions of copies of us, being simulated under a wide variety of torture conditions (and other conditions, but torture seems to be the theme here). An extortionist in our world can only create a relatively small number of simulations of us, relatively small enough that it is not worth taking them into account. The distribution of simulation types in this world bears no relation to the distribution of simulations we could possibly be in.

If we want to gain information about what sort of simulation we are in, evidence needs to come directly from properties of our universe (stars twinkling in a weird way, messages embedded in π), rather than from properties of simulations nested in our universe.

So I'm safe from the AI ... for now.

Comment author: magfrump 01 February 2010 07:00:07PM *  4 points [-]

According to some people we here at less wrong are good at determining the truth. Other people are notoriously not.

I don't know that Less Wrong is the appropriate venue for this, but I have felt for some time that I trust the truth-seeking capability here and that it could be used for something more productive than arguments about meta-ethics (no offense to the meta-ethicists intended). I also realize that people are fairly supportive of SIAI here in terms of giving spare cash away, but I feel like the community would be a good jumping-off point for a polling organization.

So I guess this leads to a few questions:

-Is anyone at LW currently involved with a polling firm?

-Is anyone (else) at LW interested in doing polls?

-Is LW an appropriate place to create a truth-seeking business, such as a pollster or a sponsor for studies?

None of these questions are immediate since I am a broke undergrad rather than an entrepreneur.

Comment author: arbimote 02 February 2010 06:17:32AM 3 points [-]

Here's an idea for how a LW-based commercial polling website could operate. Basically it is a variation on PredictionBook with a business model similar to TopCoder.

The website has business clients, and a large number of "forecasters" who have accounts on the website. Clients pay to have their questions added to the website, and forecasters give their probability estimates for whichever questions they like. Once the answer to a question has been verified, each forecaster is financially rewarded using some proper scoring rule. The more money assigned to a question, the higher the incentive for a forecaster to have good discrimination and calibration. Some clever software would also be needed to combine and summarize data in a way that is useful to clients.

The main advantage of this over other prediction markets is that the scoring rule encourages forecasters to give accurate probability estimates.

View more: Prev | Next