All of kokotajlod's Comments + Replies

Sweet. I too will write something not about coronavirus.

2Ben Pace
I have had a helluva day preparing for the debate+meetup tomorrow. I'll try to get something out before I go to bed, might be short, might be about covid, sorry about that.

"It's a badly formulated question, likely to lead to confusion." Why? That's precisely what I'm denying.

"So, can you specify what this cluster is? Can you list the criteria by which a behaviour would be included in or excluded from this cluster? If you do this, you have defined blackmail."

That's precisely what I (Stuart really) am trying to do! I said so, you even quoted me saying so, and as I interpret him, Stuart said so too in the OP. I don't care about the word blackmail except as a means to an end; I'm trying to come up with criter... (read more)

Dunno what Username was thinking, but here's the answer I had in mind: "Why is it obvious? Because the Problem of Induction has not yet been solved."

You make it sound like those two things are mutually exclusive. They aren't. We are trying to define words so that we can understand and manipulate behavior.

"I don't know what blackmail is, but I want to make sure an AI doesn't do it." Yes, exactly, as long as you interpret it in the way I explained it above.* What's wrong with that? Isn't that exactly what the AI safety project is, in general? "I don't know what bad behaviors are, but I want to make sure the AI doesn't do them."

*"In other words there are a cluster of behaviors th... (read more)

0Lumifer
It's a badly formulated question, likely to lead to confusion. So, can you specify what this cluster is? Can you list the criteria by which a behaviour would be included in or excluded from this cluster? If you do this, you have defined blackmail.

"You want to understand and prevent some behaviors (in which case, start by tabooing culturally-dense words like "blackmail")"

In a sense, that's exactly what Stuart was doing all along. The whole point of this post was to come up with a rigorous definition of blackmail, i.e. to find a way to say what we wanted to say without using the word.

1Dagon
Now I'm really confused. Are you trying to define words, or trying to understand (and manipulate) behaviors? I'm hearing you say something like "I don't know what blackmail is, but I want to make sure an AI doesn't do it". This must be a misunderstanding on my part. I guess you might be trying to understand WHY some people don't like blackmail, so you can decide whether you want to to guard against it, but even that seems pretty backward.
0Lumifer
To avoid reinventing the wheel, I suggest looking at legal definitions of blackmail, as well as reading a couple of law research articles on the topic. Lawyers and judges had to deal with this issue for a very long time and they need to have criteria which produce a definite answer.

As I understand it, the idea is that we want to design an AI that is difficult or impossible to blackmail, but which makes a good trading partner.

In other words there are a cluster of behaviors that we do NOT want our AI to have, which seem blackmailish to us, and a cluster of behaviors that we DO want it to have, which seem tradeish to us. So we are now trying to draw a line in conceptual space between them so that we can figure out how to program an AI appropriately.

2Dagon
You and Stuart seem to have different goals. You want to understand and prevent some behaviors (in which case, start by tabooing culturally-dense words like "blackmail"). He wants to understand linguistic or legal definitions (so tabooing the word is counterproductive).
1Vaniver
I think there's also an element that if you know how to identify and resist blackmail yourself, not only do you win more, but also it's easier for you to choose to not blackmail others.

OH ok I get it now: "But clearly re-arranging terms doesn't change the expected utility, since that's just the sum of all terms." That's what I guess I have to deny. Or rather, I accept that (I agree that EU = infinity for both A and B) but I think that since A is better than B in every possible world, it's better than B simpliciter.

The reshuffling example you give is an example where A is not better than B in every possible world. That's the sort of example that I claim is not realistic, i.e. not the actual situation we find ourselves in. Why? W... (read more)

Again, thanks for this.

"The problem with your solution is that it's not complete in the formal sense: you can only say some things are better than other things if they strictly dominate them, but if neither strictly dominates the other you can't say anything."

As I said earlier, my solution is an argument that in every case there will be an action that strictly dominates all the others. (Or, weaker: that within the set of all hypotheses of probability less than some finite N, one action will strictly dominate all the others, and that this action w... (read more)

0ike
I just moved the right hand side down by two spaces. The sum still stays the same, but the relative inequality flips. Why would you think that? I don't really see where you argued for that, could you point me at the part of your comments that said that?

This was helpful, thanks!

As I understand it, you are proposing modifying the example so that on some H1 through HN, choosing A gives you less utility than choosing B, but then thereafter choosing B is better, because there is some cost you pay which is the same in each world.

It seems like the math tells us that any price would be worth it, that we should give up an unbounded amount of utility to choose A over B. I agree that this seems like the wrong answer. So I don't think whatever I'm proposing solves this problem.

But that's a different problem than the... (read more)

0ike
The problem there, and the problem with Pascal's Mugging in general, is that outcomes with a tiny amount of probability dominate the decisions. A could be massively worse than B 99.99999% of the time, and still naive utility maximization says to pick B. One way to fix it is to bound utility. But that has its own problems. The problem with your solution is that it's not complete in the formal sense: you can only say some things are better than other things if they strictly dominate them, but if neither strictly dominates the other you can't say anything. I would also claim that your solution doesn't satisfy framing invariants that all decision theories should arguably follow. For example, what about changing the order of the terms? Let us reframe utility as after probabilities, so we can move stuff around without changing numbers. E.g. if I say utility 5, p:.01, that really means you're getting utility 500 in that scenario, so it adds 5 total in expectation. Now, consider the following utilities: 1<2 p:.5 2<3 p:.5^2 3<4 p:.5^3 n<n+1 p:.5^n ... etc. So if you're faced with choosing between something that gives you the left side or the right side, choose the right side. But clearly re-arranging terms doesn't change the expected utility, since that's just the sum of all terms. So the above is equivalent to: 1>0 p:.5 2>0 p:.5^2 3>2 p:.5^3 4>3 p:.5^4 n>n-1 p:.5^n So your solution is inconsistent if it satisfies the invariant of "moving around expected utility between outcomes doesn't change the best choice".

It's arbitrary, but that's OK in this context. If I can establish that this works when the ratio is 1 in a billion, or lower, then that's something, even if it doesn't work when the ratio is 1 in 10.

Especially since the whole point is to figure out what happens when all these numbers go to extremes--when the scenarios are extremely improbable, when the payoffs are extremely huge, etc. The cases where the probabilities are 1 in 10 (or arguably even 1 in a billion) are irrelevant.

Update: The conclusion of that article is that the expected utilities don't converge for any utility function that is bounded below by a computable, unbounded utility function. That might not actually be in conflict with the idea I'm grasping at here.

The idea I'm trying to get at here is that maybe even if EU doesn't converge in the sense of assigning a definite finite value to each action, maybe it nevertheless ranks each action as better or worse than the others, by a certain proportion.

Toy model:

The only hypotheses you consider are H1, H2, H3, ... etc. ... (read more)

1ike
In your example, how much should you spend to choose A over B? Would you give up an unbounded amount of utility to do so?

True. So maybe this only works in the long run, once we have more than 30 bits to work with.

Yes, but I don't think that's relevant. Any use of complexity depends on the language you specify it in. If you object to what I've said here on those grounds, you have to throw out Solomonoff, Kolmogorov, etc.

2Ramana Kumar
More specifically, it seems that your c must include information about how to interpret the X bits. Right? So it seems slightly wrong to say "R is the largest number that can be specified in X bits of information" as long as c stays fixed. c might grow as the specification scheme changes. Alternatively, you might just be wrong in thinking that 30 bits are enough to specify 3^^^^3. If c indicates that the number of additional universes is specified by a standard binary-encoded number, 30 bits only gets you about a billion.
2ChristianKl
It only takes less than 30 bits if your language supports the ^^^^ notation and that's not standard notation.

Yes, I've read it, but not at the level of detail where I can engage with it. Since it is costly for me to learn the math necessary to figure this out for good, I figured I'd put the basic idea up for discussion first just in case there was something obvious I overlooked.

Edit: OK, now I think I understand it well enough to say how it interacts with what I've been thinking. See my other comment .

I disagree with your characterization of 0. You say that it is incompatible with physicalism, but that seems false. Indeed it seems to be a very mainstream physicalist view to say "I am a physical object--my brain. So a copy of me would have the same experiences, but it would not be me."

1TAG
Seconded. Physics and physicalism allow qualitively identical things to be numerically distinct.
0turchin
Objection accepted.

How do I do step II? I can't seem to find the relevant debates. I found one debate with the same title as the minimum wage one I argued about, but I don't see my argument appearing there.

Yep. "The melancholy of haruhi suzumiya" can be thought of as an example of something in the same reference class.

This is an interesting idea! Some thoughts:

Doesn't acausal trade (like all trade) depend on enforcement mechanisms? I can see how two AI's might engage in counterfactual trade, since they can simulate each other and see that they self-modify to uphold the agreement, but I don't think a human would be able to do it.

Also, I'd like to hear more about motivations for engaging in counterfactual trade. I get the multiverse one, though I think that's a straightforward case of acausal trade rather than a case of counterfactual trade, since you would be trading with a really existing entity in another universe. But can you explain the second motivation more?

The point you raise is by far the strongest argument I know of against the idea.

However, it is a moral objection rather than a decision-theory objection. It sounds like you agree with me on the decision theory component of the idea: that if we were anthropically selfish, it would be rational for us to commit to making ancestor-simulations with afterlives. That's an interesting result in itself, isn't it? Let's go tell Ayn Rand.

When it comes to the morality of the idea, I might end up agreeing with you. We'll see. I think there are several minor considerations in favor of the proposal, and then this one massive consideration against it. Perhaps I'll make a post on it soon.

This is a formal version of a real-life problem I've been thinking about lately.

Should we commit to creating ancestor-simulations in the future, where those ancestor-simulations will be granted a pleasant afterlife upon what appears to their neighbors to be death? If we do, then arguably we increase the likelihood that we ourselves have a pleasant afterlife to look forward to.

1Chris_Leong
This is pretty much like one guy destroying evidence about global warming so that everyone else should predict a lower probability of a catastrophe. It fails for the same reasons.
2Vulture
I'm pretty sure there's something wrong with this argument, but I can't seem to put my finger on what it is. It reminds of this post, in a way.
1lackofcheese
It seems to me that you should only do this if everyone has utility functions that are completely anthropically selfish (i.e. they only care about their own subjective experience). Otherwise, wouldn't it be cruel to intentionally simulate a world with so many unpleasant characteristics that we could otherwise remove if we weren't focused on making the simulation subjectively indistinguishable from our own world? As such, I don't think we should commit to any such thing.

Thanks for the response. Yes, it depends on how much interaction I have with human beings and on the kind of people I interact with. I'm mostly interested in my own case, of course, and I interact with a fair number of fairly diverse, fairly intelligent human beings on a regular basis.

If you're a social butterfly who regularly talks with some of the smartest people in the world, the AI will probably struggle

Ah, but would it? I'm not so sure, that's why I made this post.

Yes, if everyone always said what I predicted, things would be obvious, but recall I... (read more)

1Armok_GoB
Another trick it could use is using chatbots most of the time, but swaping them out for real people only for the moments you are actually talking about deep stuff. Maybe you have deep emotional conversations with your family a few hours a week. Maybe once per year, you have a 10 hour intense discussion with Eliezer. That's not a lot out of 24 hours per day, the vast majority of the computing power is still going into simulating your brain. Edit: another; the chatbots might have some glaring failure modes if you say the wrong thing, unable to handle edge cases, but whenever you encounter then the sim is restored from a backup 10 min earlier and the specific bug is manually patched. If this went on for long enough the chatbots would become real people, and also bloat slow, but it hasn't happened yet. or maybe the patches that dont come up in long enoguh get commented out.

Yes, but I'm not sure there is a difference between an AI directly puppeting them, and an AI designing a chatbot to run as a subroutine to puppet them, at least if the AI is willing to monitor the chatbot and change it as necessary. Do you think there is?

Also, it totally is a fruitful line of thinking. It is better to believe the awful truth than a horrible lie. At least according to my values. Besides, we haven't yet established that the truth would be awful in this case.

I'm surprised that it sounded that way to you. I've amended my original post to clarify.

Yes, this is the sort of consideration I had in mind. I'm glad the discussion is heading in this direction. Do you think the answer to my question hinges on those details though? I doubt it.

Perhaps if I was extraordinarily unsuspicious, chatbots of not much more sophistication than modern-day ones could convince me. But I think it is pretty clear that we will need more sophisticated chatbots to convince most people.

My question is, how much more sophisticated would they need to be? Specifically, would they need to be so much more sophisticated that they wou... (read more)

2helltank
The thing is, if you get suspicious you don't immediately leap to the conclusion of chatbots. Nobody glances around, realizes everyone is bland and stupid and thinks," I've been fooled! An AI has taken over the world and simulated chatbots as human beings!" unless they suffer from paranoia. Your question, "How much more sophisticated would they need to be" is answered by the question "depends". If you live as a hermit in a cave up in the Himalayas, living off water from a mountain stream and eating nothing but what you hunt or gather with your bare hands, the AI will not need to use chatbots at all. If you're a social butterfly who regularly talks with some of the smartest people in the world, the AI will probably struggle(let's be frank; if we were introduced to a chatbot imitating Eliezer Yudkosky the difference would be fairly obvious). If you've interacted a lot with your friends and family, and have never been once suspicious that they are a chatbot, then with our current level of AI technology it is unlikely(but not impossible) that they are actually chatbots. [Please note that if everyone says what you expect them to say that would make it fairly obvious something is up, unless you happen to be a very, very good predictor of human behavior.]

That's exactly what I had in mind, although I did specify that the controller would never simulate anybody besides me to the level required to make them people.

3HungryHobo
Define "make them people". Because it sounds like what you're talking about is just the problem of other minds with "chatbot" substituted for "zombie"

In the picture you just drew, the ideal being is derived from a series of better beings, thus it is (trivially) easier to imagine a better being than to imagine an ideal being.

I see it differently: The ideal being maximizes all good qualities, whereas imperfect beings have differing levels of the various good qualities. Thus to compare a non-ideal being to an ideal being, we only need to recognize how the ideal being does better than the non-ideal being in each good quality. But to compare two non-ideal beings, we need to evaluate trade-offs between their ... (read more)

Ok, thanks.

I also don't see any reason to go from "the FAI doesn't care about identity* to "I shouldn't think identity exists."

I don't either, now that I think about it. What motivated me to make this post is that I realized that I had been making that leap, thanks to applying the heuristic. We both agree the heuristic is bad.

Why are we talking about a bad heuristic? Well, my past self would have benefited from reading this post, so perhaps other people would as well. Also, I wanted to explore the space of applications of this heuristic, to see if I had been unconsciously applying it in other cases without realizing it. Talking with you has helped me with that.

Hmm, okay. I'd be interested to hear your thoughts on the particular cases then. Are there any examples that you would endorse?

1Wes_W
I agree with the points about Boltzmann Brains and mind substrates. In those cases, though, I'm not sure the FAI heuristic saves you any work, compared to just directly asking what the right answer is. Almost certainly not true if taken verbatim; one of the critical traits of an FAI (as opposed to a regular AGI) is that certain traits must remain stable under self-improvement. An FAI would care very strongly about certain kinds of changes. But with a less literal reading, I can see what you're going for here - yes, an ideal FAI might be indifferent to copying/deletion except to the extent that those help or hinder its goals. I'm not sure how that belief, applied to oneself, cashes out to anything at all, at least not with current technology. I also don't see any reason to go from "the FAI doesn't care about identity" to "I shouldn't think identity exists." (Disclaimer: I am not a decision theorist. This part is especially likely to be nonsense.) You should use which one? The less snappy version is that TDT and UDT both have problem cases. We don't really know yet what an ideal decision theory looks like. Second, I doubt any human can actually implement a formal decision theory all the time, and doing it only part-time could get you "valley of bad rationality"-type problems. Third, I suspect you could easily run into problems like what you might get by saying "an ideal reasoner would use Solomonoff Induction, so I should too". That's a wonderful idea, except that even approximating it is computationally insane, and in practice you won't get to use any of the advantages that make Solomonoff Induction theoretically optimal. If you instead mean things like "an ideal FAI would cooperate in PD-like scenarios given certain conditions", then sure. But again, I'm not sure the FAI heuristic is saving you any work. A factual FAI might, for mere practical reasons. I don't see why an ideal FAI normatively should ignore them, though.

The fact that your post was upvoted so much makes me take it seriously; I want to understand it better. Currently I see your post as merely a general skeptical worry. Sure, maybe we should never be very confident in our FAI-predictions, but to the extent that we are confident, we can allow that confidence to influence our other beliefs and decisions, and we should be confident in some things to some extent at least (the alternative, complete and paralyzing skepticism, is absurd) Could you explain more what you meant, or explain what you think my mistake is in the above reasoning?

0polymathwannabe
Of course, Bayesians want to be Mr. degrees-of-belief Carneades, not Mr. know-nothing Berkeley. Far be it from me to suggest that we ought to stop making models. It just worried me that you were so willing to adjust your behavior based on inherently untrustworthy predictions. An acquaintance of mine liked to claim that superhuman intelligence was one of Superman's powers. The idea immediately struck me as contradictory: nothing Superman does will ever be indicative of superhuman intelligence as long as the scriptwriter is human. My point here is the same: Your model of an ideal FAI will fall short of accurately simulating the ideal just as much as your own mind falls short of being ideal.

It is easier to determine whether you are doing "better" than your current self than it is to determine how well you line up with a perceived ideal being.

Really? That doesn't seem obvious to me. Could you justify that claim?

0christianoudard
An "ideal" being is many layers of "better" than you are, whereas something that is simply better is only one layer better. To get to ideal, you would have to imagine someone better, then imagine what that person would consider better, and so forth, until you hit a state where there are no further improvements to be made.

Thanks!

But if the UFAI can't parlay that takes out much of the fun, and much of the realism too.

Also, if Hard Mode has no FAI tech at all, then no one will research AI on Hard Mode and it will just devolve into a normal strategy game.

Edit: You know, this proposal could probably be easily implemented as a mod for an existing RTS or 4X game. For example, imagine a Civilization mod that added the "AI" tech that allowed you to build a "Boxed AI" structure in your cities. This quadruples the science and espionage production of your city, at ... (read more)

2MugaSofer
Hmm, that's a good point. I'm just worried that people might view an additional player as much less of a threat than a superintelligent AI. Hence the necessity of making tech tree advancement random, with player actions only providing modifiers.

There are at least two distinct senses in which consciousness can be binary. The first sense is the kind you are probably thinking about: the range between e.g. insects, dogs, and humans, or maybe between early and late-stage Alzheimers.

The second sense is the kind that your interlocutors are (I surmise) thinking about. Imagine this: A being that is functionally exactly like you, and that is experiencing exactly what you are experiencing, except that it is experiencing everything "only half as much." It still behaves the same way as you, and it s... (read more)

0A1987dM
IOW halfway between a person and a p-zombie?
5Sophronius
I don't think the first sense and the second are mutually exclusive. A dog has half as much processing power as a human = a dog can think the same thoughts but only at half the speed. A dog has half as much consciousness as a human = a dog is only half as aware as a human of what's going on. And yes, I definitely think that this is how it works. For example, when I get up in the morning I am much less aware of what's going on than when I am fully awake. Sometimes I pause and go "wait what am I doing right now?" And of course there's those funny times when you end up going to your old school by accident because you're not aware of what you're doing until it's too late... The only part I disagree with is "A being that is functionally exactly like you, and that is experiencing exactly what you are experiencing". A being that is only half as conscious is going to have a different brain, so will act differently. You definitely notice when someone is only at 50% consciousness. Finally, I submit that consciousness works just like will power: It is a limited resource which you can allocate to certain tasks/thoughts, and by training it you can get more of it. A good rationalist needs a large pool of consciousness.

At the moment, in order for a Creator/FAI team to win (assuming you're sticking with Diplomacy mechanics) they first >have to collect 18 supply centres between them and then have the AI transfer all its control back to the human; I don't >think even the friendliest of AIs would willingly rebox itself like that.

This is exactly what I had in mind. :) It should be harder for FAI to win than for UFAI to win, since FAI are more constrained. I think it is quite plausible that one of the safety measures people would try to implement in a FAI is "Wha... (read more)

This would be the ideal. Like I said though, I don't think I'll be able to make it anytime soon, or (honestly) anytime ever.

But yeah, I'm trying to design it to be simple enough to play in-browser or as an app, perhaps even as a Facebook game or something. It doesn't need to have good graphics or a detailed physics simulator, for example: It is essentially a board game in a computer, like Diplomacy or Risk. (Though it is more complicated than any board game could be)

I think that the game, as currently designed, would be an excellent source of fictional evidence for the notions of AI risk and AI arms races. Those notions are pretty important. :)

(utility monsters are awful, for some reason, even though by assumption they generate huge amounts of utility, oh dear!)

Utility monsters are awful, possibly for no reason whatsoever. That's OK. Value is complex. Some things are just bad, not because they entail any bad thing but just because they themselves are bad.

3satt
You're allowed to define utility monsters as terminally awful; it's just not going to convince me.

That's not something the average person will think upon hearing the term, especially since "AGI" tends to connote something very intelligent. I don't think it is a strong reason not to use it.

5AlexMennen
Actually, I think people often will think that when they hear the term. "Safety research" implies a focus on how to prevent a system from causing bad outcomes while achieving its goal, not on getting the system to achieve its goal in the first place, so "AGI Safety" sounds like research on how to prevent a not-necessarily-friendly AGI from becoming powerful enough to be dangerous, especially to someone who does not see an intelligence explosion as the automatic outcome of a sufficiently intelligent AI.

It is nice to see people thinking about this stuff. Keep it up, and keep us posted!

Have you read the philosopher Derek Parfit? He is famous for arguing for pretty much exactly what you propose here, I think.

Doubt: Doesn’t this imply that anthropic probabilities depend on how big a boundary the mind draws around stuff it considers “I”? Self: Yes. Doubt: This seems to render probability useless.

I agree with Doubt. If can make it 100% probable that I'll get superpowers tomorrow merely by convincing myself that only superpowered future-versions of me cou... (read more)

0Skeptityke
Yes, the answer to that question is mind dependent. But this isn't really a big deal. If a person decides that there is an 80% probability that a banana will appear in front of them, their P(A banana will appear in front of me) is 0.8. If a Midwesterner decides that they are guaranteed not to be in Kansas after waking up from a coma, their P(I am in Kansas) is about 0. If I decide that I am definitely not a brain in a vat, my P(I am a vat brain) is about 0. I suspect there is still some way to get non-stupid probabilities out of this mess from a series of observations by observer-moments, though I don't know how to do it. Intuitively, the problem with deciding that your P(I am a vat brain) is 0 is that your pre-existing series of observations could have been made by a vat brain.

No, the analogy I had in mind was this:

What People Saw: Acupuncture* being correlated with health, and [building things according to theories developed using the scientific method] being correlated with [having things that work very well]

What People Thought Happened: Acupuncture causing health and [building things according to theories developed using the scientific method] causing [having things that work very well]

What Actually Happened: Placebo effect and Placebo effect (in the former case, involving whatever mechanisms we think cause the placebo effect... (read more)

I didn't mean to imply that the placebo effect is a complete mystery. As you say, perhaps it is pretty well understood. But that doesn't touch my overall point which is that before modern medicine (and modern explanations for the placebo effect) people would have had plenty of evidence that e.g. faith healing worked, and that therefore spirits/gods/etc. existed.

0gwern
Yes, but they would have been dead wrong about everything they thought about the supernatural, not just the placebo effect. Thus if anyone were to suggest The example of the placebo effect would work against this theory: "people were completely and totally wrong about beliefs affecting reality before and it turned out to be some artifacts of selection bias / regression to the mean / relaxation / evolutionarily-based allocation of bodily resources, and so I disbelieve in your suggestion even more than I would just on its merits because clearly people are not good at this sort of thinking".

Similarly, modern theories about how to discover the habits of God in governing Creation (the Laws of Nature) are pretty sound as well. Or so theists say.

A better example than Amiens Cathedral would be the Placebo Effect. For most of human history, people with access to lots of data (but no notion of the Placebo Effect) had every reason to believe that e.g. witch doctors, faith healing, etc. was all correct.

Warning: Rampant speculation about a theory of low probability: Consider the corresponding theory about science. Maybe there is a Placebo Effect going ... (read more)

0gwern
I think you underestimate how much we know about how placebo effects operate.

If we are in a time loop we won't be trying to escape it, but rather exploit it.

For example: Suppose I find out that the entire local universe-bubble is in a time loop, and there is a way to build a spaceship that will survive the big crunch in time for the next big bang. Or something like that.

Well, I go to my backyard and start digging, and sure enough I find a spaceship complete with cryo-chambers. I get in, wait till the end of the universe, and then after the big bang starts again I get out and seed the Earth with life. I go on to create a wonderful c... (read more)

0Shmi
As long as every loop is identical to every other loop, why not. I don't know if you would call it "exploit", since, classically, your actions are predefined (and already happened infinitely many times), and quantum-mechanically, they are determined up to a chance (or, equivalently, every outcome happens in one of the Everett branches).

Thanks for the info. Hmm. What do you mean by "There is no entering or exiting the loop?" Could the loop be big enough to contain us already?

I'm not concerned about traveling backwards in time to change the past; I just want to travel backwards in time. In fact, I hope that I wouldn't be able to change the past. Consistency of that sort can be massively exploited.

0Shmi
Note the reversibility issue. After a complete loop, you end up in exactly the same state as you started with. All the dissipated energy and information must somehow return. Unless you are willing to wait for the Poincare recurrence, this is not very likely to happen. And in this case the wait time is so large as to be practically infinite.

That time-traveling universe is interesting. Physics question: Is it at all possible, never mind how likely, that our own universe contains closed timelike curves? What about closed timelike curves that we can feasibly exploit?

5Shmi
Very very very unlikely. Hawking once wrote a paper called Chronology protection conjecture, arguing that any time loop would self-destruct before forming. Even if they existed, it's not like you can travel them to do things. There is no entering or exiting the loop. Everything in the groundhog day has been there forever and will be there forever. Because there is no difference between "first time through the loop" and "n-th time through the loop". This is all predicated on classical general relativity. Quantum gravity may change things. But no one knows anything about quantum gravity.

Something about the name-dropping and phrasing in the "super-committee" line is off-putting. I'm not sure how to fix it, though.

Agreed. Maybe it is because it feels like you are talking down to us with the name-dropping? Perhaps this should be tested with people who are unfamiliar with LW and AI-related ideas, to see if they have the same reaction.

Yep, it means the same thing, or close enough. Of course there are measurement problems, but the intent behind the pay is for it to reward rational thinking in the usual sense.

Yeah. I should think about how to get around this, and glean useful information from their expertise.

The original AI will have a head start over all the other AI's, and it will probably be controlled by a powerful organization. So if its controllers give it real power soon, they will be able to give it enough power quickly enough that it can stop all the other AI's before they get too strong. If they do not give it real power soon, then shortly after there will be a war between the various new AI's being built around the world with different utility functions.

The original AI can argue convincingly that this war will be a worse outcome than letting it take... (read more)

3leplen
Do humans typically give power to the person with the most persuasive arguments? Is the AI going to be able to gain power simply by being right about things?

I would expect that to lead to the creation of AIs with a similar codebase but more or less tweaked utility functions

That's the point.

5Vulture
You'll have to expand on how exactly this would be beneficial to the original AI.

In the space of possible futures, it is much better than e.g. tiling the universe with orgasmium. So much better, in fact, that in the grand scheme of things it counts as OK.

-1Lumifer
I evaluate an "OK future" on an absolute scale, not relative. Relative scales lead you there.
Load More