Comment author: torekp 17 March 2013 03:16:48PM 1 point [-]

It's a leap of faith to suppose that even our universe, never mind levels I-III, is exhausted by its mathematical properties, as opposed to simply mathematically describable. And I don't really see what it buys you. I suppose it's equally a leap of faith to suppose that our universe has more properties than that, but I just prefer not to leap at all.

Comment author: ESRogs 14 May 2016 08:11:05AM 1 point [-]

What would it mean for our universe not to be exhausted by its mathematical properties? Isn't whether a property seems mathematical just a function of how precisely you've described it?

Comment author: paulfchristiano 19 March 2016 09:04:51PM 2 points [-]

In that case, there would be severe principle-agent problems, given the disparity between power/intelligence of the trainer/AI systems and the users. If I was someone who couldn't directly control an AI using your scheme, I'd be very concerned about getting uneven trades or having my property expropriated outright by individual AIs or AI conspiracies, or just ignored and left behind in the race to capture the cosmic commons. I would be really tempted to try another AI design that does purport to have the AI serve my interests directly, even if that scheme is not as "safe".

Are these worse than the principal-agent problems that exist in any industrialized society? Most humans lack effective control over many important technologies, both in terms of economic productivity and especially military might. (They can't understand the design of a car they use, they can't understand the programs they use, they don't understand what is actually going on with their investments...) It seems like the situation is quite analogous.

Moreover, even if we could build AI in a different way, it doesn't seem to do anything to address the problem, since it is equally opaque to an end user who isn't involved in the AI development process. In any case, they are in some sense at the mercy of the AI developer. I guess this is probably the key point---I don't understand the qualitative difference between being at the mercy of the software developer on the one hand, and being at the mercy of the software developer + the engineers who help the software run day-to-day on the other. There is a slightly different set of issues for monitoring/law enforcement/compliance/etc., but it doesn't seem like a huge change.

(Probably the rest of this comment is irrelevant.)

To talk more concretely about mechanisms in a simple example, you might imagine a handful of companies who provide AI software. The people who use this software are essentially at the mercy of the software providers (since for all they know the software they are using will subvert their interests in arbitrary ways, whether or not there is a human involved in the process). In the most extreme case an AI provider could effectively steal all of their users' wealth. They would presumably then face legal consequences, which are not qualitatively changed by the development of AI if the AI control problem is solved. If anything we expect the legal system and government to better serve human interests.

We could talk about monitoring/enforcement/etc., but again I don't see these issues as interestingly different from the current set of issues, or as interestingly dependent on the nature of our AI control techniques. The most interesting change is probably the irrelevance of human labor, which I think is a very interesting issue economically/politically/legally/etc.

I agree with the general point that as technology improves a singleton becomes more likely. I'm agnostic on whether the control mechanisms I describe would be used by a singleton or by a bunch of actors, and as far as I can tell the character of the control problem is essentially the same in either case.

I do think that a singleton is likely eventually. From the perspective of human observers, a singleton will probably be established relatively shortly after wages fall below subsistence (at the latest). This prediction is mostly based on my expectation that political change will accelerate alongside technological change.

Comment author: ESRogs 15 April 2016 03:51:52AM 0 points [-]

I agree with the general point that as technology improves a singleton becomes more likely. I'm agnostic on whether the control mechanisms I describe would be used by a singleton or by a bunch of actors, and as far as I can tell the character of the control problem is essentially the same in either case.

I wonder -- are you also relatively indifferent between a hard and slow takeoff, given sufficient time before the takeoff to develop ai control theory?

(One of the reasons a hard takeoff seems scarier to me is that it is more likely to lead to a singleton, with a higher probability of locking in bad values.)

Comment author: Lumifer 29 January 2016 05:39:06PM 2 points [-]

An interesting comment:

The European champion of Go is not the world champion, or even close. The BBC, for example, reported that “Google achieves AI ‘breakthrough’ by beating Go champion,” and hundreds of other news outlets picked up essentially the same headline. But Go is scarcely a sport in Europe; and the champion in question is ranked only #633 in the world. A robot that beat the 633rd-ranked tennis pro would be impressive, but it still wouldn’t be fair to say that it had “mastered” the game. DeepMind made major progress, but the Go journey is still not over; a fascinating thread at YCombinator suggests that the program — a work in progress — would currently be ranked #279.

Comment author: ESRogs 31 January 2016 06:24:38AM 0 points [-]

It will be interesting to see how much progress they've made since October.

My guess is that they think they're going to win (see for example David Silver's "quiet confidence" here: https://www.youtube.com/watch?v=g-dKXOlsf98&t=5m9s).

Comment author: ESRogs 09 December 2015 10:15:10AM 10 points [-]

Gwern has written an article for Wired, allegedly revealing the true identity of Satoshi Nakamoto:

http://www.wired.com/2015/12/bitcoins-creator-satoshi-nakamoto-is-probably-this-unknown-australian-genius/

Comment author: ESRogs 11 December 2015 08:26:42AM 1 point [-]

Follow-up -- after we've all had some time to think about it, I think this is the best explanation for who this would-be SN is:

https://www.reddit.com/r/Bitcoin/comments/3w9xec/just_think_we_deserve_an_explanation_of_how_craig/cxuo6ac

Comment author: ESRogs 09 December 2015 10:15:10AM 10 points [-]

Gwern has written an article for Wired, allegedly revealing the true identity of Satoshi Nakamoto:

http://www.wired.com/2015/12/bitcoins-creator-satoshi-nakamoto-is-probably-this-unknown-australian-genius/

Comment author: moridinamael 04 December 2015 03:40:54PM *  6 points [-]

This advice may go against other advice, but it's a tactic that has served me well: in making early-career decisions, such as your choice of major, always ask yourself which choice preserves future options.

For example, let's say you are considering a major, and you are equally interested in Architecture, Literature, and Engineering as careers.

Under my analysis, I would ask, which of these choices preserves the most options?

If you choose to pursue a degree in Literature, it is unlikely that you will be able to parley those skills into any kind of job in Architecture or in Engineering.

If you choose Architecture, you will find it very difficult (though not entirely impossible) to switch into Engineering for a graduate degree. However, you may find that you can try to pivot into some kind of Literary existence more easily.

If you choose Engineering, you'll find that Architectural schools will be eager to accept you for a graduate program, and the difficulty of switching from Engineering to a Literature program will probably be equal to the difficulty of switching from Architecture.

So, under this analysis, Engineering is the choice that preserves the most future options. At the point of choosing your college major, you're too young to be screening off future possibilities. Unless you're completely gung-ho about Literature, and feel a real certainty about what you want, it's best to keep more cards in your hand and let yourself make that exclusionary choice when you're older and wiser.

You may find that reading the classics in your spare time and writing a little bit of fiction now and then more than satisfies your Literary impulses, in which case, you'll be glad you didn't commit yourself to it as a career.

Conversely, if you commit to Engineering and find that you hate it, it's always easier to pivot to the other options.

As a general rule, things that are perceived as more difficult are easier to pivot away from, because the admissions gatekeepers for the perceived-as-less-difficult options will find you impressive due to where you're coming from. This heuristic is valid at all levels, for example, if you decide to go the Engineering route, choose the subdiscipline of Engineering that everybody else says is the hardest, scariest one. You can always punt to one of the easier ones if you don't find it to be a good fit, but it's much harder to go uphill from where you start.

Comment author: ESRogs 07 December 2015 10:42:56PM 1 point [-]

As a complement to this advice (which I think is good), it's important to make sure you still explore. Don't be so worried about making sure you do the thing that maximizes optionality that you're afraid to fail and don't try things.

So if you think you should study math rather than econ (as per Kaj's comment), then start with math as your default, but make sure to also take an econ class to see if you're so much more interested in it / better at it that it's worth it to specialize.

Comment author: Fluttershy 02 December 2015 10:48:36AM 7 points [-]

Four years ago, I asked three members of my close family who were likely to give me something for Christmas to make a donation to GiveWell/AMF (GiveWell's top charity at that point) instead of getting something for me. This wasn't burdensome at all for me, because I didn't have many unmet material needs at this point.

Anyways, in my case, it turned out that my upper middle class American relatives, who were culturally "normal", rather than being culturally close to any EA/Silicon Valley/rationality circles, were quite offended by this suggestion. This may have had something to do with my presentation--I don't remember myself being particularly good at politics or speaking back then, and I tried to be nice, but perhaps I was too bold, or too culturally insensitive. Still, I was quite surprised at how poorly my request was received.

Of the three family members I talked to, two told me that I was being unrealistic, and that I needed to realize how things worked in the real world, or something like that. They got me something comparable to what they'd gotten me the previous year. The third one actually made a donation to the Carter Center, but only after bemoaning how I didn't appreciate how hard making money was in the real world--I think they had liked the Carter Center because a friend had worked there, or something.

A couple family members I hadn't talked to heard about my request, and I later heard that one had been talking with other members of my family about how she had become "very worried" that I was going to become "too altruistic". Another actually bought a chicken (or goat?) in my name through Heifer International. That was interesting, since I had thought that I been clear that I preferred GiveWell's top charities over other charities.

I guess that I completely stopped being vocal about EA after that point. Still, I've often wondered if the type of EAs who hold birthday and Christmas fundraisers are more, or less culturally normal-feeling to the average first-worlder than my family is.

(Also, since this comment is about how I'm terrible at understanding how to get the tone right on EA things, I apologize if the tone of this comment itself is somewhat off.)

Comment author: ESRogs 07 December 2015 10:31:07PM 2 points [-]

Four years ago, I asked three members of my close family who were likely to give me something for Christmas to make a donation to GiveWell/AMF (GiveWell's top charity at that point) instead of getting something for me.

Do you remember what you said? Was it written (like a facebook post) or spoken?

Comment author: ESRogs 07 December 2015 10:05:00PM 0 points [-]

Is there a deadline for when the survey will close?

Comment author: moridinamael 16 September 2015 08:48:46PM *  0 points [-]

I like this because it's something to point to when arguing with somebody with an obvious bias toward anthropomorphizing the agents.

You show them a model like this, then you say, "Oh, the agent can reduce its movement penalty if it first consumes this other orange glowing box. The orange glowing box in this case is 'humanity' but the agent doesn't care."

edit: Don't normally care about downvotes, but my model of LW does not predict 4 downvotes for this post, am I missing something?

Comment author: ESRogs 18 September 2015 12:07:03AM 0 points [-]

I was also surprised to see your comment downvoted.

That said, I don't think I see the value of the thing you proposed saying, since the framing of reducing the movement penalty by consuming an orange box which represents humanity doesn't seem clarifying.

Why does consuming the box reduce the movement penalty? Is it because, outside of the analogy, in reality humanity could slow down or get in the way of the AI? Then why not just say that?

I wouldn't have given you a downvote for it, but maybe others also thought your analogy seemed forced and are just harsher critics than I.

Comment author: Yvain 17 September 2015 05:33:50AM *  14 points [-]

I don't know if this solves very much. As you say, if we use the number 1, then we shouldn't wear seatbelts, get fire insurance, or eat healthy to avoid getting cancer, since all of those can be classified as Pascal's Muggings. But if we start going for less than one, then we're just defining away Pascal's Mugging by fiat, saying "this is the level at which I am willing to stop worrying about this".

Also, as some people elsewhere in the comments have pointed out, this makes probability non-additive in an awkward sort of way. Suppose that if you eat unhealthy, you increase your risk of one million different diseases by plus one-in-a-million chance of getting each. Suppose also that eating healthy is a mildly unpleasant sacrifice, but getting a disease is much worse. If we calculate this out disease-by-disease, each disease is a Pascal's Mugging and we should choose to eat unhealthy. But if we calculate this out in the broad category of "getting some disease or other", then our chances are quite high and we should eat healthy. But it's very strange that our ontology/categorization scheme should affect our decision-making. This becomes much more dangerous when we start talking about AIs.

Also, does this create weird nonlinear thresholds? For example, suppose that you live on average 80 years. If some event which causes you near-infinite disutility happens every 80.01 years, you should ignore it; if it happens every 79.99 years, then preventing it becomes the entire focus of your existence. But it seems nonsensical for your behavior to change so drastically based on whether an event is every 79.99 years or every 80.01 years.

Also, a world where people follow this plan is a world where I make a killing on the Inverse Lottery (rules: 10,000 people take tickets; each ticket holder gets paid $1, except a randomly chosen "winner" who must pay $20,000)

Comment author: ESRogs 17 September 2015 11:44:46PM 0 points [-]

if we use the number 1, then we shouldn't wear seatbelts, get fire insurance, or eat healthy to avoid getting cancer, since all of those can be classified as Pascal's Muggings

Isn't this dealt with in the above by aggregating all the deals of a certain probability together?

(amount of deals that you can make in your life that have this probability) * (PEST) < 1

Maybe the expected number of major car crashes or dangerous fires, etc that you experience are each less than 1, but the expectation for the number of all such things that happen to you might be greater than 1.

There might be issues with how to group such events though, since only considering things with the exact same probability together doesn't make sense.

View more: Next