All of anon85's Comments + Replies

anon8500

I see you've not bothered reading any of my replies and instead just made up your own version in your head.

I read all of your replies. What are you referring to? Also, this is uncharitable/insulting.

Believe it or not there are a lot of people who'll do things like insist that that's not the case or insist that you just have to wish carefully enough hence the need for the article.

To be honest, I'm not sure what we're even disagreeing about. Like, sure, some genies are unsafe no matter how you phrase your wish. For other genies, you can just wish for ... (read more)

anon8500

If you have to speak "carefully enough" then you're taking a big risk though you may luck out and get what you want, they're not safe.

If your argument is that unless a powerful being is extremely safe, then they're not extremely safe, this is true by definition. Obviously, if a genie sometimes doesn't give you what you want, there is some risk that the genie won't give you what you want. I thought a more substantial argument was being made, though - it sounded like EY was claiming that saying "I wish for whatever I should wish for" i... (read more)

0HungryHobo
I see you've not bothered reading any of my replies and instead just made up your own version in your head. Your mom example falls quite cleanly into the third catagory if it doesn't fall cleanly into the first. Unless a powerful being understands your values well enough to take them into account and actually wants to take them into account then it's not extremely safe. Yes. Believe it or not there are a lot of people who'll do things like insist that that's not the case or insist that you just have to wish carefully enough hence the need for the article.
anon8500

Examples of what? Of hypothetical intelligent minds? I feel like there are examples all over fiction; consider genies themselves, which often grant wishes in a dangerous way (but you can sometimes get around it by speaking carefully enough). Again, I agree that some genies are never safe and some are always safe, but it's easy to imagine a genie which is safe if and only if you specify your wish carefully.

Anyway, do you concede the point that EY's article contains no arguments?

0HungryHobo
If you have to speak "carefully enough" then you're taking a big risk though you may luck out and get what you want, they're not safe. EY's article contains arguments, you just seem to have picked up on something that wasn't what he was arguing about. It's like someone started a speech with "Good evening ladies and gentlemen." and your criticism was that he failed to prove that it was evening, failed to prove that there was a mix of genders in the audience and that the entirety of the rest of the speech failed to contain any arguments about whether the men in the audience were in fact gentlemen. It contained a very clear and well made argument for why simply trying to word your wish carefully was a fools errand. You may notice how it starts with an overly complex wish from the "open source wish project". It then gives examples of how simply adding clauses to the wish to get your mother out doesn't help much because you value so many things as a human that you'd have to add so many thousands of disclaimers, clauses and rules that it would be insane while missing even one could mean disaster(from your point of view) which is extremely unsafe.
anon8500

I'm making 2 points:

  1. His metaphor completely fails conceptually, because I'm perfectly capable of imagining genies that fall outside the three categories.

  2. Perhaps the classification works in some other setting, such as AIs. However, the article never provided any arguments for this (or any arguments at all, really). Instead, there was one single example (seriously, just one example!) which was then extrapolated to all genies.

0HungryHobo
Ok, so, do you actually have any examples that fall outside the 3 categories: 1:Powerful+Safe 2:Powerful+Unsafe 3:Not very powerful such that it doesn't matter so much if they're safe or unsafe.
anon8500

At age 5 you could safely wish for "I wish for you to do what I should wish for" and at worst you'd be a little disappointed if what she came up with wasn't as fun as you'd have liked.

I would have gotten the wrong flavor of ice cream. It was strictly better to specify the flavor of ice cream I preferred. Therefore, the statement about the 3 types of genies is simply false. It might be approximately true in some sense, but even if it is, the article never gives any arguments in favor of that thesis, it simply gives one example.

2HungryHobo
Wait, to be clear, you're calling getting the wrong flavor of icecream a "safety" issue? Do you have any examples that actually fall outside the 3 types? Your mother is likely not powerful and nor is she a superintelligence. So far the only example you've given has fallen squarely in the third category but even if scaled up would probably fit quite well in the first. I'd also note that the claim you're taking issue with is a metaphor for explaining things, he's not claiming that magical genies actually exist of any category.
anon8530

That sounds pretty similar to a Deist's God, which created the universe but does not interfere thereafter. Personally, I'd just shave it off with Ocam's razor.

Also, it seems a little absurd to try to infer things about our simulators, even supposing they exist. After all, their universe can be almost arbitrarily different from ours.

1Alex Vermillion
If you're still at this level of confidence in the argument, I'd check Bostrom's argument somewhere online, it's pretty straightforwards over all. In particular, your above comment has a similar mental flavor to the argument: One of these is likely true: 1. The universe came into being as-is while you were reading this 2. The universe came into being a long time ago and then things happened that caused this state I select 1 because it's simpler to assume less time in the past. It seems like you're making a similar error, but I can't really tell if it's a form of some named error or not.
anon8520

Does the simulation hypothesis have any predictive power? If so, what does it predict? Is there any way to falsify it?

2turchin
It changes probability spectre of possible futures. It makes more probable that I live in interesting historical simulation and may be simulation with miracles. But this shift is not strong enough so I could conclude "Ups, my life is ordinary and boring and so it is real". The problem here is that creators of simulation could use all their power to prevent me from knowing that I am in simulation. That is why Popper-style scientific method need to be use here with caution.
anon85-20

Oh, yes, me too. I want to engage in one-shot PD games with entirelyuseless (as opposed to other people), because he or she will give me free utility if I sell myself right. I wouldn't want to play one-shot PDs against myself, in the same way that I wouldn't want to play chess against Kasparov.

By the way, note that I usually cooperate in repeated PD games, and most real-life PDs are repeated games. In addition, my utility function takes other people into consideration; I would not screw people over for small personal gains, because I care about their happiness. In other words, defecting in one-shot PDs is entirely consistent with being a decent human being.

anon85-10

Cool, so in conclusion, if we met in real life and played a one-shot PD, you'd (probably) cooperate and I'd defect. My strategy seems superior.

3gjm
And yet I somehow find myself more inclined to engage in PD-like interactions with entirelyuseless than with your good self.
anon8520

I never liked that article. It says "there are three types of genies", and then, rather than attempting to prove the claim or argue for it, it just provides an example of a genie for which no wish is safe. I mean, fine, I'm convinced that specific genie sucks. But there may well be other genies that don't know what you want but have the ability to give it to you if you ask (when I was 5 years old, my mom was such a genie).

2HungryHobo
Your mother was human and her goals were likely very tightly aligned with yours. She was probably an extremely safe genie and was likely more concerned with what you actually needed rather than what you asked for. If a 5 year old asked for a metric ton of candy and a bag of huge fireworks she's likely such a safe genie that she'd simply not grant your wish even if it was what you really wanted, even if she fully understood why you wanted it and even if she was fully capable of granting it. At age 5 you could safely wish for "I wish for you to do what I should wish for" and at worst you'd be a little disappointed if what she came up with wasn't as fun as you'd have liked.
anon8500

But since you're making it clear that your code is quite different, and in a particular way, I would defect against you.

You don't know who I am! I'm anonymous! Whoever you'd cooperate with, I might be that person (remember, in real life I pretend to have a completely different philosophy on this matter). Unless you defect against ALL HUMANS, you risk cooperating when facing me, since you don't know what my disguise will be.

0entirelyuseless
I will take that chance into account. Fortunately it is a low one and should hardly be a reason to defect against all humans.
anon8500

You can see which side of the room you are on, so you know which one you are.

If I can do this, then my clone and I can do different things. In that case, I can't be guaranteed that if I cooperate, my clone will too (because my decision might have depended on which side of the room I'm on). But I agree that the cloning situation is strange, and that I might cooperate if I'm actually faced with it (though I'm quite sure that I never will).

People don't actually have the same code, but they have similar code. If the code in some case is similar enough tha

... (read more)
0entirelyuseless
I would cooperate with you if I couldn't distinguish my code from yours, even if there might be minor differences, even in a one-shot case, because the best guess I would have of what you would do is that you would do the same thing that I do. But since you're making it clear that your code is quite different, and in a particular way, I would defect against you.
anon85-30

Yes. The universe is deterministic. Your actions are completely predictable, in principle. That's not unique to this thought experiment. That's true for every thing you do. You still have to make a choice. Cooperate or defect?

Um, what? First of all, the universe is not deterministic - quantum mechanics means there's inherent randomness. Secondly, as far as we know, it's consistent with the laws of physics that my actions are fundamentally unpredictable - see here.

Third, if I'm playing against a clone of myself, I don't think it's even a valid PD. Can th... (read more)

0entirelyuseless
Even if you're playing against a clone, you can distinguish the copies by where they are in space and so on. You can see which side of the room you are on, so you know which one you are. That means one of you can get utility without the other one getting it. People don't actually have the same code, but they have similar code. If the code in some case is similar enough that you can't personally tell the difference, you should follow the same rule as when you are playing against a clone.
anon85-20

Well there is no causal influence. Your opponent is deterministic. His choice may have already been made and nothing you do will change it. And yet the best decision is still to cooperate.

If his choice is already made and nothing I do will change it, then by definition my choice is already made and nothing I do will change it. That's why my "decision" in this setting is not even well-defined - I don't really have free will if external agents already know what I will do.

2Houshalter
Yes. The universe is deterministic. Your actions are completely predictable, in principle. That's not unique to this thought experiment. That's true for every thing you do. You still have to make a choice. Cooperate or defect?
anon8500

The most obvious example of cooperating due to acausal dependence is making two atom-by-atom-identical copies of an agent and put them in a one-shot prisoner's dilemma against each other. But two agents whose decision-making is 90% similar instead of 100% identical can cooperate on those grounds too, provided the utility of mutual cooperation is sufficiently large.

I'm not sure what "90% similar" means. Either I'm capable of making decisions independently from my opponent, or else I'm not. In real life, I am capable of doing so. The clone situa... (read more)

4Houshalter
I think you are making this more complicated than it needs to be. You don't need to worry about your code. All you need to know that it's an exact copy of you playing. And that he will make the same decision you do. No matter how hard you think about your "code" or wish he would make a different choice, he will just do the same thing about you. In real games with real humans, yes, usually. As I said, I don't think these cases are common enough to worry about. But I'm just saying they exist. But it is more general than just clones. If you know your opponent isn't exactly the same as you, but still follows the same decision algorithm in this case, the principle is still valid. If you cooperate, he will cooperate. Because you are both following the same process to come to a decision. Well there is no causal influence. Your opponent is deterministic. His choice may have already been made and nothing you do will change it. And yet the best decision is still to cooperate.
anon8500

I don't know enough about this to tell if (2) had more influence than (3) initially. I'm glad you agree that (2) had some influence, at least. That was the main part of my point.

How long did discussion of the Basilisk stay banned? Wasn't it many years? How do you explain that, unless the influence of (2) was significant?

4Houshalter
Imagine you are playing against a clone of yourself. Whatever you do, the clone will do the exact same thing. If you choose to cooperate, he will choose to cooperate. If you choose to defect, he chooses to defect. The best choice is obviously to cooperate. So there are situations where cooperating is optimal. Despite there not being any causal influence between the players at all. I think these kinds of situations are so exceedingly rare and unlikely they aren't worth worrying about. For all practical purposes, the standard game theory logic is fine. But it's interesting that they exist. And some people are so interested by that, that they've tried to formalize decision theories that can handle these situations. And from there you can possibly get counter-intuitive results like the basilisk.
4Rob Bensinger
Causation isn't necessary. You're right that correlation isn't quite sufficient, though! What's needed for rational cooperation in the prisoner's dilemma is a two-way dependency between A and B's decision-making. That can be because A is causally impacting B, or because B is causally impacting B; but it can also occur when there's a common cause and neither is causing the other, like when my sister and I have similar genomes even though my sister didn't create my genome and I didn't create her genome. Or our decision-making processes can depend on each other because we inhabit the same laws of physics, or because we're both bound by the same logical/mathematical laws -- even if we're on opposite sides of the universe. (Dependence can also happen by coincidence, though if it's completely random I'm not sure how'd you find out about it in order to act upon it!) The most obvious example of cooperating due to acausal dependence is making two atom-by-atom-identical copies of an agent and put them in a one-shot prisoner's dilemma against each other. But two agents whose decision-making is 90% similar instead of 100% identical can cooperate on those grounds too, provided the utility of mutual cooperation is sufficiently large. For the same reason, a very large utility difference can rationally mandate cooperation even if cooperating only changes the probability of the other agent's behavior from '100% probability of defection' to '99% probability of defection'. I disagree! "Code-sharing" risks confusing someone into thinking there's something magical and privileged about looking at source code. It's true this is an unusually rich and direct source of information (assuming you understand the code's implications and are sure what you're seeing is the real deal), but the difference between that and inferring someone's embarrassment from a blush is quantitative, not qualitative. Some sources of information are more reliable and more revealing than others; but the same un
anon8500

Just FYI, if you want a productive discussion you should hold back on accusing your opponents of fallacies. Ironically, since I never claimed that you claimed Eliezer engages in habitual banning on LW, your accusation that I made a strawman argument is itself a strawman argument.

Anyway, we're not getting anywhere, so let's disengage.

anon8500

What utility do you think is gained by discussing the basilisk?

An interesting discussion that leads to better understanding of decision theories? Like, the same utility as is gained by any other discussion on LW, pretty much.

Strawman. This forum is not a place where things get habitually banned.

Sure, but you're the one that was going on about the importance of the mindset and culture; since you brought it up in the context of banning discussion, it sounded like you were saying that such censorship was part of a mindset/culture that you approve of.

0ChristianKl
Not every discussion on LW has the same utility. You engage in a pattern of simplifying the subject and then complaining that your flawed understanding doesn't make sense. LW doesn't have a culture with habitual banning discussion. Claiming that it has it is wrong. I'm claiming that particular actions of Eliezer come out of being concerned about safety. I don't claim that Eliezer engages in habitual banning on LW because of those concerns. It's a complete strawman that you are making up.
-6mwengler
2Rob Bensinger
Defecting gives you a better outcome than cooperating if your decision is uncorrelated with the other players'. Different humans' decisions aren't 100% correlated, but they also aren't 0% correlated, so the rationality of cooperating in the one-shot PD varies situationally for humans. Part of the reason why humans often cooperate in PD-like scenarios in the real world is probably that there's uncertainty about how iterated the PD is (and our environment of evolutionary adaptedness had a lot more iterated encounters than once-off encounters). But part of the reason for cooperation is probably also that we've evolved to do a very weak and probabilistic version of 'source code sharing': we've evolved to (sometimes) involuntarily display veridical evidence of our emotions, personality, etc. -- as opposed to being in complete control of the information we give others about our dispositions. Because they're at least partly involuntary and at least partly veridical, 'tells' give humans a way to trust each other even when there are no bad consequences to betrayal -- which means at least some people can trust each other at least some of the time to uphold contracts in the absence of external enforcement mechanisms. See also Newcomblike Problems Are The Norm.
anon8520

I agree that resolving paradoxes is an important intellectual exercise, and that I wouldn't be satisfied with simply ignoring an ontological argument (I'd want to find the flaw). But the best way to find such flaws is to discuss the ideas with others. At no point should one assign such a high probability to ideas like Roko's basilisk being actually sound that one refuses to discuss them with others.

anon8500

It seems unlikely that they would, if their gun is some philosophical decision theory stuff about blackmail from their future. I don't expect that gun to ever fire, no matter how many times you click the trigger.

0hairyfigment
That is not what I said, and I'm also guessing you did not have a grandfather who taught you you gun safety.
anon8500

Probably not a quick fix, but I would definitely say Eliezer gives significant chances (say, 10%) to there being some viable version of the Basilisk, which is why he actively avoids thinking about it.

If Eliezer was just angry at Roko, he would have yelled or banned Roko; instead, he banned all discussion of the subject. That doesn't even make sense as a "slashing out" reaction against Roko.

3Rob Bensinger
It sounds like you have a different model of Eliezer (and of how well-targeted 'lashing out' usually is) than I do. But, like I said to V_V above: The point I was making wasn't that (2) had zero influence. It was that (2) probably had less influence than (3), and its influence was probably of the 'small probability of large costs' variety.
anon85-10

Somehow, blackmail from the future seems less plausible to me than every single one of your examples. Not sure why exactly.

4Richard_Kennaway
How plausible do you find TDT and related decision theories as normative accounts of decision making, or at least as work towards such accounts? They open whole new realms of situations like Pascal's Mugging, of which Roko's Basilisk is one. If you're going to think in detail about such decision theories, and adopt one as normative, you need to have an answer to these situations. Once you've decided to study something seriously, the plausibility heuristic is no longer available.
anon85-10

If you are a programmer and think your code is safe because you see no way things could go wrong, it's still not good to believe that it isn't plausible that there's a security hole in your code.

Let's go with this analogy. The good thing to do is ask a variety of experts for safety evaluations, run the code through a wide variety of tests, etc. The think NOT to do is keep the code a secret while looking for mistakes all by yourself. If you keep your code out of the public domain, it is more likely to have security issues, since it was not scrutinized by the public. Banning discussion is almost never correct, and it's certainly not a good habit.

0ChristianKl
No, if you don't want to use code you don't give the code to a variety of experts for safety evaluations but you simply don't run the code. Having a public discussion is like running the code untested on a mission critical system. What utility do you think is gained by discussing the basilisk? Strawman. This forum is not a place where things get habitually banned.
anon8510

Ideas that aren't proven to be impossible are possible. They don't have to be plausible.

anon85-10

That even seems to be false in Eliezer's case, and Eliezer definitely isn't 'LessWrong'.

It seems we disagree on this factual issue. Eliezer does think there is a risk of acausal blackmail, or else he wouldn't have banned discussion of it.

3hairyfigment
I believe he thinks that sufficiently clever idiots competing to shoot off their own feet will find some way to do so.
6Rob Bensinger
Sorry, I'll be more concrete; "there's a serious risk" is really vague wording. What would surprise me greatly is if I heard that Eliezer assigned even a 5% probability to there being a realistic quick fix to Roko's argument that makes it work on humans. I think a larger reason for the ban was just that Eliezer was angry with Roko for trying to spread what Roko thought was an information hazard, and angry people lash out (even when it doesn't make a ton of strategic sense).
anon85-20

I'm not the person you replied to, but I mostly agree with (a) and reject (b). There's no way you can could possibly know enough about a not-yet-existing entity to understand any of its motivations; the entities that you're thinking about and the entities that will exist in the future are not even close to the same. I outlined some more thoughts here.

1Richard_Kennaway
... Despite the other things I've said here, that is my attitude as well. But I recognise that when I take that attitude, I am not solving the problem, only ignoring it. It may be perfectly sensible to ignore a problem, even a serious one (comparative advantage etc.). But dissolving a paradox is not achieved by clinging to one of the conflicting thoughts and ignoring the others. (Bullet-swallowing seems to consist of seizing onto the most novel one.) Eliminating the paradox requires showing where and how the thoughts went wrong.
0[anonymous]
... Despite the other things I've said here, that is my attitude as well. But I recognise that when I take that attitude, I am not solving the problem, only ignoring it. It may be perfectly sensible to ignore a problem, even a serious one (comparative advantage etc.). But dissolving a paradox is not achieved merely by clinging to one of the conflicting thoughts and ignoring the others. (Bullet-swallowing seems to consist of seizing onto the most implausible one.) Eliminating the paradox requires showing where and how the thoughts went wrong.
0ChristianKl
Finding an idea plausible has little to do with being extremely suspicious of fancy arguments claiming to prove it. Idea that aren't proven to be impossible are plausible even when there are no convincing arguments in favor of them.
4Richard_Kennaway
So are: God, superintelligent AI, universal priors, radical life extension, and any really big idea whatever; as well as the impossibility of each of these. Plausibility is fine as a screening process for deciding where you're going to devote your efforts, but terrible as an epistemological tool.
4Rob Bensinger
If you're saying 'LessWrongers think there's a serious risk they'll be acausally blackmailed by a rogue AI', then that seems to be false. That even seems to be false in Eliezer's case, and Eliezer definitely isn't 'LessWrong'. If you're saying 'LessWrongers think acausal trade in general is possible,' then that seems true but I don't see why that's ridiculous. Is there something about acausal trade in general that you're objecting to, beyond the specific problems with Roko's argument?
anon8500

I think saying "Roko's arguments [...] weren't generally accepted by other Less Wrong users" is not giving the whole story. Yes, it is true that essentially nobody accepts Roko's arguments exactly as presented. But a lot of LW users at least thought something along these lines was plausible. Eliezer thought it was so plausible that he banned discussion of it (instead of saying "obviously, information hazards cannot exist in real life, so there is no danger discussing them").

In other words, while it is true that LWers didn't believe Roko... (read more)

6ChristianKl
If you are a programmer and think your code is safe because you see no way things could go wrong, it's still not good to believe that it isn't plausible that there's a security hole in your code. You rather practice defense in depth and plan for the possibility that things can go wrong somewhere in your code, so you add safety precautions. Even when there isn't what courts call reasonable doubt a good safety engineer still adds additional safety procautions in security critical code. Eliezer deals with FAI safety. As a result it's good for him to have mindset of really caring about safety. German nuclear power station have trainings for their desk workers to teach the desk workers to not cut themselves with paper. That alone seems strange to outsiders but everyone in Germany thinks that it's very important for nuclear power stations to foster a culture of safety even when that means something going overboard.
0Richard_Kennaway
Why would they be correct? The basilisk is plausible.
5Rob Bensinger
The wiki article talks more about this; I don't think I can give the whole story in a short, accessible way. It's true that LessWrongers endorse ideas like AI catastrophe, Hofstadter's superrationality, one-boxing in Newcomb's problem, and various ideas in the neighborhood of utilitarianism; and those ideas are weird and controversial; and some criticism of Roko's basilisk are proxies for a criticism of one of those views. But in most cases it's a proxy for a criticism like 'LW users are panicky about weird obscure ideas in decision theory' (as in Auerbach's piece), 'LWers buy into Pascal's Wager', or 'LWers use Roko's Basilisk to scare up donations/support'. So, yes, I think people's real criticisms aren't the same as their surface criticisms; but the real criticisms are at least as bad as the surface criticism, even from the perspective of someone who thinks LW users are wrong about AI, decision theory, meta-ethics, etc. For example, someone who thinks LWers are overly panicky about AI and overly fixated on decision theory should still reject Auerbach's assumption that LWers are irrationally panicky about Newcomb's Problem or acausal blackmail; the one doesn't follow from the other.
anon8500

my variation: choose the next candidate after 1/e trials that is better than 90% of existing trials. Why?: if you have a low number of candidates: worked solution - 10 candidates. you should (according to the secretary problem) interview 4 candidates, then select the next one that is better than the ones before.

Why n/e, and not some other number? Why 90%, and not some other amount? Come to think of it, shouldn't the value of the candidates matter, and not just the rank? For example, if I know my candidates' utility is sampled from [-1000,1000] and the f... (read more)

0Elo
yes; the problem of distribution is going to mess with things. If for example you knew that candidates utilities spiked at age 25, you should game the system and aim for 25 year olds. If you had prior knowledge of the candidates being all -1000 utility except for one which is +1000, then you shouldn't rely on this system at all. The premise of the problem is that n candidates can be ranked 1 to n. (which is not necessarily true for real life). The nature of the standard solution to the secretary problem is to give you the best candidate 1/e of the time; the last candidate 1/e of the time and other candidates the rest of the time. Without challenging half of the known world of my Internet friends to derive their own mathematical theory as to solving the secretary problem; starting with a neat solution and dragging an applicable one out of it is my best option. If you think that I am putting too much pseudo into the mathematics of the application; I'd encourage you to say so. If you think this is too far from applicable then I'd also encourage you to say so. (Please tell me I am wrong, I would rather be wrong than nice, and wrong than vague) There are certainly flaws in applying this known problem/solution to real life. As many other people have pointed out other prominent edge cases, (returning to partners after a few years; having children without being married). I came up with this concept when talking to a person who needed clarity on a similar issue of deciding whether he should settle down; and found it applicable enough to help him; and also while honing it down I found it applicable enough to myself to help me. Are you suggesting it shouldn't be applicable at all? Or also that it doesn't work for you? I will concede that this idea will not work for many many many people.
anon8520

By changing the strategy from "first candidate better than the ones seen in the first n/e" to anything else, you lose all the rigorous mathematical backing that made the secretary problem cool in the first place. Is your solution optimal? Near-optimal? Who knows; it depends on your utility function and the distribution of candidates, and probably involves ugly integrals with no closed-form solution.

The whole point of the secretary problem is that a very precise way of stating the problem has a cool mathematical answer (the n/e strategy). But this precise statement of the problem is almost always useless in practice, so there's very little insight gained.

0Elo
Update: yes; secretary problem has a cool and clean mathematical "next bestest candidate after 1/e trials" solution. Real life is a lot more complicated. if you work off that solution it has a 1/e chance of selecting the last candidate. which personally is atrocious odds to be playing around with. Considering the above mentioned opportunity cost ticking-time race, I need better odds than that. even if it sacrifices my chances of finding the best candidate. if you are at the 1/e*n point and you have passed the best candidate you will end up at the last candidate. If you have any suspicion that you have passed the best (or very good candidates) maybe its time to change your rule to select the best candidate excluding one known candidate. or (again) the candidate that is better than 90% of existing trials.
0Elo
I didn't mathematically model my variation. I strongly considered doing so... The reason being that the difference in my head was as clear as an addition sum. It really seems like a minor change. I am going to explain the difference again and I'd like you to try to explain what I might have overlooked in modifying the original solution. existing solution: choose the next candidate after 1/e trials that is better than all existing trials. my variation: choose the next candidate after 1/e trials that is better than 90% of existing trials. Why?: if you have a low number of candidates: worked solution - 10 candidates. you should (according to the secretary problem) interview 4 candidates, then select the next one that is better than the ones before. ---------------------------------------- I believe the sum from ((1/e)*n)->n of 1/n will yield the chance of passing the best candidate in the 1/e trials. This was the largest factor I was trying to avoid by changing the criteria to >90% of all trials. I will ask around and get back to you.
anon85120

The secretary problem is way overused, and very rarely has any application in practice. This is because it maximizes the probability of finding the best match, and NOT the expectation over the utility of the match your get. This is almost never what you want in practice; in practice, you don't care much between a match with utility 1000 and a match with utility 999, you just want to avoid a match with utility -1000.

1dankuck
Does the secretary problem say anything about getting the second-best match? That seems to be the crux and this is the counter-argument I'm most swayed by.
0Elo
This was my solution to that problem. I find the secretary problem to have a very high risk for low numbers. by the time you get to higher numbers; (>73) the chance of ending up at the last candidate is down at 0.5% or 1/1000. By using my solution your greatly lower that risk, because not only does the top candidate need to be in the first 1/e trials but the top n% of candidates need to be in the 1/e trials (in my example 10%) to cause you to be stuck with the last candidate. Of course the reasons for doing that extra weighting are more salient when concern for lost opportunity of exploration time is also factored in. Unfortunately "how much do I care about searching through the partner space for 10 years more" is a different answer for different people (read: different utilities of care/pain/effort factor) Does this make sense? (/answer your comment?)
anon8570

You might be interested in Aaronson's proposed theory for why it might be physically impossible to copy a human brain. He outlined it in "The Ghost in the Quantum Turing Machine": http://arxiv.org/abs/1306.0159

In that essay he discusses a falsifiable theory of the brain that, if true, would mean brain states are un-copyable. So Yudkowsky's counter-argument may be a little too strong: it is indeed consistent with modern physics for brain simulation to be impossible.

anon85100

That might be Eliezer's stated objection. I highly doubt it's his real one (which seems to be something like "not releasing the logs makes me seem like a mysterious magician, which is awesome"). After all, if the goal was to make the AI-box escape seem plausible to someone like me, then releasing the logs - as in this post - helps much more than saying "nya nya, I won't tell you".

2entirelyuseless
Yes, it's not implausible that this motive is involved as well.
anon8500

Okay, let's suppose for a second that I buy that teaching students to be goal oriented helps them significantly. That still leaves quite a few questions:

  1. Many school boards already try to teach students to be goal oriented. Certainly "list out realistic goals" was said to me countless times in my own schooling. What do you plan to do differently?

  2. There seems to be no evidence at all that LW material is better for life outcomes than any other self-help program, and some evidence that it's worse. Consider this post (again by Scott): http://lesswrong.com/lw/9p/extreme_rationality_its_not_that_great/

0Gleb_Tsipursky
I plan to teach students actually how to be goal oriented. It's the difference between telling people "lose weight" and specifically giving them clear instructions for how to do is. Here is an example of how I do so in a videotaped workshop. I would like to have an experimental attitude to LW content, and will look forward to see the results of my experiments. I don't intend to do the extreme rationality stuff, and expect more of it than it can deliver. We'll see, I guess :-)
anon8500

There's also quite a bit of reason to be skeptical of that evidence. Here's slatestarcodex's take: http://slatestarcodex.com/2015/03/11/too-good-to-be-true/

2Gleb_Tsipursky
Yup, I'm aware of Scott's dislike of the growth mindset hypothesis, he's a bit on the extreme spectrum on that one. However, even in the post itself, he notes that there are several studies that show the benefits of teaching students to be goal oriented. There's lots of research out there that teaching students metacognition is helpful, for example this chapter cites a lot of studies. I'd say that overall the probabilistic evidence supports the hypothesis that teaching people to be goal oriented and self-reflective about their ways of achieving their goals will help them have better results in achieving those goals.
anon8500

Do you have any evidence that LW materials help people refine and achieve their goals?

Helping people refine and achieve their goals is pretty damn difficult: school boards, psychiatrists, and welfare programs have been trying to do this for decades. For example, are you saying that teaching LW material in schools will improve student outcomes? I would bet very strongly against such a prediction.

0Gleb_Tsipursky
There's actually quite a bit of evidence on how helping students refine and achieve their goals helps them learn better, for example here.
anon8510

Modern SGD mechanisms are powerful global optimizers.

They are heuristic optimizers that have no guarantees of finding a global optimum. It's strange to call them "powerful global optimizers".

Solomonoff induction is completely worthless - intractable - so you absolutely don't want to do that anyway.

I believe that was my point.

1jacob_cannell
Not at all strange. As I pointed out in the post above - modern SGD functions as a global optimizer due to clever tricks such as over-parameterization, dropout, and momentum. These optimizers are also suprisingly powerful as measured in terms of actual solution quality vs computational cost - indeed they dominate most of the landscape. It is completely disingenous to call the dominate solution anything other than powerful. In practice strong guarantees are not important. What actually matters is finding a solution that is good enough relative to computational costs, where 'good enough' is determined according to some more complex business/economic utility metric, where solution quality follows diminishing returns. For sufficiently complex problems, finding the actual global optimum is rarely ideal - it is far too far to the right on the asymptotic diminishing returns curve.
anon8510

My goal is convincing people to have more clear and rational, evidence-thinking, as informed by LW materials.

Is there an objective measure by which LW materials inform more "clear and rational" thought? Can you define "clear and rational"? Or actually, to use LW terminology, can you taboo "clear" and "rational" and restate your point?

Regardless, as Brian Tomasik points out, helping people be more rational contributes to improving the world, and thus the ultimate goal of the EA movement.

But does it contribute to improving the world in an effective way?

0Gleb_Tsipursky
Well, I'd say that "clear and rational" is the same as "arriving at the correct answer to make the best decision to refine and achieve goals." So yes, I would say it does contribute to improving the world in an effective way, because helping people both understand their goals better (refine goals) and then achieve their goals helps people have better lives and thus improves flourishing.
anon85-10

If by "spreading rationality" you mean spreading LW material and ideas, then a potential problem is that it causes many people to donate their money to AI friendliness research instead of to malaria nets. Although these people consider this to be "effective altruism", as an AI skeptic it's not clear to me that this is significantly more effective than, say, donating money to cancer research (as non-EA people often do).

0Gleb_Tsipursky
My goal is convincing people to have more clear and rational, evidence-thinking, as informed by LW materials. Some people may choose to donate to AI, and others to EA - as you can see from the blog I cited, I specifically highlight the benefits of the EA movement. Regardless, as Brian Tomasik points out, helping people be more rational contributes to improving the world, and thus the ultimate goal of the EA movement.
anon8500

You cannot cherry pick a single year (a pretty non-representative year given the recession) in which the growth of a few sub-Saharan African countries was faster than the average growth of the stock market.

I didn't cherry-pick anything; that was the first google image result, so it's the one I linked to. I didn't think it's any different from a typical year. Is it? If so, what was special that year? If you're concerned that the US was in a recession, you can simply compare sub-Saharan Africa to the typical 6-7% stock market returns instead of comparing... (read more)

anon8510

The rate of return on the stock market is around 10%

You didn't adjust for inflation; it's actually around 6 or 7%.

This is much faster than the rate of growth of sub-Saharan economies.

Depends on the country:

http://en.wikipedia.org/wiki/Gross_domestic_product#/media/File:Gdp_real_growth_rate_2007_CIA_Factbook.PNG

Actually foreign aid might have a negative rate of return since most of the transfers are consumed rather than reinvested. Which isn't a problem per say - eventually you have to convert capital into QALYs even if that means you stop growing

... (read more)
2pianoforte611
You cannot cherry pick a single year (a pretty non-representative year given the recession) in which the growth of a few sub-Saharan African countries was faster than the average growth of the stock market. to refute the claim that the stock market grows faster than sub-Saharan economies. A more complete data set shows that indeed the sub-Saharan economy has grown much slower than the stock market. This shouldn't be a controversial point. So what you are arguing is that the most efficient use of money to gain QALYs (not the average) has decreased exponentially and faster than the growth of capital over time? That seems very difficult to argue while taking into account increased knowledge and technology. But I have no idea how to calculate that.
anon8510

The first is: more wasteful economically. This seems pretty robust, investments in sub-Saharan Africa have historically generated much less wealth than investments in other countries. Moreover wealth continues to grow via reinvestment.

It's not clear what you mean by this. Do you mean investments in Africa have generated less wealth for the investor? That might be true, but it doesn't mean they have generated less wealth overall. How would you measure this?

he second is: more wasteful ethically. This is harder to defend, but I think it is a reasonable c

... (read more)
2pianoforte611
The rate of return on the stock market is around 10%. This is much faster than the rate of growth of sub-Saharan economies. Actually foreign aid might have a negative rate of return since most of the transfers are consumed rather than reinvested. Which isn't a problem per say - eventually you have to convert capital into QALYs even if that means you stop growing it (if you are an effective altruist). The question is how much, and when? I didn't actually come up with the argument that investing now and donating later is more efficient. Robin Hanson did, and there has been some back and forth there which I highly recommend (so as not to retread over old arguments). Even if QALYs per dollar decrease exponentially and faster than the growth of capital (which you've asserted without argument - I simply think that no one knows), there is still the issue of whether investment followed by donation (to high marginal QALY causes), is more effective than direct donation. Its a very difficult optimization problem and while I don't know the answer to it, I'm disappointed by how overconfident people are that they know the answer.
anon8560

Yeah, you're probably right. I was probably just biased because the timeline is my main source of disagreement with AI danger folks.

anon8540

I think point 1 is very misleading, because while most people agree with it, hypothetically a person might assign 99% chance of humanity blowing itself up before strong AI, and < 1% chance of strong AI before the year 3000. Surely even Scott Alexander will agree that this person may not want to worry about AI right now (unless we get into Pascal's mugging arguments).

I think most of the strong AI debate comes from people believing in different timelines for it. People who think strong AI is not a problem think we are very far from it (at least conceptually, but probably also in terms of time). People who worry about AI are usually pretty confident that strong AI will happen this century.

7Houshalter
In my experience the timeline is not usually the source of disagreement. They usually don't believe that AI would want to hurt humans. That the paperclip maximizer scenario isn't likely/possible. E.g. this popular reddit thread from yesterday. I guess that would be premise number 3 or 4, that goal alignment is a problem that needs to be solved.
1IlyaShpitser
My reading of that article is: "I am stumping for my friends."
anon8500

I suppose the Bayesian answer to that is that a probability distribution is a description of one's knowledge, and that in principle, every state of knowledge, including total ignorance, can be represented as a prior distribution. In practice, one may not know how to do that. Fundamentalist Bayesians say that that is a weakness in our knowledge, while everyone else, from weak Bayesians to Sunday Bayesians, crypto-frequentists, and ardent frequentists, say it's a weakness of Bayesian reasoning. Not being a statistician, I don't need to take a view, although

... (read more)
Load More