waitingforgodel comments on Best career models for doing research? - Less Wrong

27 Post author: Kaj_Sotala 07 December 2010 04:25PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (999)

You are viewing a single comment's thread. Show more comments above.

Comment author: waitingforgodel 08 December 2010 01:13:07PM 2 points [-]

I understand your argument re: very weak belief... but it seems silly.

How is this different than positing a very small chance that a future dictator will nuke the planet unless I mail a $10 donation to green peace?

Comment author: Desrtopa 08 December 2010 01:17:12PM 3 points [-]

Do you have any reason to believe that it's more likely that a future dictator, or anyone else, will nuke the planet if you don't send a donation to Greenpeace than if you do?

Comment author: Vladimir_Nesov 08 December 2010 01:31:49PM 2 points [-]

How is this different than positing a very small chance that a future dictator will nuke the planet unless I mail a $10 donation to green peace?

I agree that you are not justified in seeing a difference, unless you understand the theory of acausal control to some extent and agree with it. But when you are considering a person who agrees with that theory, and makes a decision based on it, agreement with the theory fully explains that decision, this is a much better explanation than most of the stuff people are circulating here. At that point, disagreement about the decision must be resolved by arguing about the theory, but that's not easy.

Comment author: David_Gerard 08 December 2010 03:06:09PM *  4 points [-]

But when you are considering a person who agrees with that theory, and makes a decision based on it, agreement with the theory fully explains that decision, this is a much better explanation than most of the stuff people are circulating here.

You appear to be arguing that a bad decision is somehow a less bad decision if the reasoning used to get to it was consistent ("carefully, correctly wrong").

At that point, disagreement about the decision must be resolved by arguing about the theory, but that's not easy.

No, because the decision is tested against reality. Being internally consistent may be a reason for doing something that it is obvious to others is just going to be counterproductive - as in the present case - but it doesn't grant a forgiveness pass from reality.

That is: in practical effects, sincere stupidity and insincere stupidity are both stupidity.

You even say this above ("There is only one proper criterion to anyone's actions, goodness of consequences"), making your post here even stranger.

(In fact, sincere stupidity can be more damaging, as in my experience it's much harder to get the person to change their behaviour or the reasoning that led to it - they tend to cling to it and justify it when the bad effects are pointed out to them, with more justifications in response to more detail on the consequences of the error.)

Think of it as a trolley problem. Leaving the post is a bad option, the consequences of removing it are then the question: which is actually worse and results in the idea propagating further? If you can prove in detail that a decision theory considers removing it will make it propagate less, you've just found where the decision theory fails.

Removing the forbidden post propagated it further, and made both the post itself and the circumstances of its removal objects of fascination. It has also diminished the perceived integrity of LessWrong, as we can no longer be sure posts are not being quietly removed as well as loudly; this also diminished the reputation of SIAI. It is difficult to see either of these as working to suppress the bad idea.

Comment author: wedrifid 08 December 2010 04:42:32PM 6 points [-]

It has also diminished the perceived integrity of LessWrong, as we can no longer be sure posts are not being quietly removed as well as loudly; this also diminished the reputation of SIAI.

More importantly it removed lesswrong as a place where FAI and decision theory can be discussed in any depth beyond superficial advocacy.

Comment author: David_Gerard 08 December 2010 07:55:05PM 3 points [-]

More importantly it removed lesswrong as a place where FAI and decision theory can be discussed in any depth beyond superficial advocacy.

The problem is more than the notion that secret knowledge is bad - it's that secret knowledge increasingly isn't possible, and increasingly isn't knowledge.

If it's science, you almost can't do it on your own and you almost can't do it as a secret. If it's engineering, your DRM or other constraints will last precisely as long as no-one is interested in breaking them. If it's politics, your conspiracy will last as long as you aren't found out and can insulate yourself from the effects ... that one works a bit better, actually.

Comment author: Vladimir_Nesov 08 December 2010 07:58:41PM 1 point [-]

More importantly it removed lesswrong as a place where FAI and decision theory can be discussed in any depth beyond superficial advocacy.

I don't believe this is true to any significant extent. Why do you believe that? What kind of questions are not actually discussed that could've been discussed otherwise?

Comment author: wedrifid 08 December 2010 09:06:55PM 6 points [-]

You are serious?

  • What qualifies as a 'Friendly' AI?
  • If someone is about to execute an AI running 'CEV' should I push a fat man on top of him and save five people from torture? What about an acausal fat man? :)
  • (How) can acausal trade be used to solve the cooperation problem inherent in funding FAI development? If I recall this topic was one that was explicitly deleted. Torture was mostly just a superficial detail.

... just from a few seconds brainstorming. These are the kinds of questions that can not be discussed without, at the very least, significant bias due to the threat of personal abuse and censorship if you are not careful. I am extremely wary of even trivial inconveniences.

Comment author: Vladimir_Nesov 08 December 2010 09:22:25PM *  0 points [-]

You are serious?

Yes.

  • What qualifies as a 'Friendly' AI?

This doesn't seem like an interesting question, where it intersects the forbidden topic. We don't understand decision theory well enough to begin usefully discussing this. Most directions of discussion about this useless question are not in fact forbidden and the discussion goes on.

(How) can acausal trade be used to solve the cooperation problem inherent in funding FAI development?

We don't formally understand even the usual game theory, let alone acausal trade. It's far too early to discuss its applications.

Comment author: wedrifid 08 December 2010 10:04:07PM 4 points [-]

This doesn't seem like an interesting question

It wasn't Vladimir_Nesov's interest that you feigned curiosity in and nor is it your place to decide what things others are interested in discussing. They are topics that are at least as relevant as such things as 'Sleeping Beauty' that people have merrily prattled on about for decades.

That you support a censorship of certain ideas by no means requires you to exhaustively challenge every possible downside to said censorship. Even if the decision were wise and necessary there is allowed to be disappointing consequences. That's just how things are sometimes.

The zeal here is troubling.

Comment author: Vladimir_Nesov 08 December 2010 10:22:48PM *  2 points [-]

It wasn't Vladimir_Nesov's interest that you feigned curiosity in and nor is it your place to decide what things others are interested in discussing.

What do you mean by "decide"? Whether they are interested in that isn't influenced by my decisions, and I can well think about whether they are, or whether they should be (i.e. whether there is any good to be derived from that interest).

I opened this thread by asking,

What kind of questions are not actually discussed that could've been discussed otherwise?

You answered this question, and then I said what I think about that kind of questions. It wasn't obvious to me that you didn't think of some other kind of questions that I find important, so I asked first, not just rhetorically.

What you implied in this comment seems very serious, and it was not my impression that something serious was taking place as a result of the banning incident, so of course I asked. My evaluation of whether the topics excluded (that you've named) are important is directly relevant to the reason your comment drew my attention.

Comment author: Vladimir_Nesov 09 December 2010 04:29:36PM -1 points [-]

On downvoting of parent comment: I'm actually surprised this comment got downvoted. It's not as long inferential depth as this one that got downvoted worse, and it looks to me quite correct. Help me improve, say what's wrong.

Comment author: Vladimir_Nesov 08 December 2010 10:27:40PM *  1 point [-]

That you support a censorship of certain ideas by no means requires you to exhaustively challenge every possible downside to said censorship.

The other way around. I don't "support censorship", instead I don't see that there are downsides worth mentioning (besides the PR hit), and as a result I disagree that censorship is important. Of course this indicates that I generally disagree with arguments for the harm of the censorship (that I so far understood), and so I argue with them (just as with any other arguments I disagree with that are on topic I'm interested in).

The zeal here is troubling.

No zeal, just expressing my state of belief, and not willing to yield for reasons other than agreement (which is true in general, the censorship topic or not).

Comment author: wedrifid 09 December 2010 12:37:21AM 4 points [-]

No zeal, just expressing my state of belief, and not willing to yield for reasons other than agreement (which is true in general, the censorship topic or not).

No, yielding and the lack thereof is not the indicator of zeal of which I speak. It is the sending out of your soldiers so universally that they reach even into the territory of other's preferences. That critical line between advocation of policy and the presumption that others must justify their very thoughts (what topics interests them and how their thoughts are affected by the threat of public shaming and censorship) is crossed.

The lack of boundaries is a telling sign according to my model of social dynamics.

Comment author: Vladimir_Nesov 08 December 2010 10:38:14PM 0 points [-]

(Should I lump everything in one comment, or is the present way better? I find it more clear if different concerns are extracted as separate sub-threads.)

Comment author: steven0461 08 December 2010 10:42:00PM 6 points [-]

It's not just more clear, it allows for better credit assignment in cases where both good and bad points are made.

Comment author: wedrifid 09 December 2010 12:16:06AM 0 points [-]

Steven beat me to it - this way works well. Bear in mind though that I wasn't planning to engage in this subject too deeply. Simply because it furthers no goal that I am committed to and is interesting only in as much as it can spawn loosely related tangents.

Comment author: Vladimir_Nesov 08 December 2010 10:36:32PM 0 points [-]

That some topics are excluded is tautological, so it's important what kind of topics were. Thus, stating "nor is it your place to decide what things others are interested in discussing" seems to be equivalent with stating "censorship (of any kind) is bad!", which is not very helpful in the discussion of whether it's in fact bad. What's the difference you intended?

Comment author: wedrifid 08 December 2010 09:40:05PM 4 points [-]

We don't formally understand even the usual game theory, let alone acausal trade. It's far too early to discuss its applications.

You do see the irony there I hope...

Comment author: XiXiDu 09 December 2010 10:08:56AM 3 points [-]

Would you have censored the information? If not, do you think it would be a good idea to discuss the subject matter on an external (public) forum? Would you be interested to discuss it?

Comment author: wedrifid 09 December 2010 10:42:15AM 6 points [-]

Would you have censored the information?

No, for several reasons. I have made no secret of the fact that I don't think Eliezer processes perceived risks rationally and I think this applies in this instance.

This is not a claim that censorship is always a bad idea - there are other obvious cases where it would be vital. Information is power, after all.

If not, do you think it would be a good idea to discuss the subject matter on an external (public) forum?

Only if there is something interesting to say on the subject. Or any interesting conversations to be had on the various related subjects that the political bias would interfere with. But the mere fact that Eliezer forbids it doesn't make it more interesting to me. In fact, the parts of Roko's posts that were most interesting to me were not even the same parts that Eliezer threw a tantrum over. As far as I know Roko has been bullied out of engaging in such conversation even elsewhere and he would have been the person most worth talking to about that kind of counterfactual.

Bear in mind that the topic has moved from the realm of abstract philosophy to politics. If you make any mistakes, demonstrate any ignorance or even say things that can be conceivably twisted to appear as such then expect that will be used against you here to undermine your credibility on the subject. People like Nesov and and jimrandom care, and care aggressively.

Would you be interested to discuss it?

Post away, if I have something to add then I'll jump in. But warily.

Comment author: Vladimir_Nesov 08 December 2010 09:42:11PM 0 points [-]

No irony. You don't construct complex machinery out of very weak beliefs, but caution requires taking very weak beliefs into account.

Comment author: wedrifid 08 December 2010 09:49:33PM 1 point [-]

The irony is present and complex machinery is a red herring.

Comment author: Vladimir_Nesov 08 December 2010 03:33:58PM 0 points [-]

But when you are considering a person who agrees with that theory, and makes a decision based on it, agreement with the theory fully explains that decision, this is a much better explanation than most of the stuff people are circulating here.

You appear to be arguing that a bad decision is somehow a less bad decision if the reasoning used to get to it was consistent ("carefully, correctly wrong").

Here, I'm talking about factual explanation, not normative estimation. The actions are explained by holding a certain belief, better than by alternative hypotheses. Whether they were correct is a separate question.

At that point, disagreement about the decision must be resolved by arguing about the theory, but that's not easy.

No, because the decision is tested against reality.

You'd need to explain this step in more detail. I was discussing a communication protocol, where does "testing against reality" enter that topic?

Comment author: David_Gerard 08 December 2010 03:38:51PM 1 point [-]

Ah, I thought you were talking about whether the decision solved the problem, not whether the failed decision was justifiable in terms of the theory.

I do think that if a decision theory leads to quite as spectacular a failure in practice as this one did, then the decision theory is strongly suspect.

As such, whether the decision was justifiable is less interesting except in terms of revealing the thinking processes of the person doing the justification (clinginess to pet decision theory, etc).

Comment author: Vladimir_Nesov 08 December 2010 03:47:32PM 1 point [-]

I do think that if a decision theory leads to quite as spectacular a failure in practice as this one did, then the decision theory is strongly suspect.

"Belief in the decision being a failure is an argument against adequacy of the decision theory", is simply a dual restatement of "Belief in the adequacy of the decision theory is an argument for the decision being correct".

Comment author: David_Gerard 08 December 2010 04:02:00PM *  0 points [-]

"Belief in the decision being a failure is an argument against adequacy of the decision theory", is simply a dual restatement of "Belief in the adequacy of the decision theory is an argument for the decision being correct".

This statement appears confusing to me: you appear to be saying that if I believe strongly enough in the forbidden post having been successfully suppressed, then censoring it will not have in fact caused it to propagated widely, nor will it have become an object of fascination and caused a reputational hit to LessWrong and hence SIAI. This, of course, makes no sense.

I do not understand how this matches with the effects observable in reality, where these things do in fact appear to have happened. Could you please explain how one tests this result of the decision theory, if not by matching it against what actually happened? That being what I'm using to decide whether the decision worked or not.

Keep in mind that I'm talking about an actual decision and its actual results here. That's the important bit.

Comment deleted 08 December 2010 04:44:33PM *  [-]
Comment author: David_Gerard 08 December 2010 07:03:48PM *  2 points [-]

No, I think you're nitpicking to dodge the question, and looking for a more convenient world.

I think at this point it's clear that you really can't be expected to give a straight answer. Well done, you win!

Comment author: Vladimir_Nesov 08 December 2010 07:09:11PM *  -2 points [-]

No, I think you're nitpicking to dodge the question, and looking for a more convenient world.

Have you tried?

I read your comment, understood the error you made, and it was about not seeing the picture clearly enough. If you describe the situation in terms of the components I listed, I expect you'll see what went wrong. If you don't oblige, I'll probably describe the solution tomorrow.

Edit in response to severe downvoting: Seriously? It's not allowed to entertain exercises about a conversational situation? (Besides, I was merely explaining an exercise given in another comment.) Believe, argument can be a puzzle to understand, and not a fight. If clumsy attempts to understand are discouraged, how am I supposed to develop my mastery?

Comment author: shokwave 08 December 2010 04:45:56PM 0 points [-]

If you believe that "decision is a failure" is evidence that the decision theory is not adequate, you believe that "decision is a success" is evidence that the decision theory is adequate.

Since a decision theory's adequacy is determined by how successful its decisions are, you appear to be saying "if a decision theory makes a bad decision, it is a bad decision theory" which is tautologically true.

Correct me if I'm wrong, but Vladimir_Nesov is not interested in whether the the decision theory is good or bad, so restating an axiom of decision theory evaluation is irrelevant.

The decision was made by a certain decision theory. The factual question "was the decision-maker holding to this decision theory in making this decision?" is entirely unrelated to the question "should the decision-maker hold to this decision theory given that it makes bad decisions?". To suggest otherwise blurs the prescriptive/descriptive divide, which is what Vladimir_Nesov is referring to when he says

Here, I'm talking about factual explanation, not normative estimation.

Comment author: David_Gerard 08 December 2010 07:08:58PM *  1 point [-]

If you believe that "decision is a failure" is evidence that the decision theory is not adequate, you believe that "decision is a success" is evidence that the decision theory is adequate.

I believe that if the decision theory clearly led to an incorrect result (which it clearly did in this case, despite Vladimir Nesov's energetic equivocation), then it is important to examine the limits of the decision theory.

As I understand it, the purpose of bothering to advocate TDT is that it beats CDT in the hypothetical case of dealing with Omega (who does not exist), and is therefore more robust, then this failure in a non-hypothetical situation suggests a flaw in its robustness, and it should be regarded as less reliable than it may have been regarded previously.

Assuming the decision was made by robust TDT.

Comment author: wedrifid 08 December 2010 08:08:01PM 6 points [-]

As I understand it, the purpose of bothering to advocate TDT is that it beats CDT in the hypothetical case of dealing with Omega (who does not exist), and is therefore more robust, then this failure in a non-hypothetical situation suggests a flaw in its robustness, and it should be regarded as less reliable than it may have been regarded previously.

The decision you refer to here... I'm assuming it is this still the Eliezer->Roko decision? (This discussion is not the most clearly presented.) If so for your purposes you can safely consider 'TDT/CDT' irrelevant. While acausal (TDTish) reasoning is at play in establishing a couple of the important premises, they are not relevant to the reasoning that you actually seem to be criticising.

ie. The problems you refer to here are not the fault of TDT or of abstract reasoning at all - just plain old human screw ups with hasty reactions.

Comment author: Vladimir_Nesov 08 December 2010 08:29:47PM *  2 points [-]

As I understand it, the purpose of bothering to advocate TDT is that it beats CDT in the hypothetical case of dealing with Omega (who does not exist), and is therefore more robust

See section 7 of the TDT paper (you'll probably have to read from the beginning to familiarize yourself with concepts). It doesn't take Omega to demonstrate that CDT errs, it takes mere ability to predict dispositions of agents to any small extent to get out of CDT's domain, and humans do that all the time. From the paper:

The argument under consideration is that I should adopt a decision theory in which my decision takes general account of dilemmas whose mechanism is influenced by "the sort of decision I make, being the person that I am" and not just the direct causal effects of my action. It should be clear that any dispositional influence on the dilemma's mechanism is sufficient to carry the force of this argument. There is no minimum influence, no threshold value.

Comment author: jimrandomh 08 December 2010 07:24:17PM 1 point [-]

I wouldn't use this situation as evidence for any outside conclusions. Right or wrong, the belief that it's right to suppress discussion of the topic entails also believing that it's wrong to participate in that discussion or to introduce certain kinds of evidence. So while you may believe that it was wrong to censor, you should also expect a high probability of unknown unknowns that would mess up your reasoning if you tried to take inferential steps from that conclusion to somewhere else.

Comment author: Vladimir_Nesov 08 December 2010 07:12:01PM *  1 point [-]

failure in a non-hypothetical situation suggests its robustness is in doubt and it should be regarded as less reliable than it may have been regarded previously.

I agree, and in this comment I remarked that we were assuming this statement all along, albeit in a dual presentation.

Comment author: waitingforgodel 08 December 2010 01:35:12PM 3 points [-]

If you're interested, we can also move forward as I did over here by simply assuming EY is right, and then seeing if banning the post was net positive

Comment author: Vladimir_Nesov 08 December 2010 01:55:36PM *  2 points [-]

If you're interested, we can also move forward as I did over here

It's not "moving forward", it's moving to a separate question. That question might be worth considering, but isn't generally related to the original one.

simply assuming EY is right,

Why would the assumption that EY was right be necessary to consider that question?

and then seeing if banning the post was net positive

I agree that it was net negative, specifically because the idea is still circulating, probably with more attention drawn to it than would happen otherwise. Which is why I started commenting on my hypothesis about the reasons for EY's actions, in an attempt to alleviate the damage, after I myself figured it out. But that it was in fact net negative doesn't directly argue that given the information at hand when the decision was made, it had net negative expectation, and so that the decision was incorrect (which is why it's a separate question, not a step forward on the original one).

Comment author: wedrifid 08 December 2010 02:02:02PM 6 points [-]

But that it was in fact net negative doesn't directly argue that given the information at hand when the decision was made, it had net negative expectation, and so that the decision was incorrect.

More than enough information about human behavior was available at the time. Negative consequences of the kind observed were not remotely hard to predict.

Comment author: Vladimir_Nesov 08 December 2010 02:04:46PM 1 point [-]

Yes, quite likely. I didn't argue with this point, though I myself don't understand human behavior enough for that expectation to be obvious. I only argued that the actual outcome isn't a strong reason to conclude that it was expected.

Comment author: waitingforgodel 08 December 2010 02:27:26PM 5 points [-]

It's not "moving forward", it's moving to a separate question.

I like the precision of your thought.

All this time I thought we were discussing if blocking future censorship by EY was a rational thing to do -- but it's not what we were discussing at all.

You really are in it for the details -- if we could find a way of estimating around hard problems to solve the above question, that's only vaguely interesting to you -- you want to know the answers to these questions.

At least that's what I'm hearing.

It sounds like the above was your way of saying you're in favor of blocking future EY censorship, which gratifies me.

I'm going to do the following things in the hope of gratifying you:

  1. Writing up a post on less wrong for developing political muscles. I've noticed several other posters seem less than savvy about social dynamics, so perhaps a crash course is in order. (I know that there are certainly several in the archives, I guarantee I'll bring several new insights [with references] to the table).

  2. Reread all your comments, and come back at these issues tomorrow night with a more exact approach. Please accept my apology for what I assume seemed a bizarre discussion, and thanks for thinking like that.

Night!

Comment author: Vladimir_Nesov 08 December 2010 03:05:30PM 0 points [-]

It sounds like the above was your way of saying you're in favor of blocking future EY censorship, which gratifies me.

I didn't address that question at all, and in fact I'm not in favor of blocking anything. I came closest to that topic in this comment.