You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

perpetualpeace1 comments on A question about Eliezer - Less Wrong Discussion

33 Post author: perpetualpeace1 19 April 2012 05:27PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (158)

You are viewing a single comment's thread. Show more comments above.

Comment author: perpetualpeace1 19 April 2012 06:39:15PM 3 points [-]

If his decision theory had a solid theoretical background, but turned out to be terrible when actually implemented, how would we know? Has there been any empirical testing of his theory?

Comment author: orthonormal 19 April 2012 11:14:58PM 4 points [-]

Permit me to substitute your question: TDT seems pretty neat philosophically, but can it actually be made to work as computer code?

Answer: Yes. (Sorry for the self-promotion, but I'm proud of myself for writing this up.) The only limiting factor right now is that nobody can program an efficient theorem-prover (or other equivalently powerful general reasoner), but that's not an issue with decision theory per se. (In other words, if we could implement Causal Decision Theory well, then we could implement Timeless Decision Theory well.) But in any case, we can prove theorems about how TDT would do if equipped with a good theorem-prover.

Comment author: TimS 19 April 2012 06:52:49PM *  4 points [-]

What does a decision theory that "has a solid theoretical background but turns out to be terrible" look like when implemented?

Comment author: perpetualpeace1 19 April 2012 07:02:25PM 5 points [-]

You play a game that you could either win or lose. One person follows, so far as he or she is able, the tenets of timeless decision theory. Another person makes a decision by flipping a coin. The coin-flipper outperforms the TDTer.

Comment author: Larks 19 April 2012 07:29:55PM 3 points [-]

I'm pretty confident that if we played an iterated prisoners dilemma, your flipping a coin each time and my running TDT, I would win. This is, however, quite a low bar.

Comment author: David_Gerard 19 April 2012 08:03:52PM *  5 points [-]

I took it as a statement of what would prove the theory false, rather than a statement of something believed likely.

Comment author: JoshuaZ 19 April 2012 08:47:01PM 0 points [-]

I think that would only be true if there were more than 2 players. Won't random coin flip and TDT be tied in the two player case?

Comment author: Luke_A_Somers 20 April 2012 07:48:22AM *  4 points [-]

No. TDT, once it figures out it's facing a coin flipper, defects 100% of the time and runs away with it.

Comment author: Larks 20 April 2012 06:12:22AM 3 points [-]

Nope, TDT will defect every time.

Comment author: [deleted] 19 April 2012 08:22:43PM 2 points [-]

I'm not sure such a poor theory would survive having a solid theoretical background.

Comment author: shokwave 20 April 2012 03:03:28AM 0 points [-]

Actually, I think the proper case is "Two players play a one-player game they can either win or lose. One person follows, so far as able, the tenets of TDT. The other decides by flipping a coin. The coin-flipper outperforms the TDTer."

I mention this because lots of decision theories struggle against a coin flipping opponent: tit-for-tat is a strong IPD strategy that does poorly against a coin-flipper.

Comment author: Random832 20 April 2012 01:41:34PM 1 point [-]

Is there any decision strategy that can do well (let's define "well" as "better than always-defect") against a coin-flipper in IPD? Any decision strategy more complex than always-defect requires the assumption that your opponent's decisions can be at least predicted, if not influenced.

Comment author: dlthomas 21 April 2012 12:29:04AM 1 point [-]

No, of course not. Against any opponent whose output has nothing to do with your previous plays (or expected plays, if they get a peak at your logic), one should clearly always defect.

Comment author: Random832 21 April 2012 05:15:01AM *  0 points [-]

Not if their probability of cooperation is so high that the expected value of cooperation remains higher than that of defecting. Or if their plays can be predicted, which satisfies your criterion (nothing to do with my previous plays) but not mine.

If someone defects every third time with no deviation, then I should defect whenever they defect. If they defect randomly one time in sixteen, I should always cooperate. (of course, always-cooperate is not more complex than always-defect.)

...I swear, this made sense when I did the numbers earlier today.