I don't understand why the Smoking Lesion is a problem for evidential decision theory. I would simply accept that in the scenario given, you shouldn't smoke. And I don't see why you assert that this doesn't lessen your chances of getting cancer, except in the same sense that two-boxing doesn't lessen your chances of getting the million.
I would just say: in the scenario give, you should not smoke, and this will improve your chances of not getting cancer.
If you doubt this, consider if the correlation were known to be 100%; every person who ever smoked up til...
I see you've moved this discussion off-site. FWIW, I commend you for trying to organize the various decision theory issues into a more accessible and organized sequence. I'd like to suggest that you take some of this and use it to improve the (almost comically sparse) decision theory articles on the LW Wiki. If that's really going to be the go-to place for LW knowledge, your efforts to summarize and present this info could really be useful there, and any redundancy with existing blog posts would be a non-issue.
I'm confused as to why you said you weren't continuing this on Less Wrong, then posted it on Less Wrong.
I've read the smoking lesion thing before, and what occurred to be is that even under EDT, the reasoning in there is wrong. What I mean was that one shouldn't simply reason it out by comparing to the average stats, but take into account the fact that they're using EDT itself. ie, they should say "given that a person is using EDT, then what's the correlation between etc etc..."
Worth referencing:
The Smoking Lesion on the wiki.
Timeless Decision Theory and Meta-Circular Decision Theory, where Eliezer discusses this problem (among others)
(By the way, your blog has some interesting posts!)
I'm sorry for being stupid.
Still, your original description of the scenario was
you have a light attached to your head. If it blinks red, it'll make you feel happy, but it'll blow up in an hour. It's not linked to the rest of your brain at all.
Now you have changed the "happy" feeling into a "decided" feeling. So the bulb has to be connected somehow to the brain to stimulate the feeling. I am not sure what "rest" refers to here.
But in general, if somebody said they decided freely, I take it as given. I don't know any better criterion how to judge whether the decision was free, whatever it means.
It's my mistake.
I meant: it's not connected to your brain at all except when making you happy/making you believe you decided.
ie. it's not taking any input from the brain at any point. Much like the lesion.
But in general, if somebody said they decided freely, I take it as given. I don't know any better criterion how to judge whether the decision was free, whatever it means.
In the specific case of the bulb-world, would you consider their decisions free, if they did?
This is part of a sequence titled "An introduction to decision theory". The previous post was Newcomb's Problem: A problem for Causal Decision Theories
For various reasons I've decided to finish this sequence on a seperate blog. This is principally because there were a large number of people who seemed to feel that this sequence either wasn't up to the Less Wrong standard or felt that it was simply covering ground that had already been covered on Less Wrong.
The decision to post it on another blog rather than simply discontinuing it came down to the fact that other people seemed to feel that the sequence had value. Those people can continue reading it at "The Smoking Lesion: A problem for evidential decision theory".
Alternatively, there is a sequence index available: Less Wrong and decision theory: sequence index