Well, if you know advance that Omega is more likely to do this than it is to impose a dilemma where it will fill both boxes only if you two-box, then I'd agree that this is an appropriate solution.
I think that if in advance you have a flat probability distribution for what sort of Omega scenarios might occur (Omega is just as likely to fill both boxes only if you would two-box in the first scenario as it is to fill both boxes only if you would one-box,) then this solution doesn't make sense.
In the transparent Newcomb's problem, when both boxes are filled, does it benefit you to be the the sort of person who would one-box? No, because you get less money that way. If Omega is more likely to impose the transparent Newcomb's problem than its inversion, then prior to Omega foisting the problem on you, it does benefit you to be the sort of person who would one-box (and you can't change what sort of person you are mid-problem.)
If Omega only presents transparent Newcomb's problems of the first sort, where the box containing more money is filled only if the person presented with the boxes would one-box, then situations where a person is presented with two transparent boxes of money and picks both will never arise. People who would one-box in the transparent Newcomb's problem come out ahead.
If Omega is equally likely to present transparent Newcomb's problems of the first sort, or inversions where Omega fills both boxes only for people it predicts would two-box in problems of the first sort, then two-boxers come out ahead, because they're equally likely to get the contents of the box with more money, but always get the box with less money, while the one-boxers never do.
You can always contrive scenarios to reward or punish any particular decision theory. The Transparent Newcomb's Problem rewards agents which one-box in the Transparent Newcomb's Problem over agents which two-box, but unless this sort of problem is more likely to arise than ones which reward agents which two-box in Transparent Newcomb's Problem over ones that one-box, that isn't an an argument favoring decision theories which say you should one-box in Transparent Newcomb's.
If you keep a flat probability distribution of what Omega would do to you prior to actually being put into a dilemma, expected-utility-maximizing still favors one-boxing in the opaque version of the dilemma (because based on the information available to you, you have to assign different probabilities to the opaque box containing money depending on whether you one-box or two-box,) but not one-boxing in the transparent version.
You can always contrive scenarios to reward or punish any particular decision theory. The Transparent Newcomb's Problem rewards agents which one-box in the Transparent Newcomb's Problem over agents which two-box, but unless this sort of problem is more likely to arise than ones which reward agents which two-box in Transparent Newcomb's Problem over ones that one-box, that isn't an an argument favoring decision theories which say you should one-box in Transparent Newcomb's.
No, Transparent Newcomb's, Newcomb's and Prisoner's Dilemma with full mutual knowl...
If you've recently joined the Less Wrong community, please leave a comment here and introduce yourself. We'd love to know who you are, what you're doing, what you value, how you came to identify as a rationalist or how you found us. You can skip right to that if you like; the rest of this post consists of a few things you might find helpful. More can be found at the FAQ.
A few notes about the site mechanics
A few notes about the community
If English is not your first language, don't let that make you afraid to post or comment. You can get English help on Discussion- or Main-level posts by sending a PM to one of the following users (use the "send message" link on the upper right of their user page). Either put the text of the post in the PM, or just say that you'd like English help and you'll get a response with an email address.
* Normal_Anomaly
* Randaly
* shokwave
* Barry Cotter
A note for theists: you will find the Less Wrong community to be predominantly atheist, though not completely so, and most of us are genuinely respectful of religious people who keep the usual community norms. It's worth saying that we might think religion is off-topic in some places where you think it's on-topic, so be thoughtful about where and how you start explicitly talking about it; some of us are happy to talk about religion, some of us aren't interested. Bear in mind that many of us really, truly have given full consideration to theistic claims and found them to be false, so starting with the most common arguments is pretty likely just to annoy people. Anyhow, it's absolutely OK to mention that you're religious in your welcome post and to invite a discussion there.
A list of some posts that are pretty awesome
I recommend the major sequences to everybody, but I realize how daunting they look at first. So for purposes of immediate gratification, the following posts are particularly interesting/illuminating/provocative and don't require any previous reading:
More suggestions are welcome! Or just check out the top-rated posts from the history of Less Wrong. Most posts at +50 or more are well worth your time.
Welcome to Less Wrong, and we look forward to hearing from you throughout the site.