Thanks! I do also rely to some extent on reasoning... for example, Chapter 3 is my argument for why we should expect to be better off with (on the margin) more scout mindset and less soldier mindset, compared to our default settings. I point out some basic facts about human psychology (e.g., the fact that we over-weight immediate consequences relative to delayed consequences) and explain why it seems to me those facts imply that we would have a tendency to use scout mindset less often than we should, even just for our own self interest.
The nice thing about argumentation (as compared to citing studies) is that it's pretty transparent -- the reader can evaluate my logic for themselves and decide if they buy it.
Hey Ozzie! Thanks for reading / reviewing.
I originally hoped to write a more “scholarly” book, but I spent months reading the literature on motivated reasoning and thought it was mostly pretty bad, and anyway not the actual cause of my confidence in the core claims of the book such as “You should be in scout mindset more often.” So instead I focused on the goal of giving lots of examples of scout mindset in different domains, and addressing some of the common objections to scout mindset, in hopes of inspiring people to practice it more often.
I left i...
I am really glad about this choice, and also made similar epistemic updates over the last few years, and my guess is if I was to write a book, I would probably make a similar choice (though probably with more first-principles reasoning and a lot more fermi-estimates, though the latter sure sounds like it would cut into my sales :P).
This doesn't really ring true to me (as a model of my personal subjective experience).
The model in this post says despair is "a sign that important evidence has been building up in your buffer, unacknowledged, and that it’s time now to integrate it into your plans."
But most of the times that I've cycled intermittently into despair over some project (or relationship), it's been because of facts I already knew, consciously, about the project. I'm just becoming re-focused on them. And I wouldn't be surprised if things like low blood sugar or anxie...
Hey, I'm one of the founders of CFAR (and used to teach the Reference Class Hopping session you mentioned).
You seem to be misinformed about what CFAR is claiming about our material. Just to use Reference Class Hopping as an example: It's not the same as reference class forecasting. It involves doing reference class forecasting (in the first half of the session), then finding ways to put yourself in a different reference class so that your forecast will be more encouraging. We're very explicit about the difference.
I've emailed experts in reference class for...
I usually try to mix it up. A quick count shows 6 male examples and 2 female examples, which was not a deliberate choice, but I guess I can be more intentional about a more even split in future?
Thanks for showing up and clarifying, Sam!
I'd be curious to hear more about the ways in which you think CFAR is over-(epistemically) hygienic. Feel free to email me if you prefer, but I bet a lot of people here would also be interested to hear your critique.
Sure, here's a CDC overview: http://www.cdc.gov/handwashing/show-me-the-science-hand-sanitizer.html They seem to be imperfect but better than nothing, and since people are surely not going to be washing their hands every time they cough, sneeze, or touch communal surfaces, supplementing normal handwashing practices with hand sanitizer seems like a probably-helpful precaution.
But note that this has turned out to be an accidental tangent since the "overhygienic" criticism was actually meant to refer to epistemic hygiene! (I am potentially also indi...
Edited to reflect the fact that, no, we certainly don't insist. We just warn people that it's common to get sick during the workshop because you're probably getting less sleep and in close contact with so many other people (many of whom have recently been in airports, etc.). And that it's good practice to use hand sanitizers regularly, not just for your own sake but for others'.
and in close contact with so many other people
So, people who commute by public transportation in a big city are just screwed, aren't they? :-)
it's good practice to use hand sanitizers regularly
I don't think so -- not for people with a healthy immune system.
Perhaps this is silly of me, but the single word in the article that made me indignantly exclaim "What!?" was when he called CFAR "overhygienic."
I mean... you can call us nerdy, weird in some ways, obsessed with productivity, with some justification! But how can you take issue with our insistence [Edit: more like strong encouragement!] that people use hand sanitizer at a 4-day retreat with 40 people sharing food and close quarters?
[Edit: The author has clarified above that "overhygienic" was meant to refer to epistemic hygiene, not literal hygiene.]
"I'm beginning to find this conversation aversive, and I'm not sure why. I propose we hold off until I've figured that out."
I read this suggested line and felt a little worried. I hope rationalist culture doesn't head in that direction.
There are plenty of times when I agree a policy of frankness can be useful, but one of the risks of such a policy is that it can become an excuse to abdicate responsibility for your effect on other people.
If you tell me that you're having an aversive reaction to our conversation, but can't tell me why, it's goi...
Yes, my version of this always goes, "I'm finding this conversation aversive and I don't know why. Hold on while I figure it out." In other words, it doesn't delay a conversation until later, but it does mean that I close my eyes for 60 seconds and think.
I also find that line a bit strange. In nearly all cases where I would expect that someone says: "I'm beginning to find this conversation aversive, and I'm not sure why" I think I would take it as a topic change to why the conversation might bring up negative emotions in the person.
If we are in an environment of open conversation and I say something that brings up an emotional trauma in another person and that person doesn't have the self-awareness to know why he's feeling unwell, that's not a good time to leave him alone.
Interesting, I have the exact opposite gut reaction. It could be rephrased in slight variations, e.g. "until we've figured that out", or, as shokwave below suggested, with a request for assistance, but in general, if someone said that to me, I would, ceteris paribus, infer that they are a self-aware and peaceful/cooperative person and that they are not holding anything in particular against me.
Whereas when someone leaves a conversation with an excuse that may or may not be genuine, it leaves me totally stressed-out because I have no idea what's g...
Yes, that makes a lot of sense!
Since we don't have any programmers on staff at the moment, we went with the less-than-ideal solution of a manual thermometer, which we update about once a day -- but it certainly would be better to have it happen automatically.
For now, I've gone with the kluge-y solution of an "Updated January XXth" note directly above the menu bar. Thanks for the comment.
several mainstream media articles about CFAR on their way, including one forthcoming shortly in the Wall Street Journal
That article's up now -- it was on the cover of the Personal Journal section of the WSJ, on December 31st. Here's the online version: More Rational Resolutions
Great one, thanks!
Agreed. I might add them to a future version of this map.
This time around I held off mainly because I was confounded by how to add them; drugs really do pervade so many of these groups, in different variants: psychadelics are strong among the counterculture and New Age culture, nootropics are more popular among rationalists and biohackers/Quantified Self, and both are popular among transhumanists. (See this H+ article for a discussion of psychadelic transhumanists.)
Well, I'd say that LW does take account of who we are. They just haven't had the impetus to do so quite as thoroughly as CFAR has. As a result there are aspects of applied rationality, or "rationality for humans" as I sometimes call it, that CFAR has developed and LW hasn't.
If it makes you feel less hesitant, we've given refunds twice. One person at a workshop last year who said he'd expected polish and suits, and another who said he enjoyed it but wasn't sure it was going to help enough with his current life situation to be worth it.
Fixed now, sorry!
Fixed! Thanks, I apparently didn't understand how links worked in this system.
Not sure what kind of evidence you're looking for here; that's just a description of our selection criteria for attendees.
Preferring utilitarianism is a moral intuition, just like preferring Life Extension. The former's a general intuition, the latter's an intuition about a specific case.
So it's not a priori clear which intuition to modify (general or specific) when the two conflict.
Right -- I don't claim any of my moral intuitions to be true or correct; I'm an error theorist, when it comes down to it.
But I do want my intuitions to be consistent with each other. So if I have the intuition that utility is the only thing I value for its own sake, and I have the intuition that Life Extension is better than Replacement, then something's gotta give.
When our intuitions in a particular case contradict the moral theory we thought we held, we need some justification for amending the moral theory other than "I want to."
I agree, and that's why my intuition pushes me towards Life Extension. But how does that fact fit into utilitarianism? And if you're diverging from utilitarianism, what are you replacing it with?
Excellent.
One doesn't have to be better than the other. That's what's in dispute.
I think making this comparison is important philosophically, because of the implications our answer has for other utilitarian dilemmas, but it's also important practically, in shaping our decisions about how to allocate our efforts to better the world.
Thanks -- but if I'm reading your post correctly, your arguments hinge on the utility experienced in Life Extension being greater than that in Replacement. Is that right? If I stipulate that the utility is equal, would your answer change?
Ah, true! I edited it again to include the original setup, so that people will know what Logos01 and drethelin are referring to.
Thanks -- I fixed the setup.
My framing was meant to be encouraging you to disproportionately question beliefs which, if false, make you worse off. But motivated skepticism is disproportionately questioning beliefs that you want to be false. That's an important difference, I think.
Are you claiming that my version is also a form of motivated skepticism (perhaps a weaker form)? Or do you think my version's fine, but that I need to make it clearer in the text how what I'm encouraging is different from motivated skepticism?
Incidentally, the filmmaker didn't capture my slide with the diagram of the revised model of rationality and emotions in ideal human* decision-making, so I've uploaded it.
The Straw Vulcan model of ideal human* decisionmaking: http://measureofdoubt.files.wordpress.com/2011/11/screen-shot-2011-11-26-at-3-58-00-pm.png
My revised model of ideal human* decisionmaking: http://measureofdoubt.files.wordpress.com/2011/11/screen-shot-2011-11-26-at-3-58-14-pm.png
*I realize now that I need this modifier, at least on Less Wrong!
Great point, in many cases, such as when you're trying to decide what school to go to, and you make the decision deliberatively but taking into account the data from your intuitive reactions to the schools.
But in other cases, such as chess-playing, aren't you mainly just deciding based on your System 1 judgments? (Admittedly I'm no chess player; that's just my impression of how it works.)
I agree you need to use System 2 for your meta-judgment about which system to use in a particular context, but once you've made that meta-judgment, I think there are some cases in which you make the actual judgment based on System 1.
Am I correctly understanding your point?
Yup, I went through the same reasoning myself -- I decided on "system 1" and "system 2" for their neutral tone, and also because they're Stanovich's preferred terms.
Good question. My intended meaning was closest to (h). (Although isn't (g) pretty much equivalent?)
Hey, thanks for the shoutout! @SilasBarta -- Yeah, I first encountered the mirror paradox in G&R, but I ended up explaining it differently than Drescher did, drawing on Gardner as well as some discussions with a friend, so I didn't end up quoting Drescher after all. I do like his explanation, though.
This was a really clarifying post for me. I had gotten to the point of noticing that "What is X?" debates were really just debates over the definition of X, but I hadn't yet taken the next step of asking why people care about how X is defined.
I think another great example of a disguised query is the recurring debate, "Is this art?" People have really widely varying definitions of "art" (e.g., some people's definition includes "aesthetically interesting," other people's definition merely requires "conceptually i...
Eliezer, you wrote:
But when you're really done, you'll know you're done. Dissolving the question is an unmistakable feeling...
I'm not so sure. There have been a number of mysteries throughout history that were resolved by science, but people didn't immediately feel as if the scientific explanation really resolved the question, even though it does to us now -- like the explanation of light as being electromagnetic waves.
I frequently find it tricky to determine whether a feeling of dissatisfaction indicates that I haven't gotten to the root of a problem,...
I like the cuteness of turning an old parlor game into a theory-test. But I suspect a more direct and effective test would be to take one true fact, invert it, and then ask your test subject which statement fits their theory better. (I always try to do that to myself when I'm fitting my own pet theory to a new fact I've just heard, but it's hard once I already know which one is true.)
Other advantages of this test over the original one proposed in the post: (1) You don't have to go to the trouble of thinking up fake data (a problematic endeavor, because the...
... By the way, you might've misunderstood the point of the Elon Musk examples. The point wasn't that he's some exemplar of honesty. It was that he was motivated to try to make his companies succeed despite believing that the most likely outcome was failure. (i.e., he is a counterexample to the common claim "Entrepreneurs have to believe they are going to succeed, or else they won't be motivated to try")