I am reading a post called The Martial Art of Rationality by Eliezer Yudkowsky, in which he makes the following claim:
"If you’re a fast learner, you might learn faster—but the art of rationality isn’t about that; it’s about training brain machinery we all have in common. And where there are systematic errors human brains tend to make—like an insensitivity to scope—rationality is about fixing those mistakes, or finding work-arounds."
this post follows one in which he explained the concept of scope insensitivity by discussing a study that found that persons asked to contribute to dealing with the consequences of fuel spill on wildlife habitat did not contribute more money to save more birds. the contributors were deemed to be insensitive to the scope of the suffering.
the point I want to discuss is whether it is entirely fair to describe scope insensitivity, as defined in this way, as a "systematic human brain error"?
it seems to me that this is bordering on saying that persons who made a different choice to yours are therefore not just wrong, but suffering from something, their brain is not working properly and they need to be taught how to make better choices. where "better" obviously means, more in line with the choice you would make.
scope insensitivity would only be irrational if saving birds were the only criteria in play. to save more birds, give more money. but this is almost never the case, people are more complex than this and they need to consider more criteria than this and each person may consider different criteria and weight them differently. to label those differences as "systematic human brain error" seems to be a very one-dimensional response.
I think we need to bear in mind that the original study did not allow for the possibility that folk did not pay more because they were unable to afford more, or because they would prefer to allocate their charitable spending to alleviate human suffering rather than animal suffering. in fact, the study explicitly said that they were unable to account for the lack of sensitivity to scope. it seems wrong, in fact plain anti-intellectual, for Yudkowsky to claim that their scope insensitivity is a "systematic human brain error"?
please discuss.
It's not about the choice in isolation, it's the mismatch between stated goals and actions. If someone says they want to save money, and they spend tens of hours of their time to avoid a $5 expense when there was a $500 expense they could have avoided with the same effort, then they aren't doing the best thing for their stated goal. Scope-insensitivity problems like this are very common, because quantifying and comparing things is a skill that not everyone has; this causes a huge amount of wasted resources and effort. That doesn't mean everything that looks like an example of scope insensitivity actually is one; people may have other, unstated goals. In the classic study with birds and oil ponds, for example, people might spend a little money to make themself look good to the experimenter.
(I would also note that, while the classic birds-and-oil-ponds example study is often used as an illustrative example, most peoples' belief that scope insensitivity exists and is a problem does not rely on that example, and other examples are easy to find.)
so you agree with me that there my be a rational reason for them not to donate more money. which implies that it is not logical or rational of Eliezer Yudkowsky to ascribe that reason to a human brain error.
thank you.