Clarity comments on Effective Altruism from XYZ perspective - Less Wrong Discussion
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (77)
Thanks for your comment.
Yes, as you stated I was working with the visible sample of EA's who aren't focused on existential risk. I feel the term in relation to existential risk is redundant since effective thinking about existential risk on Lesswrong.
The crowding out effect occurs not just as the individual level (which isn't applicable to individual EA's given room for more funding consideration), but also at the movement level. Because EA's act en-bloc, and factor into their considerations 'what are other people not funding', they compete the supply a demand for donations against established institutional donors like the Gate's Foundation. One might wonder then that if that was true, why those Foundations don't close the funding gaps as a priority - and it looks like someone is trying to answer that here. Admittedly, I haven't got to reading the article fully but from a quick skim it looks like the magnitude of donations of high impact philanthropists is such it compensates for the 'ineffectiveness of their cause', since those charities Givewell recommends have less room for more funding - which becomes a higher order consideration at that scale. The obvious counterexample to this is GiveDirectly, but I wouldn't be suprised if the reason philanthropists don't like them is because of fear of setting a precedence (sp?) against productive mutualistic exchange.
I can't find the original post about the buck stopping after a bit of Googling. I'd like to keep looking into this!
The post I'm referring to is here, but I should note that EY used the phrase in a different context, and my view on terminal values does not reflect his view. My critique of the idea that all human values are complex is that it presupposes too narrow of an interpretation of "values". Let's talk about "goals" instead, defined as follows:
I took the definition from this blogpost I wrote a while back. The comment section there contains a long discussion on a similar issue where I elaborate on my view of terminal values.
Anyway, the way my definition of "goals" seems to differ from the interpretation of "values" in the phrase "human values are complex" is that "goals" allow for self-modification. If I could, I would self-modify into a utilitarian super-robot, regardless of whether it was still conscious or not. According to "human values are complex", I'd be making a mistake in doing so. What sort of mistake would I be making?
The situation is as follows: Unlike some conceivable goal-architectures we might choose for artificial intelligence, humans do not have a clearly defined goal. When you ask people on the street what their goals are in life, they usually can't tell you, and if they do tell you something, they'll likely revise it as soon as you press them with an extreme thought experiment. Many humans are not agenty. Learning about rationality and thinking about personal goals can turn people into agents. How does this transition happen? The "human values are complex" theory seems to imply that we introspect, find out that we care/have intuitions about 5+ different axes of value, and end up accepting all of them for our goals. This is probably how quite a few people are doing it, but they're victim of a gigantic typical mind fallacy if they think that's the only way to do it. Here's what happened to me personally (and incidentally, to about "20+" agents I know personally and to all the hedonistic utilitarians who are familiar with Lesswrong content and still keep their hedonistic utilitarian goals):
I started out with many things I like (friendship, love, self-actualization, non-repetitiveness, etc) plus some moral intuitions (anti-harm, fairness). I then got interested in ethics and figuring out the best ethical theory. I turned into a moral anti-realist soon, but still wanted to find a theory that incorporates my most fundamental intuitions. I realized that I don't care intrinsically about "fairness" and became a utiltiarian in terms of my other-regarding/moral values. I then had the decision to what extent I should invest into utilitarianism/altruism, and how much into values that are more about me specifically. I chose altruism, because I have a strong, OCD-like tendency for doing things either fully or not at all, and I thought saving for retirement, eating healthy etc is just as bothersome as trying to be altruistic, because I don't strongly self-identify with a 100-year-old version of me anyway, so might as well try to make sure that all future sentience will be suffering-free. I still take a lot of care about my long-term happiness and survival, but much less so than if I had the goal to live forever, and as I said I would instantly press the "self-modify into utilitarian robot" button, if there was one. I'd be curious to hear whether I am being "irrational" somewhere, whether there was a step involved that was clearly mistaken. I cannot imagine how that would be the case, and the matter seems obvious to me. So every time I read the link "human values are complex", it seems like an intellectually dishonest discussion stopper to me.