I found the Nonviolent Communication method extremely helpful for feeling more connected to my friends.
I've noticed a related phenomenon where, when someone acquires a new insight, they judge its value by how difficult it was to understand, instead of by how much it improves their model of the world. It's the feeling of "Well, I hadn't thought of that before, but I suppose it's pretty obvious." But of course this is a mistake because the important part is "hadn't thought of that before," no matter whether you think you could've realized it in hindsight. (The most pernicious version of this is "Oh, yeah, I totally knew that already. I just hadn't make it so explicit.")
Awesome. PM me if you want to talk more about effective altruism. (I'm currently staffing the EA Summit, so I may not reply swiftly.)
How many rationalists does it take to change a lightbulb?
Just one. They’ll take any excuse to change something.
How many effective altruists does it take to screw in a lightbulb?
Actually, it’s far more efficient if you convince someone else to screw it in.
How many Giving What We Can members does it take to change a lightbulb?
Fifteen have pledged to change it later, but we’ll have to wait until they finish grad school.
How many MIRI researchers does it take to screw in a lightbulb?
The problem is that there are multiple ways to parse that, and while it might naively seem like the ambiguity is harmless, it would actually be disastrous if any number of MIRI researchers tried to screw inside of a lightbulb.
How many CFAR instructors does it take to change a lightbulb?
By the time they’re done, the lightbulb should be able to change itself.
How many Leverage Research employees does it take to screw in a lightbulb?
I don’t know, but we have a team working to figure that out.
How many GiveWell employees does it take to change a lightbulb?
Not many. I don't recall the exact number; there’s a writeup somewhere on their site, if you care to check.
How many cryonicists does it take to change a lightbulb?
Two; one to change the lightbulb, and one to preserve the old one, just in case.
How many neoreactionaries does it take to screw in a lightbulb?
We’d be better off returning to the dark.
I agree when it comes to asking questions about the facts of the situation. On the other hand, asking nonjudgmental questions about the person's feelings is a good way to establish rapport, if that's your goal. (See also)
The counterargument would be to claim that cows > pigs > chickens in intelligence/complexity
My understanding is that pigs > cows >> chickens. Poultry vs mammal is a difficult question that depends on nebulous value judgments, but I thought it was fairly settled that beef causes less suffering/mass than other mammals.
I've found that the process of creating the cards is helpful because it forces me to make the book's major insight explicit. I usually use cloze tests to run through a book's major points. For example, my card for The Lean Startup is:
"The Lean Startup process for continuous improvement is (1) {{c1::identify the hypothesis to test}}, (2) {{c2::determine metrics with which to evaluate the hypothesis}}, (3) {{c3::build a minimum viable product}}, (4) {{c4::use the product to get data and test the hypothesis}}."
This isn't especially helpful if you just remember what the four phrases are, so I use this as a cue to think briefly about each of those concepts.
I frequently give my friends detailed feedback and analysis on their writing. They know about my speed reading thing, and none of them have noticed any change in the quality of my feedback.
This is a difficult problem whose implications go well beyond evaluating charities. Many people seem to defer their evaluation of experts to the experts, but then you have to figure out how to qualify those experts, and I haven't yet seen a good solution to that.
Some heuristics that I use instead:
—Does the expert produce powerful, visible effects in their domain of expertise which non-experts can't duplicate? If so, they're probably reliable within their domain. (For example, engineers can build bridges and authors can make compelling stories, so they're probably reliable in those fields.) This is only useful in cases where a non-expert can evaluate the product's quality; it won't help a non-mathematician evaluate theoretical physics.
—Are the domain experts split into many disagreeing camps? If so, at most one camp is right, which means most of the experts are wrong, and the field isn't reliable. (So this rules out, e.g., experts on nutrition.) This one is a tools for assessing domains of expertise, and won't tell you much about individual experts.