Comment author: CarlShulman 04 January 2012 07:32:50AM *  6 points [-]

Hi Holden,

I just read this thread today. I made a clarification upthread about the description of my comment above, under Louie's. Also, I'd like to register that I thought your characterization of that interview as such was fine, even without the clarifications you make here.

They both argue that standard Bayesian inference indicates against the literal use of non-robust expected value estimates, particularly in "Pascal's Mugging" type scenarios.

As a technical point, I don't think these posts address "Pascal's Mugging" scenarios in any meaningful way.

Bayesian adjustment is a standard part of Pascal's Mugging. The problem is that Solomonoff complexity priors have fat tails, because describing fundamental laws of physics that allow large payoffs is not radically more complex than laws that only allow small payoffs. It doesn't take an extra 10^1000 bits to describe a world where an action generates 2^(10^1000) times as many, e.g. happy puppies. So we can't rule out black swans a priori in that framework (without something like an anthropic assumption that amounts to the Doomsday Argument).

The only thing in your posts that could help with Pascal's Mugging is the assumption of infinite certainty in a distribution without relevantly fat tails or black swans, like a normal or log-normal distribution. But that would be an extreme move, taking coherent worlds of equal simplicity and massively penalizing the ones with high payoffs, so that no evidence that could fit in a human brain could convince us we were in the high-payoff worlds. Without some justification, that seems to amount to assuming the problem away, not addressing it.

Disclaimer 1: This is about expected value measured in the currency of "goods" like happy puppies, rather than expected utility, since agents can have bounded utility, e.g. simply not caring much more about saving a billion billion puppies rather than a billion. This seems fairly true of most people, at least emotionally.

Disclaimer 2: Occam's razor priors give high value to Pascal's Mugging cases, but they also give higher expectations to all other actions. For instance, the chance that space colonization will let huge populations be created increases the expected value of reducing existential risk by many orders of magnitude to total utilitarians. But it also greatly increases the expected payoffs of anything else that reduces existential risk by even a little. So if vaccinating African kids is expected to improve the odds of human survival going forward (not obvious but plausible) then its expected value will be driven to within sight of focused existential risk reductions, e.g. vaccination might be a billionth the cost-effectiveness of focused risk-reduction efforts but probably not smaller by a factor of 10^20. By the same token, different focused existential risk interventions will compete against one another, so one will not want to support the relatively ineffective ones.

Comment author: HoldenKarnofsky 17 January 2012 08:04:23PM 1 point [-]

Carl, it looks like we have a pretty substantial disagreement about key properties of the appropriate prior distribution over expected value of one's actions.

I am not sure whether you are literally endorsing a particular distribution (I am not sure whether "Solomonoff complexity prior" is sufficiently well-defined or, if so, whether you are endorsing that or a varied/adjusted version). I myself have not endorsed a particular distribution. So it seems like the right way to resolve our disagreement is for at least one of us to be more specific about what properties are core to our argument and why we believe any reasonable prior ought to have these properties. I'm not sure when I will be able to do this on my end and will likely contact you by email when I do.

What I do not agree with is the implication that my analysis is irrelevant to Pascal's Mugging. It may be irrelevant for people who endorse the sorts of priors you endorse. But not everyone agrees with you about what the proper prior looks like, and many people who are closer to me on what the appropriate prior looks like still seem unaware of the implications for Pascal's Mugging. If nothing else, my analysis highlights a relationship between one's prior distribution and Pascal's Mugging that I believe many others weren't aware of. Whether it is a decisive refutation of Pascal's Mugging is unresolved (and depends on the disagreement I refer to above).

Comment author: Louie 28 December 2011 09:44:59PM 3 points [-]

Thanks for the helpful comments! I was uninformed about all those details above.

These posts are not about GiveWell's process.

One of the posts has the sub-heading "The GiveWell approach" and all of the analysis in both posts use examples of charities you're comparing. I agree you weren't just talking about the GiveWell process... you were talking about a larger philosophy of science you have that informs things like the GiveWell process.

I recognize that you're making sophisticated arguments for your points. Especially the assumptions that you claim simply must be true to satisfy your intuition that charities should be rewarded for transparency and punished otherwise. Those seem wise from a "getting things done" point of view for an org like GiveWell -- even when there is no mathematical reason those assumptions should be true -- but only a human-level tit-for-tat shame/enforcement mechanism you hope eventually makes this circularly "true" through repeated application. Seems fair enough.

But adding regression adjustments to cancel out the effectiveness of any charity which looks too effective to be believed (based on the common sense of the evaluator) seems like a pretty big finger on the scale. Why do so much analysis in the beginning if the last step of the algorithm is just "re-adjust effectiveness and expected value to equal what feels right"? Your adjustment factor amounts to a kind of Egalitarian Effectiveness Assumption: We are all created equal at turning money into goodness. Or perhaps it's more of a negative statement, like, "None of us is any better than the best of us at turning money into goodness" -- where the upper limit on the best is something like 1000x or whatever the evaluator has encountered in the past. Any claims made above the best limit gets adjusted back down -- those guys were trying to Pascal's Mug us! That's the way in which there's a blinding effect. You disbelieve the claims of any groups who claims to be more effective per capita than you think is possible.

Comment author: HoldenKarnofsky 29 December 2011 12:37:24AM 8 points [-]

Louie, I think you're mischaracterizing these posts and their implications. The argument is much closer to "extraordinary claims require extraordinary evidence" than it is to "extraordinary claims should simply be disregarded." And I have outlined (in the conversation with SIAI) ways in which I believe SIAI could generate the evidence needed for me to put greater weight on its claims.

I wrote more in my comment followup on the first post about why an aversion to arguments that seem similar to "Pascal's Mugging" does not entail an aversion to supporting x-risk charities. (As mentioned in that comment, it appears that important SIAI staff share such an aversion, whether or not they agree with my formal defense of it.)

I also think the message of these posts is consistent with the best available models of how the world works - it isn't just about trying to set incentives. That's probably a conversation for another time - there seems to be a lot of confusion on these posts (especially the second) and I will probably post some clarification at a later date.

Comment author: Louie 28 December 2011 05:14:02AM -5 points [-]

A few corrections.

  • I know that Holden interviewed two other supporters of ours... but I don't think he interviewed 2 other employees. If he did, why did he only publish the unhelpful notes from the one employee he spoke to who didn't know anything?

  • SIAI didn't give Jasen to GiveWell to be interviewed -- Holden chose him unilaterally -- not because he was a good choice, but because Jasen is from New York (just like Holden).

  • I'm unaware of Holden sending his notes to anyone at SIAI prior to publication. Who did he send them to? I never saw them.

  • My guess is Holden sent his notes back to Jasen and called that "sending them to SIAI for feedback". In other words, no one at SIAI who is a leader, or a board member, or someone who understands the plans/finances of the organization saw the notes prior to publication. If Holden had sent the notes to any of the board members of Singularity Institute, they would have sent him tons of corrections.

  • To clarify, I didn't say the interview itself was a lie. I said calling it an interview with SIAI was a lie. I stick by that characterization.

Comment author: HoldenKarnofsky 28 December 2011 03:04:16PM *  20 points [-]

Hi, here are the details of whom I spoke with and why:

  • I originally emailed Michael Vassar, letting him know I was going to be in the Bay Area and asking whether there was anyone appropriate for me to meet with. He set me up with Jasen Murray.
  • Justin Shovelain and an SIAI donor were also present when I spoke with Jasen. There may have been one or two others; I don't recall.
  • After we met, I sent the notes to Jasen for review. He sent back comments and also asked me to run it by Amy Willey and Michael Vassar, who each provided some corrections via email that I incorporated.

A couple of other comments:

  • If SIAI wants to set up another room for more funding discussion, I'd be happy to do that and to post new notes.
  • In general, we're always happy to post corrections or updates on any content we post, including how that content is framed and presented. The best way to get our attention is to email us at info@givewell.org

And a tangential comment/question for Louie: I do not understand why you link to my two LW posts using the anchor text you use. These posts are not about GiveWell's process. They both argue that standard Bayesian inference indicates against the literal use of non-robust expected value estimates, particularly in "Pascal's Mugging" type scenarios. Michael Vassar's response to the first of these was that I was attacking a straw man. There are unresolved disagreements about some of the specific modeling assumptions and implications of these posts, but I don't see any way in which they imply a "limited process" or "blinding to the possibility of SIAI's being a good giving opportunity." I do agree that SIAI hasn't been a fit for our standard process (and is more suited to GiveWell Labs) but I don't see anything in these posts that illustrates that - what do you have in mind here?

Comment author: HoldenKarnofsky 12 November 2011 09:31:44PM 4 points [-]

A few quick notes:

  • As I wrote in my response to Carl on The GiveWell Blog, the conceptual content of this post does not rely on the assumption that the value of donations (as measured in something like "lives saved" or "DALYs saved") is normally distributed. In particular, a lognormal distribution fits easily into the above framework. .

  • I recognize that my model doesn't perfectly describe reality, especially for edge cases. However, I think it is more sophisticated than any model I know of that contradicts its big-picture conceptual conclusions (e.g., by implying "the higher your back-of-the-envelope [extremely error-prone] expected-value calculation, the necessarily higher your posterior expected-value estimate") and that further sophistication would likely leave the big-picture conceptual conclusions in place.

  • JGWeissman is correct that I meant "maximum" when I said "inflection point."

Comment author: HoldenKarnofsky 29 August 2011 04:31:00PM *  9 points [-]

Hello all,

Thanks for the thoughtful comments. Without responding to all threads, I'd like to address a few of the themes that came up. FYI, there are also interesting discussions of this post at The GiveWell Blog , Overcoming Bias , and Quomodocumque (the latter includes Terence Tao's thoughts on "Pascal's Mugging").

On what I'm arguing. There seems to be confusion on which of the following I am arguing:

(1) The conceptual idea of maximizing expected value is problematic.

(2) Explicit estimates of expected value are problematic and can't be taken literally.

(3) Explicit estimates of expected value are problematic/can't be taken literally when they don't include a Bayesian adjustment of the kind outlined in my post.

As several have noted, I do not argue (1). I do aim to give with the aim of maximizing expected good accomplished, and in particular I consider myself risk-neutral in giving.

I strongly endorse (3) and there doesn't seem to be disagreement on this point.

I endorse (2) as well, though less strongly than I endorse (3). I am open to the idea of formally performing a Bayesian adjustment, and if this formalization is well done enough, taking the adjusted expected-value estimate literally. However,

  • I have examined a lot of expected-value estimates relevant to giving, including those done by the DCP2 , Copenhagen Consensus , and Poverty Action Lab , and have never once seen a formalized adjustment of this kind.

  • I believe that often - particularly in the domains discussed here - formalizing such an adjustment in a reasonable way is simply not feasible and that using intuition is superior. This is argued briefly in this post, and Dario Amodei and Jonah Sinick have an excellent exchange further exploring this idea at the GiveWell Blog.

  • If you disagree with the above point, and feel that such adjustments ought to be done formally, then you do disagree with a substantial part of my post; however, you ought to find the remainder of the post more consequential than I do, since it implies substantial room for improvement in the most prominent cost-effectiveness estimates (and perhaps all cost-effectiveness estimates) in the domains under discussion.

All of the above applies to expected-value calculations that take relatively large amounts of guesswork, such as in the domain of giving. There are expected-value estimates that I feel are precise/robust enough to take literally.

Is it reasonable to model existential risk reduction and/or "Pascal's Mugging" using log-/normal distributions? Several have pointed out that existential risk reduction and "Pascal's Mugging" seem to involve "either-or" scenarios that aren't well approximated by log-/normal distributions. I wish to emphasize that I'm focused on the prior over expected value of one's actions and on the distribution of error in one's expected-value estimate. (The latter is a fuzzy concept that may be best formalized with the aid of concepts such as imprecise probability. In the scenarios under discussion, one often must estimate the probability of catastrophe essentially by making a wild guess with a wide confidence interval, leaving wide room for "estimate error" around the expected-value calculation.) Bayesian adjustments to expected-value estimates of actions, in this framework, are smaller (all else equal) for well-modeled and well-understood "either-or" scenarios than for poorly-modeled and poorly-understood "either-or" scenarios.

For both the prior and for the "estimate error," I think the log-/normal distribution can be a reasonable approximation, especially when considering the uncertainty around the impact of one's actions on the probability of catastrophe.

The basic framework of this post still applies, and many of its conclusions may as well, even when other types of probability distributions are assumed.

My views on existential risk reduction are outside the scope of this post. The only mention I make of existential risk reduction is to critique the argument that "charities working on reducing the risk of sudden human extinction must be the best ones to support, since the value of saving the human race is so high that 'any imaginable probability of success' would lead to a higher expected value for these charities than for others." Note that Eliezer Yudkowsky and Michael Vassar also appear to disapprove of this argument, so it seems clear that disputing this argument is not the same as arguing against existential risk reduction charities.

For the past few years we have considered catastrophic risk reduction charities to be lower on GiveWell's priority list for investigation than developing-world aid charities, but still relatively high on the list in the scheme of things. I've recently started investigating these causes a bit more, starting with SIAI (see LW posts on my discussion with SIAI representatives and my exchange with Jaan Tallinn). It's plausible to me that asteroid risk reduction is a promising area, but I haven't looked into it enough (yet) to comment more on that.

My informal objections to what I term EEV. Several have criticized the section of my post giving informal objections to what I term the EEV approach (by which I meant explicitly estimating expected value using a rough calculation and not performing a Bayesian adjustment). This section was intended only as a very rough sketch of what unnerves me about EEV; there doesn't seem to be much dispute over the more formal argument I made against EEV; thus, I don't plan on responding to critiques of this section.

Comment author: patrissimo 02 January 2011 07:04:23AM 4 points [-]

Completely agree with your general point on marginal analysis (although I'm a TDT skeptic), and am a fan of GiveWell, but this is trivially wrong:

It is not possible for everyone to behave this way in elections: no voter is able to consider the existing distribution of votes before casting their own.

This seems to assume away information about the size of the electorate as well as any predictive power about the outcome. Surely the marginal benefit of a Presidential vote in a small swing state is massively higher than in a large solidly Democratic state, for example. And in addition to historical results, there is polling data in advance of the election to improve predictions.

Besides this being theoretically true, we can see it empirically from the spending patterns of both Presidential campaigns and political parties on Congressional races. They allocate money to the states / races where they believe it will do the most marginal good, which is often a very inequal distribution. Thus they do, in fact "consider the existing distribution of votes before casting" their advertising dollars.

Comment author: HoldenKarnofsky 03 January 2011 09:43:55PM 3 points [-]

Patrissimo, fair enough. I was thinking that voters can't vote with the same degree of knowledge of the existing situation that they can have with blood donations. Arguments over TDT certainly seem more relevant to voting than to blood donations. But you are right that voters have lots of relevant information about the likely distribution of votes that can be productively factored into their decisions regardless of the TDT debate. Glad to hear you're a fan of GiveWell.

Comment author: Perplexed 25 December 2010 07:08:34AM *  5 points [-]

I take it that you're suggesting marginal analysis based on the standard correct classical causal decision theory (in which no one is responsible for saving a life by donating blood unless someone would have actually died had that donation not been made) out of either belated humility about the probability of an SIAI-originating decision theory being correct, or because you're planning to actually convince someone and you don't want to invoke Hofstadterian superrationality in place of the standard correct decision theory?

:)

My guess would be that at the margin, a blood donation saves less than 0.00001 lives. (Otherwise, compensation would be increased for the paid donors). But, if you want to use a TDT/UDT style analysis, here are some relevant statistics from the American Red Cross:

  • The number of blood donations collected in the U.S. in a year: 16 million (2006).
  • The number of patients who receive blood in the U.S. in a year: 5 million (2006).

Given these numbers, I would estimate that roughly 0.5 million (US) lives are saved (more accurately, extended) by blood products annually. If you adopt the assumption that all blood comes from voluntary, uncompensated donations, and divide those 0.5 million lives among the 16 million annual donations, you get one life saved for every 32 pints donated - not as much as jsteinhardt hoped, but still significant enough to earn a major warm-and-fuzzy.

Comment author: HoldenKarnofsky 29 December 2010 05:01:06PM 15 points [-]

This is Holden Karnofsky, the co-Executive Director of GiveWell, which is referenced in the top-level article and elsewhere on this thread.

I think there is an important difference between discussing the marginal impact of a blood donation and the marginal impact of a vote. When it comes to blood donations, it is possible for everyone to simultaneously follow the rule: "Give blood only when the supply of donations is low enough that an additional donation would have high expected impact", with a reasonable outcome. It is not possible for everyone to behave this way in elections: no voter is able to consider the existing distribution of votes before casting their own.

I am only casually familiar with TDT/UDT, but it seems to me that that "Give blood only when the supply of donations is low enough that an additional donation would have high expected impact" should get about the same amount of credit under TDT/UDT as giving blood, and thus the extra impact of actually giving blood (as opposed to following that rule) is small regardless of what decision theory one is using.

I bring this up because the discussion of marginal blood donations is parallel to analysis GiveWell often does of the marginal impact of donations. We do everything we can to understand the marginal (not average) impact of a donation and recommend organizations on this basis, and we believe this is a very important and unique element of what we offer (more on this issue). We try to push donors to underfunded charities and away from overfunded ones, and I do not think the validity of this depends on any controversial (even controversial-within-Less-Wrong) view on decision theory, though I am open to arguments that it does.

Comment author: shokwave 25 December 2010 03:09:41PM *  4 points [-]

It is possible that Donor A may choose to donate fully to GiveWell for many reasons, including a prior assumption that it's 50:50 or better without checking easily available facts. This reflects badly on Donor A, not GiveWell, and does not in any way make a case for calling GiveWell "ethically questionable". The most you could possibly say is "GiveWell does not overly pander to the lowest common denominator enough" but these people are already donating their money to Make a Wish foundation or something equally silly.

I belabour this point because charities run solely on their appearance as ethical, and to the extent that your comments deprive GiveWell of possible donations on the basis of spurious claims, you're doing a bad thing.

Comment author: HoldenKarnofsky 29 December 2010 04:39:24PM *  16 points [-]

This is Holden Karnofsky, the co-Executive Director of GiveWell. As a frequent Less Wrong reader, I'm really glad to see the thoughtful discussion here. Thanks to Yvain for calling attention both to GiveWell and to the general topic of effective giving.

First off, much of this content overlaps with our own, so people interested in this thread might also find the following links interesting:

I'm mostly posting to clarify a few things regarding the concerns that have been raised about GiveWell (by aeschenkarnos).

  • We regret the astroturfing that aeschenkarnos brought up. This incident is disclosed, along with other mistakes we've made, on our shortcomings list , which is accessible via a top-level link on our navigation bar.
  • Regarding the split between grants to charities and funds spent on our own operations:
    • Early in our existence, we relied on making grants of our own to charities. We weren't able to point them to any benefits that would come from our recommendations (since we were new and had no track record of influencing donations), so rather than inviting them to be reviewed, we invited them to apply for grants (subject to certain conditions such as public disclosure of application materials). Grantmaking is no longer important to our process and we no longer solicit donations to be regranted, though we still occasionally receive them. That explains why the % of our funds spent on grants has fallen a lot, though it hasn't hit zero.
    • At this point, we actively solicit donations to GiveWell only when dealing with institutional funders or with people who have a relationship with us. When dealing with the general public, we put the solicitation on behalf of recommended charities - rather than ourselves - front and center. Our top charities page, linked prominently from our front page and navigation bar and in other places throughout the site, links to "donate" pages for top charities ( here's the one for our top-rated charity VillageReach ) that allow us to track donations, but otherwise take no part in the donation process (the money does not touch our bank account). These "donate" pages also are linked from charity reviews. The only way to get to the "Donate to GiveWell" page is under "About GiveWell." If donors make a considered decision to support us rather than our top charities, we want them to be able to do so, but our site is designed to push the casual user to our top charities.
    • In 2009 we tracked ~$1 million in donations to our top charities as as result of our research, while our own operating (non-grant) expenses were under $300k. We expect 2010 to have a higher "donations to top charities" figure on similar operating expenses. We are still new and hope the ratio will improve substantially over time.
    • We have a policy of regranting unrestricted funds if our reserves go above a certain level; we don't believe in building a massive endowment for ourselves. This is the only condition under which we regrant unrestricted funds. We don't want donors to fear that we might blindly pile up reserves without limit (we won't), but we don't want to get into all the details of our "Excess reserves" policy on the Donate page, so we went with the language: "we may use these funds for operating expenses or grants to charities, at our discretion."
    • Bottom line - grantmaking used to be an important part of what we do but it isn't now; the % of our funds spent on grants is not a meaningful figure.
  • Regarding Charity Navigator:
    • I believe Yvain is correct to say that Charity Navigator does not evaluate effectiveness (and admits this) and that GiveWell does. See also this recent New York Times article on planned changes at Charity Navigator and Charity Navigator's disclosure of the full details of its current methodology.
    • I agree with alexanderis that "number of charities rated" is higher for Charity Navigator primarily because its research is not as in-depth. I believe Charity Navigator would agree with this as well.
    • I believe that Charity Navigator has a significantly higher profile than GiveWell, overall, and know of no evidence suggesting otherwise. However, GiveWell does have a higher profile within certain communities, including Less Wrong. I attribute our higher profile on Less Wrong to specific individuals including Michael Vassar, Anna Salomon, Carl Shulman, Razib at GNXP, and multifoliaterose. I don't believe any of these individuals have plugged GiveWell in ignorance of Charity Navigator (in fact I have probably discussed the differences specifically with each of them).

We've worked to find the best, most cost-effective charities (in terms of actual impact per marginal dollar) and write up all the details of our analysis. We welcome more comments and questions about our work, whether here, on our blog, or via email.

View more: Prev