All of Benjamin_Todd's Comments + Replies

I'm open to that and felt unsure the post was a good idea after I released it. I had some discussion with him on twitter afterwards, where we smoothed things over a bit: https://x.com/GaryMarcus/status/1888604860523946354

Thanks for the comments! 

I'm especially keen to explore bottlenecks (e.g. another suggestion I saw is that to reach 1 billion a year would require 10x current global lithium production to supply the batteries.)

A factor of 2 for increased difficultly due to processing intensity seems reasonable, and I should have thrown it in. (Though my estimates were to an order of magnitude so this probably won't change the bottom line, and on the other side, many robots will weigh <100kg and some will be non-humanoid.)

Thanks, and fair points!

Note that it you convert only half the car factories, you can still produce 0.5 billion robots per year, so it doesn't change the basics picture that much. (It's all order of magnitude stuff.)

I talk a little about some other estimates - a standard trajectory would be 20-30 years on the long end. ASI enabled could be even faster than 5yr. I agree it would be nice to flesh these out more.

Also agree it would be good to figure out the conversion efficiency better. One factor on the other side is robots involve lighter parts, which apparently makes it easier. Ideally we'd also check for other input factors that could bottleneck production -eg lithium for batteries at over 100m.

That's helpful! Makes me think the all in hardware costs could be off by a factor of 2x. 

I did wonder about maintenance costs, but I figured they wouldn't change the picture too much because I only assume an avg 3 year lifetime for the robot, and figured they wouldn't need a huge amount of maintenance to make it to that point.

Moreover, if there's worthwhile maintenance that extends the lifetime further, then the hardware costs could end up cheaper than my per year estimate. 

I'm also envisioning the costs after a big scale up, and there would be robot repair shops as numerous as car repair, rather than needing to fly in specialists.

That said, I agree it would be interesting to look at how much is spent on car maintenance per year on a car vs. capital costs. (I expect it would be under 10%?)

5frontiersummit
The average American drives 45,000 miles in three years, but a car operated 20/7 (like your robot) would accumulate about a million miles in that timeframe. Probably it would go through 2 engines and 3 transmissions if it could even be kept on the road. All things being equal it would need 22x as much maintenance than the average of the US fleet, so probably more like 220% of the capital cost. A really nice printer/photocopier-combo costs about $10,000 like your robot, and is built from of motors, cameras, and computers just like your robot. While it's mature technology and built to generally high quality standards, if you try running copies 24/7 you will quickly be on a first-name basis with local Kyocera guy.

I'd be happy to put the opening bunch of paragraphs. I was feeling reluctant to cross-post because I often update my articles as I learn more about a topic, and I don't want to keep multiple versions in sync (especially for a lower priority article).

6Ben Pace
That makes sense. We have something of a solution to this where users with RSS crossposting can manually re-sync the post from the triple-dot memu. I'll DM you about how to set it up if you want it.

Yes - if anyone reading knows more about manufacturing and could comment on how easy it would be to convert, that would be very helpful.

I also agree it would be interesting to try to do more analysis of how much ASI and robotics could speed up construction of robot factories, by looking at different bottlenecks and how much they could help.

I'm not sure a robot workforce would have a huge effect initially, since there's already a large pool of human workers (though maybe you get some boost by making everything run 24/7). However, at later stages it might become hard to hire enough human workers, while with robots you could keep scaling.

Thanks, great comment.

Seems like we roughly agree on the human-only case. My thinking was that the profit margin would initially be 90-99%, which would create huge economic incentives. Though incentives and coordination were probably stronger in WW2, which could make things slower. Also 10x per year for 5 years sounds like a lot – helpful to point out they didn't quite achieve that in WW2.

With ASI, I agree something like another 5x speed-up sounds plausible.

I agree (1) and (2) are possibilities. However, from a personal planning pov, you should focus preparing for scenarios (i) that might last a long time (ii) where you can affect what happens, since that's where the stakes are.

Scenarios where we all die soon can be mostly be ignored, unless you think they make up most of the probability. (Edit: to be clear it does reduce the value of saving vs. spending, just don't think it's a big effect unless probabilities are high.)

I think (3) is the key way to push back. 

I feel unsure all my preferences are either ... (read more)

1wassname
I would disagree: unless you can change the probability. In which case they can still be significant in your decision making, if you can invest time or money or effort to decrease the probability.

True, though I think many people have the intuition that returns diminish faster than log (at least given current tech).

For example, most people think increasing their income from $10k to $20k would do more for their material wellbeing than increasing it from $1bn to $2bn.

I think the key issue is whether new tech makes it easier to buy huge amounts of utility, or that people want to satisfy other preferences beyond material wellbeing (which may have log or even close to linear returns).

I agree the lower bound for output isn't very tight. I'd be very interested to hear other simple rules of thumb you could use to provide a tighter one.

I'll add a note to the section on input tokens that since they don't require KV cache, it's possible to get much closer to the upper bound. 

The opportunities for algorithmic improvements go far beyond the parallelization and mixture of experts methods you mention.

 

I agree. I'd be very interested in anyone's forecasts for how they might evolve.

I've been working with (very roughly) another ~10x or so improvement in "inference efficiency" by 2030 (or how to measure this and make sure it's independent from other factors). 

By this I mean that if we were able to train a model with 10^26 FLOP this year, achieving a fixed level of learning efficiency, it would require 10X FLOP to generate useful output, while by 2030 it would only require X FLOP to get the same output. 

Thanks that's interesting!

Can I double check, do you think this affects the bottom lines?

The bottom line is supposed to be that FLOP/s vs. FLOP per forward pass can be used as an upper bound, and memory bandwidth vs. model size can be used as an lower bound, and real life efficiency falls somewhere in the middle depending on a many factors (inc. length of KV cache), which I don't try to get into, but is plausibly around 15% of the upper bound for GPT-4 on H100s.

Are you saying that the lower bound for output tokens should maybe be even lower, because the KV cache can be larger than the model weights?

7ryan_greenblatt
The lower bound of "memory bandwidth vs. model size" is effectively equivalent to assuming that the batch size is a single token. I think this isn't at all close to realistic operating conditions and thus won't be a very tight lower bound. (Or reflect the most important bottlenecks.) I think that the KV cache for a single sequence won't be larger than the model weights for realistic work loads, so the lower bound should still be a valid lower bound. (Though not a tight one.) I think the bottom line number you provide for "rough estimate of actual throughput" ends up being pretty reasonable for output tokens and considerably too low for input tokens. (I think input tokens probably get more like 50% or 75% flop utilization rather than 15%. See also the difference in price for anthropic model.) That said, it doesn't seem like a good mechanism for estimating throughput will be to aggregate the lower and upper bounds you have as the lower bound doesn't have much correspondence with actual bottlenecks. (For instance, this lower bound would miss that mamba would get much higher throughput.) I also think that insofar as you care about factors of 3-5 on inference efficiency, you need to do different analysis for input tokens and output tokens. (I also think that input tokens get pretty close to the pure FLOP estimate. So, another estimation approach you can use if you don't care about factors of 5 is to just take the pure flop estimate and then halve it to be account for other slow downs. I think this estimate gets input tokens basically right and is wrong by a factor of 3-5 for output tokens.) It seems like your actual mechanism for making this estimate for the utilization on output tokens was to take the number from semi-analysis and extrapolate it to other GPUs. (At least the number matches this?) This does seem like a reasonable approach, but it isn't particularly tethered to your lower bound.

You need to add in the endowments of the colleges as well. The richest college at Cambridge (Trinity) has an endowment of about $1.5bn; whereas the richest college at Oxford has only about $300m.

0joaolkf
Cambridge's total colleges endowments is 2.8 and Oxford's 2.9. But the figures above already include this.

What are the chances of being a billionaire or getting $30m plus if you go to Harvard rather than an elite uni?

And then what about HBS rather than Harvard?

Agree - Glassdoor is mainly designed to appeal to job seekers. The way they get their data is by only granting access if you reveal your salary. So the salary data ends up tilted towards the people who are seeking jobs.

There's also a sampling problem. Google has ~10,000 engineers, but there's probably only ~100 who earn $1mn+. Large companies normally only have a couple of responses, so even if you sampled everyone randomly, you'd only get ~1 top earner in the sample.

0John_Maxwell
Still interesting to know that there are so many GS VPs getting paid so little though, isn't it? I would expect that if what you say is true, and GS VPs are well-paid like the consensus from other sources indicates, we'd see very few salary reports from people at the VP level (they're not seeking jobs!) But in fact there are 102 VP salaries, contrasted with 127 Senior Analyst Salaries and 536 Associate salaries... it seems like associates occur at at least a 5 to 1 ratio with VPs within the organization, if not a higher ratio.

Hi Jonah,

Great posts.

I agree these figures show it's plausible that the value of donations in finance are significantly larger than the direct economic contribution of many jobs, though I see it as highly uncertain. When you're working in highly socially valuable sectors like research or some entrepreneurship, it seems to me that the two are roughly comparable, with big error bars.

However, I don't think this shows it's plausible that earning to give is likely to be the path towards doing the most good. There are many careers that seem to offer influence ov... (read more)

1gjm
Do you (meaning, I guess, anyone reading this) have a good idea of just how altruistic typical people considering "earning to give" are? I mean: a perfect altruist will indeed be looking simply for "the path towards doing the most good", but someone who merely cares much more than most about the welfare of the world's poorest people (or the dangers of runaway artificial intelligence, or eradicating disease, or ending aging, or whatever) will presumably attach some weight to their own earnings. It seems like it could easily be true that (1) there are other things a smart hard-working person could do that do more good overall than ETG, but (2) ETG handily beats them in terms of the actual preferences of most people contemplating ETG. (Though there might be benefits in not acknowledging those actual preferences too openly: e.g., doing so might make people feel less good about ETG and therefore less inclined to do it, or it might encourage them to put a higher weight on their own personal gain and therefore give less.)
1JonahS
Thanks for all of these thoughts I have the same intuition, but I would be interested in hearing about whether you have object level reasons for thinking so. Quoting from an email that I wrote I would be interested in seeing more analysis of flow-through effects of interventions in the developed world: when it comes to general efforts to increase economic growth / tech speedup, I don't see an object level case for there being disproportionate flow-through effects coming from working in the developed world (though I still give the possibility substantial weight on priors).

Glassdoor rarely properly includes the top paid employees (those people don't fill out the survey). According to Goldman's own figures, mean compensation per employee (across all employees) is ~$400k. It'll be significantly higher if you're in front office. Your expected earnings from a Goldman job are roughly the mean earnings multiplied by the expected number of years you'll stay at the firm.

0John_Maxwell
Why? It seems like if anything the most salary-obsessed people would be driven to achieve high salaries and driven to compare their salaries on Glassdoor.

I think both research and advocacy (both to governments and among individuals) are highly important, and it's very unclear which is more important at the margin.

It's too simple to say basic research is more important, because advocacy could lead to hugely increased funding for basic research.

We've collated a list of all the approaches that seem to be on the table in the effective altruism community for improving the long-run future. There's some other options, including funding GiveWell and GCRI. This doc also explains a little more of the reasoning behind the approaches. If you like more detail on how 80k might help reduce the risk of extinction, drop me an email at ben@80000hours.org.

In general, I think the question of how best to improve the long-run future is highly uncertain, but has high value of information, so the most important activi... (read more)

Note that Toby is a trustee of CEA and did most of his government consulting due to GWWC, not the FHI, so it's not clear that FHI wins out in terms of influence over government.

Moreover, if your concern is influence over government, CEA could still beat FHI (even if FHI is doing very high level advocacy) by acting as a multiplier on the FHI's efforts (and similar orgs): $1 donated to CEA could lead to more than $1 of financial or human capital delivered to the FHI or similar. I'm not claiming this is happening, but just pointing out that it's too simple to say FHI wins out just because they're doing some really good advocacy.

Disclaimer: I'm the Executive Director of 80,000 Hours, which is part of CEA.

2Sean_o_h
Re: point 1: The bulk of our policy consultations to date have actually been Nick Bostrom, although Anders Sandberg has done quite a bit, Toby has been regularly consulting with the UK government recently, and I've been doing some lately (mostly wearing my CSER hat, but drawing on my FHI expertise, so I would give FHI credit there ;) ) and others have also done bits and pieces.
0joaolkf
I don't have the numbers of the top of my head, but the bulk of the consultations in my list are due to Nick. I believe there are even much more done by him previous to FHI even existing back in the 90s. Nonetheless, I would guess he is probably very much willing to transfer the advocacy to CEA and similar organizations, as it seems to be already happening. In my opinion, that isn't FHI main role at all, even though they been doing it a lot. As a wild guess, I would be inclined to say he probably actively rejects a few consultations by now. As I said, we need research. Influence over the government is useless - and perhaps harmful - without it. While they work together, I'm not sure advocacy and influence over the government are quite the same. I think advocacy here might just be seen as close to advertising and movement building, which in turn will create political pressure. Quite another thing is to be asked by the government to offer ones opinion.

Read the response to poor cause choice and inconsistent attitude toward rigor as "while some EAs might be donating without enough thought, lots of others are investing most of their resources in doing more research"

The monoculture problem is something we often think about how to fix at 80k. We haven't come up with great solutions yet though.

I also argued that the decline in the FB group is not obviously important. And if it's difficult to avoid, but many movements started by a small group of smart people nevertheless go on to achieve a lot, that's also evidence that it's not important.

Hi Ben,

Thanks for the post. I think this is an important discussion. Though I'm also sympathetic to Nick's comment that a significant amount of extra self-reflection is not the most important thing to EA's success.

I just wanted to flag that I think there are attempts to deal with some of these issues, and explain why I think some of these issues are not a problem.

Philosophical difficulties

Effective altruism was founded by philosophers, so I think there's enough effort going into this, including population ethics. (See Nick's comment)

Poor cause choices

There... (read more)

3benkuhn
Hi Ben, Thanks for responding. I've responded to points below. The point of this argument wasn't that organizations aren't working on it. In fact the existence of this research strengthens my point, which was that people are donating now anyway despite the fact that it looks like we know very little now and the attitude towards giving now vs. later seems to be "well there's a good case for either one" rather than "we really need to figure this out because we may be pouring money down the drain", which is evidence that people are stopping thinking at the level of "doesn't obviously conflict with EA principles". Again, the issue isn't that nobody is trying to solve these, it's that most people are way more worried about the charity analysis issue than ancillary issues that are just as important. If our knowledge of e.g. cost-effectiveness of global health interventions was as limited as our knowledge elsewhere, would people be donating to global health charities? I doubt it. I've been following 80k and have not noticed this phenomenon. Can you give some examples? This is definitely not all we can do (unless you take a tautologically broad interpretation of "make an active effort to reach out"). For instance, if a substantial fraction of effective altruists were raging sexists, it would be wise to fix our group norms before going "hey women! there's this thing called effective altruism!" Even supposing it is all we can do, is there anything we're actually doing about it? The point of the critique was not to list easily avoidable problems, but to list bad problems. If decline in quality of people is inevitable, then we better find some solutions to the problems it brings (e.g. epistemic inertia), or the decline of EA is inevitable too.

Hi, I'd like to clarify that we prioritise people who are optimising around positive impact, not earning to give. If someone takes earning to give seriously, then we view that as a good indicator, but we speak to lots of people who aren't considering earning to give careers.

I started writing a response, but decided it would be better to summarise my general thoughts on degree choice and post them on our blog. So see our latest thoughts on how to pick a degree.

Insofar as this particular situation goes, I haven't thought about it much, so take this with a pi... (read more)

FYI: There has been a discussion on 80,000 Hours (started by me) about the value of this project and how to maximise it.

Hey John, we discuss this quite a bit in the interview (esp 1st and 2nd questions). Happy to take further more specific questions here though.

People might prefer this pair:

Peter Buffett and Zizek on why philanthropists do more harm than good and Will MacAskill's response on Qz.com

3ikrase
... Did that guy really just write that article without any concrete claim of harm due to capitalism? He seemed to make an oblique reference to outsourcing and implied that he thought that financial speculation leads to some form of poverty.

Sorry Luke, I didn't want to bother you so didn't ask, but I should have guessed you would have found this :)

Hi Luke,

This is certainly really important for 80k - it's on our list of strategic considerations to investigate.

We haven't looked into it in depth already, beyond knowledge of some relevant psychology literature (e.g. being primed by images of money has been found to make people more selfish in a couple of (probably dodgy) studies).

We've put a couple of measures in place which seem like they might help to mitigate the types of drift that don't involve updating on new information. First, making a public commitment to make the world a better place in an ef... (read more)

Givewell is effectively attempting to work out which charities most increase human welfare for dollar. So, a charity 'fails' if it becomes clearly less effective than the next best.

Heh almost, but the argument only seems to apply to xrisk. I don't see much reason to think EA movement building is the most effective way to fight global poverty.

There must be some, and it we'd certainly like to investigate which areas of industry are the most harmful. But in general, it's pretty hard for a career to result in the deaths of 600 people, which is a lower bound for what you could do with $1m (you could also fund SI for 1-2 years...). The most common harmful careers seem to inflict economic damage, and since the average dollar is spent on stuff which produces much less welfare than malaria nets or catastrophic risk research, you have to do a lot of it to outweigh your donations, like maybe 1-2 orders o... (read more)

That's a fairly common and very interesting question. Carl's got some thoughts on it, which we'll hopefully get written up. It's closely linked to two big and controversial issues: how good is economic growth and how good is technological progress? It's a case of weighing your contribution to that against the extra donations you can make.

Quite a few 80k members are interested in entrepreneurship. We'd definitely like to investigate these kinds of questions. But haven't found anyone yet.

that's a good one. It's going on the list. We have an upcoming series about happiness and career choice. The first (and one of the upcoming posts) are partially relevant. Drethelin's suggesting a good general strategy. If spending's the problem, you could also consider giving up on altruism in that domain, and making a difference in some other way. This is an example of macrooptimisation

Good question. We tend to take our charity evaluation from Givewell (though we've started our evaluation in some areas). So, we wouldn't be able to easily answer this. I don't think we've ever come across a charity which openly states its terms of surrender. What I can say is that the charities that tend to get recommended have a very focused method (e.g. distributing malaria nets) with a measurable outcome (less malaria), so it's pretty obvious if their failing, and that would cause them to lose funding.

0[anonymous]
I could be mistaken and I hope you will correct me of I am wrong. That sounds like equating a measurable outcome with success. Like a company that invested five hundred dollars, made a penny, and called itself profitable. A profit was made, but... no. One net distributed, one life saved, I will not say that's no good at any cost. But some bottom line of failure, of surrender, should be part of the evaluation. Charities that crow the most about 'raising awareness' or prayer are the worst offenders, confusing activity with achievement. They do more than nothing, but... no.

It's not all about donating. What's different about us is that we really try to weigh up different career options in terms of how much difference they make. We understand 'making a difference' to mean 'making good stuff happen that wouldn't have happened otherwise.' So, we wouldn't just recommend working for a traditional NGO if someone was going to do that job equally well if you didn't take it. Or if it didn't seem to be particularly cost-effective. In carrying this out, we take an evidence-based approach, paying attention to heuristics and biases. We'd ... (read more)

This seems to essentially be the question 'how can we best reduce xrisk?' We've got people ready to write about this in the fall, if not earlier. As a teaser, it seems like you can make a pretty good argument for EA movement building dominating most of the other approaches.

0Giles
Good thing I'm doing that then :-) On the other hand, my map says that people in the EA movement will say that EA movement building is the bestest thing, people in the SI will say that it's FAI research, etc. etc. Once you've filtered for strategically-minded people, you'd expect them all to already be doing whatever they thought was most effective (though out of the people I have in mind, not everyone is motivated by xrisk reduction, or not exclusively). Looking forward to what your team has to say on the matter though, definitely.

(Just making this more visible.)

Don't read this until you've already thought about your questions!

But here's what we're already working on:

  1. Which people can have the most impact in research careers? When does working in research trump funding research?

    1. How should we factor our own happiness into career decisions? What leads to job satisfaction and how realistic is it to take jobs in industries we're not passionate about?

    2. Among the 'effective altruist' and xrisk organisations, which have the greatest need for more funding or skills of various sorts?

  2. W

... (read more)

Apologies - I wasn't intending to hide the fact that I help to run 80k. If I were, hopefully I would have done a better job than using my real name. Point taken about it being a cross posting on the 80k blog, but I did think the content would be of special interest to LWers, and it hasn't been cross posted anywhere else.

2Xachariah
From what little I know regarding 80,000 hours it actually is worthwhile and one could make a good argument for being worthwhile to advertise or just discuss here. It's just that it strongly pattern matched to other advertising, like when LG came to a technology forum I frequented so they could tell us about their fantastic products for cheap! I wouldn't be averse to a different discussion talking about how it helps people achieve optimal philanthropy. Effective altruism is something I (and I'd wager a lot of people on this forum) find interesting. Edit: Just one quick question, why is 80,000 hours called 80,000 hours? It's a striking name, but a cursory read of the FAQ didn't explain.
2John_Maxwell
I don't think there's anything wrong with the fact you cross-posted it. Maybe your reception would have been a little better if you explicitly identified it as a cross-post and explained why you thought LWers would be interested, but from a consequentialist perspective, who cares?

This is the first post about 80k on LW by an 80k volunteer/staff member, and like Randaly says, the only two posts in the last 6 months to significantly feature 80k were about arguments for and against professional philanthropy.

Apologies for the 'collage of buzzwords' impression. I didn't include a detailed description of 80k and its purpose, like the THINK post, because I wasn't intending it to be an advert. Rather, I was intending it to be a survey. For this reason I also didn't include much detail about what our existing work is about, hoping not to bia... (read more)

5[anonymous]
Clarifying your mission statement is something you should do in the main body of your post, not something that should be somewhat buried in a comment thread.

Thanks for the write up Larks. We're currently looking for people to get involved with writing and researching similar articles. If you might be interested, email me: ben@80000hours.org