I often talk to developers who prefer not destroying the world by accident (specifically by accelerating AGI risk), but neither them nor me can decide if specific companies qualify for this.
Could someone knowledgable help? A few short replies could probably change someone's career decisions
Can you help with future questions?
Please subscribe to the this comment. I'll reply to it only when there's a new open question.
Thank you!
Adding: Reply anonymously here
The worst thing you could possibly do is work for the capabilities section of existing AGI enterprises like Google Brain, DeepMind or OpenAI. This includes, obviously, the "AI alignment" companies that really just do capabilities research, and does not include the sections within these companies that do genuine alignment research. Dan Heydricks has an excellent sequence here on how not to fuck this up. Use your critical thinking and and ask simple questions to find out which position is which.
The second worst thing in terms of expected impact would be to work at or support pioneering ML research at a general company like Facebook's division, that isn't necessarily explicitly trying to engineer AGI but effectively involves a dayjob of burning the capabilities commons.
Below that would be to work on straightforward ML tooling that has generalist applications; things like working on frameworks (PyTorch, wandb.ai, etc.), computer hardware designed explicitly for ML, or for companies like Scale.
Somewhere deep below that is making money for or investing in the parent companies that pioneer these things (Facebook, Microsoft, Google). Depending on specifics you can lump in certain more general types of computer engineering work here.
After that though, I think if you just donate a reasonable fraction of your income to charity, or AI alignment enterprises, you're probably net positive. It's really not that complicated: if you're making or contributing to research that pushes the boundary of artificial intelligence, then... stop doing that.
This appears to be the crux of our disagreement. I do not think the venn diagram is clear at all. But if I had to guess, I think there is a large overlap between "make an AI that doesn't spew out racist garbage" and "make an AI that doesn't murder us all".