ETA: I'll be adding things to the list that I think belong there.
I'm assuming a high level of credence in classic utilitarianism, and that AI-Xrisk is significant (e.g. roughly >10%), and timelines are not long (e.g. >50% ASI in <100years). ETA: For the purpose of this list, I don't care about questioning those assumptions.
Here's my current list (off the top of my head):
- not your comparitive advantage
- consider other Xrisks more threatening (top contenders: bio / nuclear)
- infinite ethics (and maybe other fundamental ethical questions, e.g. to do with moral uncertainty)
- S-risks
- simulation hypothesis
- ETA: AI has high moral value in expectation / by default
- ETA: low tractability (either at present or in general)
- ETA: Doomsday Argument as overwhelming evidence against futures with large number of minds
Also, does anyone want to say why they think none of these should change the picture? Or point to a good reference discussing this question? (etc.)
Yes, or it might just never be truly "solved". I agree that complexity theory seems fairly likely to hold (something like) a solution.
Not really. I don't think about infinite ethics much, which is probably one of the reasons it seems likely to change my mind. I expect that if I spent more time thinking about it, I would just become increasingly convinced that it isn't worth thinking about.
But it definitely troubles me that I haven't taken the time to really understand it, since I feel like I am in a similar epistemic state to ML researchers who dismiss Xrisk concerns and won't take the time to engage with them.
I guess there's maybe a disanalogy there, though, in that it seems like people who *have* thought more about infinite ethics tend to not be going around trying to convince others that it really actually matters a lot and should change what they work on or which causes they prioritize.
-------------------
I guess the main way I can imagine changing my views by studying infinite ethics would be to start believing that I should actually just aim to increase the chances of generating infinite utility (to the extent this is actually a mathematically coherent thing to try to do), which doesn't necessarily/obviously lead to prioritizing Xrisk, as far as I can see.
The possibility of such an update seems like it might make studying infinite ethics until I understand it better a higher priority than reducing AI-Xrisk.