Alright, I'll take you up on it:
Failure Mode I: The AI doesn't do anything useful, because there's no way of satisfying every contextual constraint.
Predicting your response: "That's not what I meant."
Failure Mode II: The AI weighs contextual constraints incorrectly and sterilizes all humans to satisfy the sort of person who believes in Voluntary Human Extinction.
Predicting your response: "It would (somehow) figure out the correct weighting for all the contextual constraints."
Failure Mode III: The AI weighs contextual constraints correctly (for a given value of "correctly") and sterilizes everybody of below-average intelligence or any genetic abnormalities that could impose costs on offspring, and in the process, sterilizes all humans.
Predicting your response: "It wouldn't do something so dumb."
Failure Mode IV: The AI weighs contextual constraints correctly and puts all people of minority ethical positions into mind-rewriting machines so that there's no disagreement anymore.
Predicting your response: "It wouldn't do something so dumb."
We could keep going, but the issue is that so far, you've defined -any- failure mode as "dumb"ness, and have argued that the AI wouldn't do anything so "dumb", because you've already defined that it is superintelligent.
I don't think you know what intelligence -is-. Intelligence does not confer immunity to "dumb" behaviors.
Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)
Since it is a steelman it isnt supposed to be what the paper is saying,
Are you maintaining, in contrast, that the maverick nanny is flatly impossible?
Sorry, I may have been confused about what you were trying to say because you were responding to someone else, and I hadn't come across the 'steelman' term before.
I withdraw 'parody' (sorry!) but ... it isn't quite what the logical structure of the paper was supposed to be.
It feels like you steelmanned it onto some other railroad track, so to speak.