moridinamael comments on [Stub] Ontological crisis = out of environment behaviour? - Less Wrong Discussion
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (4)
That's a good observation. These situations are analogous in some ways:
I wonder if it could be possible to permanently anchor an agent to its original ontology. To specify that the ontology with which it initialized is the perspective that it is required to use when evaluating its utility function. The agent is permitted the build whatever models it needs to build, but it's only allowed to assign value using the primitive concepts. So:
(Or perhaps the agent is allowed to re-define its value system within the new, more accurate ontology, but it isn't allowed to do so until it comes up with a sufficiently good mapping that the prior ontology and the new ontology give the same answers on questions of value. And if it can never accomplish that, then it simply never uses the new mapping.)
On the one hand, we do ultimately want agents who can grow to understand everything. And we don't want them to stop caring about humans the moment they stop seeing "humans" and start seeing "quivering blobs of cellular machinery".
Another thought is that AIs won't necessarily be as preoccupied with what is "real" as humans sometimes are. Just because an agent realizes that its whole world model is "not sufficiently fundamental" doesn't immediately imply that it discards the prior model wholesale.
That actually seems like what humans do. Human confusions about moral philosophy even seem quite like an ontological crisis in an AI.