A self-modifying AI is built to serve humanity. The builders know, of course, that this is much riskier than it seems, because its success would render their own observations extremely rare. To solve the problem, they direct the AI to create billions of simulated humanities in the hope that this will serve as a Schelling point to them, and make their own universe almost certainly simulated.
Plausible?
OK, suppose I come to you while you're sleeping, and add/remove a single neuron. Will you wake up in your model? Yes, because while you're naturally sleeping, much more neurons change. Now imagine that I alter your entire brain. Now, the answer seems to be no. Therefore, there must be some minimal change to your brain to ensure that a different person will wake up (i.e. with different consciousness/qualia). This seems strange.
You don't assume that the person who wakes up always has different consciousness with the person who fell asleep, do you?
It would be the same computer, but different working session. Anyway, I doubt such analogies are precise and allow for reliable reasoning.
Alter how? Do I still have memories of this argument? Do I share any memories with my past self? If I share all memories, then probably it's still me. If all have gone, then most likely not. (Identifying self with memories has its own problems, but let's gloss over them for now.) So I'm going to interpret your "remove a neuron" as "remove a memory", and then your question becomes "how many memories can I lose and still be me"? That's a difficult ques... (read more)