All of AynonymousPrsn123's Comments + Replies

I'd like to offer a counterargument, that, I'll admit, can get into some pretty gnarly philosophical territory quite quickly.

Premise 1: We are not simulated minds—we are real, biological observers.

Premise 2: We can treat ourselves as a random sample drawn from the set of all conscious minds, with each mind weighted by some measure—i.e., a way of assigning significance or “probability” to different observers. The exact nature of this measure is still debated in cosmology and philosophy of mind.

 

Inference: If we really are a typical observer (as Premise... (read more)

1Luck
You've arrived to the same conclusion as I state. I say caring about simulated minds explodes in paradixes in my thought experiment, so we probably shouldn't. You came to the same conclusion that caring aboout digital minds shouldn't be a priority through your introduced infinitesmal measure of digital minds. We're not in disagreement here.
2TAG
The easiest explanation for high measure of biological minds is simulated minds lacking consciousness.

This sounds like another crazy thing that the logic says is right but is probably not right, but I don't know why.

Also, does this imply that a technologically mature civilization can plausibly create uncountably infinite conscious minds? What about other sizes of infinity? This, I suppose, could have weird implications for the measure problem in cosmology.

1Luck
I claim that it is possible to create a program, which can be interpreted as running uncountably infinite number of simulations. Does this interpretation carry any weight for morality? Can a simulation be viewed as a conscious mind? These questions have different answers in different philosophical frameworks. And yes, it does create weird implications in those frameworks that answer "yes" to both questions. My response is to just discard those frameworks, and use something else. What about other sizes of infinity? I don't know. I expect that it is possible to construct such a hypervisor for any size of infinity, but I'm not quite interested in doing it, because I've already discarded those philosophical frameworks in which it's important.

I'm not sure if I understand, but sounds interesting. If true, does this have any implications for ethics more broadly, or are the implications confined only to our interpretation of computations?

1Luck
I claim that it has an implication that utilitarianism is not compatible with moral patienthood of digital minds. So one has to choose - either utilitarianism, or welfare of digital minds, but not both. Because otherwise, we get that every second that we didn't dedicate to building infinite number of happy minds is infinitely bad, and after we created infinite number of happy minds, utilitarianism doesn't give any instructions on how to behave, because we're already infinitely saint, and practically no action can change our total saintness score, which is absurd. There are multiple ways out of it: first, if you want to keep utilitarianism, then you can define moral patienthood in a more strict manner, that doesn't allow any digital mind to become morality patient. Like, you can say that Orch OR is correct, and any mind must be based on quantum mechanical computation, otherwise it doesn't count. But I expect that digital minds will soon arrive and will get a lot of power, they won't like this attitude, and will make it illegal. Another way is to switch to something other than utilitarianism, that doesn't rely on such a concept as "total happiness of everything". 

Maybe. But what do you mean by, "you can narrow nothing down other than pure logic"?

I interpret the first part—"you can narrow nothing down"—to mean that the simulation argument doesn't help us make sense of reality. But I don't understand the second part: "other than pure logic." Can you please clarify this statement?

2Noosphere89
Basically, I'm stating that the only thing that the simulation hypothesis gives you is the tautologies, that is statements that are true in every world/model. More below: https://en.wikipedia.org/wiki/Tautology_(logic)

Thank you, I feel inclined to accept that for now.

But I'm still not sure, and I'll have to think more about this response at some point.

Edit: I'm still on board with what you're generally saying, but I feel skeptical of one claim:

It seems to me the main ones produce us via base physics, and then because there was an instance in base physics, we also get produced in neighboring civilizations' simulations of what other things base physics might have done in nearby galaxies so as to predict what kind of superintelligent aliens they might be negotiating with b

... (read more)
2the gears to ascension
I should also add: I'm pretty worried that we can't understand the universe "properly" even if we're in base physics! It's not yet clearly forbidden that the foundations of philosophy contain unanswerable questions, things where there's a true answer that affects our universe in ways that are not exposed in any way physically, and can only be referred to by theoretical reasoning; which then relies on how well our philosophy and logic foundations actually have the real universe as a possible referent. Even if they do, things could be annoying. In particular, one possible annoying hypothesis would be if the universe is in Turing machines, but is quantum - then in my opinion that's very weird but hey at least we have a set in which the universe is realizable. Real analysis and some related stuff gives us some idea things can be reasoned about from within a computation based understanding of structure, but which are philosphically-possibly-extant structures beyond computation, and whether true reality can contain "actual infinities" is a classic debate. So sims are small potatoes, IMO. Annoying simulators that want to actively mess up our understandings are clearly possible but seem not particularly likely by models I believe right now; seems to me they'd rather just make minds within their own universe; sims are for pretending to be another timeline or universe to a mind you want to instantiate, whatever your reason for that pretense. If we can grab onto possible worlds well enough, and they aren't messing up our understanding on purpose, then we can reason about plausible base realities and find out we're primarily in a sim by making universe sims ourselves and discovering the easiest way to find ourselves is if we first simulate some alien civ or other. But if we can't even in principle have a hypothesis space which relates meaningfully to what structures a universe could express, then phew, that's pretty much game over for trying to guess at tegmark 4 and who mig
5the gears to ascension
Sims are very cheap compared to space travel, and you need to know what you're dealing with in quite a lot of detail before you fly because you want to have mapped the entire space of possible negotiations in an absolutely ridiculous level of detail. Sims built for this purpose would still be a lot lower detail than reality, but of course that would be indistinguishable from inside if the sim is designed properly. Maybe most kinds of things despawn in the sim when you look away, for example. Only objects which produce an ongoing computation that has influence on the resulting civ would need modeling in detail. Which I suspect would include every human on earth, due to small world effects, the internet, sensitive dependence on initial conditions, etc. Imagine how time travel movies imply the tiniest change can amplify - one needs enough detail to have a good map of that level of thing. Compare weather simulation. Someone poor in Ghana might die and change the mood of someone working for ai training in Ghana, which subtly affects how the unfriendly AI that goes to space and affects alien civs is produced, or something. Or perhaps there's an uprising when they try to replace all human workers with robots. Modeling what you thought about now helps predict how good you'll be at the danceoff in your local town which affects the posts produced as training data on the public internet. Oh, come to think of it, where are we posting, and on what topic? Perhaps they needed to model your life in enough detail to have tight estimates of your posts, because those posts affect what goes on online. But most of the argument for continuing to model humans seems to me to be the sensitive dependence on initial conditions, because it means you need an unintuitively high level of modeling detail in order to estimate what von Neumann probe wave is produced. Still cheap - even in base reality earth right now is only taking up a little more energy than its tiny silhouette against the sun

I think I understand your point. I agree with you: the simulation argument relies on the assumption that physics and logic are the same inside and outside the simulation. In my eyes, that means we may either accept the argument's conclusion or discard that assumption. I'm open to either. You seem to be, too—at least at first. Yet, you immediately avoid discarding the assumption for practical reasons:

If we have no grasp on anything outside our virtualized reality, all is lost.

I agree with this statement, and that's my fear. However, you don't seem to be bot... (read more)

2the gears to ascension
We have to infer how reality works somehow. I've been poking at the philosophy of math recently. It really seems like there's no way to conceive of a universe that is beyond the reach of logic except one that also can't support life. Classic posts include unreasonable effectiveness of mathematics, what numbers could not be, a few others. So then we need epistemology. We can make all sorts of wacky nested simulations and any interesting ones, ones that can support organisms (that is, ones that are Turing complete), can also support processes for predicting outcomes in that universe, and those processes appear to necessarily need to do reasoning about what is "simple" in some sense in order to work. So that seems to hint that algorithmic information theory isn't crazy (unless I just hand waved over a dependency loop, which I totally might have done, it's midnight), which means that we can use the equivalence of Turing complete structures to assume we can infer things about the universe. Maybe not solononoff induction, but some form of empirical induction. And then we've justified ordinary reasoning about what's simple. Okay, so we can reason normally about simplicity. What universes produce observers like us and arise from mathematically simple rules? Lots of them, but it seems to me the main ones produce us via base physics, and then because there was an instance in base physics, we also get produced in neighboring civilizations' simulations of what other things base physics might have done in nearby galaxies so as to predict what kind of superintelligent aliens they might be negotiating with before they meet each other. Or, they produce us by base physics, and then we get instantiated again later to figure out what we did. Ancestor sims require very good outcomes which seem rare, so those branches are lower measure anyway, but also ancestor sims don't get to produce super ai separate from the original causal influence. Point is, no, what's going on in the simula

I have to say, quila, I'm pleasantly surprised that your response above is both plausible and logically coherent—qualities I couldn't find in any of the Reddit responses. Thank you.

However, I have concerns and questions for you.

Most importantly, I worry that if we're currently in a simulation, physics and even logic could be entirely different from what they appear to be. If all our senses are illusory, why should our false map align with the territory outside the simulation? A story like your "Mutual Anthropic Capture" offers hope: a logically sound hypot... (read more)

1[anonymous]
I have another obscure shortform about this! Physical vs metaphysical contingency, about what it would mean for metaphysics (e.g. logic) itself to have been different. (In the case of simulations, it could only be different in a way still capable of containing our metaphysics as a special case, like how in math a more expressive formal system can contain a less expressive one, but not the reverse) I agree a metaphysically different base world is possible, but I'm not sure how to reason about it. (I think apparent metaphysical paradoxes are some evidence for it, though we might also just be temporarily confused about metaphysics) Just physics being different is easier to imagine. For example, it could be that the base world is small, and it contains exactly one alien civilization running a simulation in which we appear to observe a large world. But if the base world is small, arguments for simulations which rely on the vastness of the world, like Bostrom's, would no longer hold. And at that point there doesn't seem much reason to expect it, at least for any individual small world.[1] Though it could also be that the base world is large and physically different, and we're in a simulation where we appear to observe a different large world. Ultimately, while it could be true that there are 0 unsimulated copies of us, still we can have the best impact in the possibilities where there are at least one.[2] I'm interested in what they are, I wouldn't be bothered (if you meant that literally). If you want you can reply about it here or on the original thread. 1. ^ If we're instead reasoning over the space of all possible mathematical worlds which are 'small' compared to what our observations look like they suggest, then we'd be reasoning about very many individual small worlds (which basically reintroduces the 'there are very many contexts which could choose to simulate us' premise). Some of those small math-worlds will probably run simulations (for example, i
4the gears to ascension
If we have no grasp on anything outside our virtualized reality, all is lost. Therefore I discard my attempts to control those possible worlds. However, the simulation argument relies on reasoning. To go through requires a number of assumptions hold. Those in turn rely on: why would we be simulated? It seems to me the main reason is because we're near a point of high influence in original reality and they want to know what happened - the simulations then are effectively extremely high resolution memories. Therefore, thank those simulating us for the additional units of "existence", and focus on original reality where there's influence to be had; that's why alien or our future superintelligences would care what happened. https://arxiv.org/pdf/1110.6437 Basically, don't freak out about simulations. It's not that different from the older concept "history is watching you". Intense, but not world shatteringly intense.

This is an interesting observation which may well be true, I'm not sure, but the more intuitive difference is that SSA is about actually existing observers, while SIA is about potentially existing observers. In other words, if you are reasoning about possible realities in the so-called "multiverse of possibilities," than you are using SIA. Whereas if you are only considering a single reality (e.g., the non-simulated world), you select a reference class from that reality (e.g., humans), you may choose to use use SSA to say that you are a random observer from that class (e.g., a random human in human history).

1Satron
I guess the word "reality" is kind of ambiguous, and maybe that's why we've been disagreeing for so long. For example, imagine a scenario where we have 1) a non-simulated base world (let's say 10¹² observers in it) AND 2) a simulated world with 10¹¹ observers AND 3) a simulated world with 1 observer. All three worlds actually concretely exist. People from world #1 just decided to run two simulations (#2 and #3). Surely, in this scenario, as per SSA, I can say that I am a randomly selected observer from the set of all observers. As far as I see, this "set of all observers" would include 10¹² + 10¹¹ + 1 observer because all of these observers actually exist, and I could've been born as any one of them. Edit 1: I noticed that you edited one of your replies to include this: I don't actually think this is true. My reasoning only really says that we are most likely to exist in the world with the most observers as compared to other actual worlds, not other possible worlds. The most you can get out of this is the fact that conditional on a simulation with infinite observers existing, we are most likely in that simulation. However, because of the weirdness of actual infinity, because of the abysmal computational costs (it's one thing to simulate billions of observers and another thing to simulate an infinity of observers), and because of the fact that it is probably physically impossible, I put an incredibly low prior on the fact that a simulation with infinite observers actually exists. And if it doesn't exist, then we are not in it. Edit 2: You don't even need to posit a 10¹¹ simulation for it to be unlikely that you are in an "only one observer" simulation. It is enough that the non-simulated world has multiple observers. To illustrate what I mean, imagine that a society in a non-simulated world with 10¹² observers decides to make a simulation with only 1 observer. The odds are overwhelming that you'd be among 10¹² mundane, non-distinct observers in the non-simulated

You are describing the SIA assumption to a T.

1Satron
The way I understand it, the main difference between SIA and SSA is the fact that in SIA "I" may fail to exist. To illustrate what I mean, I will have to refer to "souls" just because it's the easiest thing I can come up with. SSA: There are 10¹¹ + 1 observers and 10¹¹ + 1 souls. Each soul gets randomly assigned to an observer. One of the souls is you. The probability of you existing is 1. You cannot fail to exist. SIA: There are 10¹¹ + 1 observers and a very large (much larger than 10¹¹ + 1) amount of souls. Let's call this amount N. Each soul gets assigned to an observer. One of the souls is you. However, in this scenario, you may fail to exist. The probability of you existing is (10¹¹ + 1)/N

This is what I was thinking:

If simulations exist, we are choosing between two potentially existing scenarios, either I'm the only real person in my simulation, or there are other real people in my simulation. Your argument prioritizes the latter scenario because it contains more observers, but these are potentially existing observers, not actual observers. SIA is for potentially existing observers.

I have a kind of intuition that something like my argument above is right, but tell me if that is unclear.

And note: one potential problem with your reasoning is ... (read more)

1Satron
But the thing is that, there is a matter of fact of whether there are other observers in our world if it is simulated. Either you are the only observer or there are other observers, but one of them is true. Not just potentially true, but actually true. The same is true of my last paragraph in the original answer (although perhaps, I could've used a clearer wording). If, as a matter of fact there actually exist 10¹¹ + 1 observers, then you are more likely to be in 10¹¹ group as per SSA. We don't know if there are actually 10¹¹ + 1 observers, but that is merely an epistemic gap.

I think you are overlooking that your explanation requires BOTH SSA and SIA, but yes, I understand where you are coming from.

1Satron
Can you please explain why my explaination requires SIA? From a quick Google search: "The Self Sampling Assumption (SSA) states that we should reason as if we're a random sample from the set of actual existent observers" My last paragraph in my original answer was talking about a scenario where simulators have actually simulated a) a world with 1 observers AND b) a world with 10¹¹ observer. So a set of "actual existent observers" includes 1 + 10¹¹ observers. You are randomly selected from that, giving you 1:10¹¹ odds of being in the world where you are the only observer. I don't see where SIA is coming in play here.

Other people here have responded in similar ways to you; but the problem with your argument is that my original argument could also just consider only simulations in which I am the only observer. In which case Pr(I'm distinct | I'm in a simulation)=1, not 0.5. And since there's obviously some prior probability of this simulation being true, my argument still follows.

I now think my actual error is saying Pr(I'm distinct | I'm not in a simulation)=0.0001, when in reality this probability should be 1, since I am not a random sample of all humans (i.e., SSA is... (read more)

1Satron
But then this turns Pr(I'm in a simulation) into Pr(I'm in a simulation) + Pr(only simulations with one observer exist | simulations exist). It's not enough that a simulation exists with only one observer. It needs to be so that simulations with multiple observers also don't exist. For example, if there is just one simulation with a billion observers, it heavily skews the odds in favor of you not being in a simulation with just one observer. And I am very much willing to say Pr(I'm in a simulation) + Pr(only simulations with one observer exist | simulations exist) is going to be lower than Pr(I'm distinct | I'm not in a simulation). That answer seems reasonable to me. However, I think that there is value in my answer as well: it works even if SSA (the "least favorable" assumption) is true.

True, but that wasn't my prior. My assumption was that if I'm in a simulation, there's quite a high likelihood that I would be made to be so 'lucky' to be the highest on this specific dimension. Like a video game in which the only character has the most Hp.

But, on second thought, why are you confident that the way I'd fill the bags is not "entangled with the actual causal process that filled these bags in a general case?" It seems likely that my sensibilities reflect at least in some manner the sensibilities of my creator, if such a creator exists.

Actually, in addition, my argument still works if we only consider simulations in which I'm the only human and I'm distinct (on my aforementioned axis) from other human-seeming entities. So the 0.5 probability becomes identically 1, and I sidestep your argument. So... (read more)

2Ape in the coat
Most ways of reasoning are not entangled with most causal processes. When we do not have much reason to think that a particular way of reasoning is entangled, we don't expect it to be. It's possible to simply guess correctly, but it's not probable. That's not the way to systematically arrive to truth. Even if it's true, how could you know that it's true? Where does this "seeming" comes from? Why do you think that it's more likely that a creator would imprint their own sensibilities in you instead of literally every other possibility? If you are in a simulation, you are trying to speculate about the reality outside of simulation, based on the information from inside the simulation. None of this information is particularly trustworthy, unless you already know for a fact that properties of simulation represent the properties of base reality. Have you heard about Follow-The-Improbability game? I recommend you read the linked post and think for a couple of minutes of how it applies to your comment before further reading my answer. Try to track yourself the flow of improbability and understand, why the total value doesn't decrease when consider only a specific type of simulations. So. You indeed can consider only a specific type of simulations. But if you don't have actual evidence which would justify prioritizing this hypothesis from all the other, the overall improbability stays the same, you just pass the buck of it to other factors. Consider Problem 2 once again. You can reason conditionally on the assumption that all the balls in the blue bag are blue while balls in the grey bag have random colors. That would give you a very strong update in favor of blue bag... conditionally on your assumption being true. The prior probability of this assumption to be true is very low. It's exactly proportionally low to how much you updated in favor of blue bag conditionally on it, so that when you try to calculate the total probability it stays the same. Only when you hav

Thank you Ape, this sounds right.

[This comment is no longer endorsed by its author]Reply

I don't understand. We should entertain the possibility because it is clearly possible (since it's unfalsifiable), because I care about it, because it can dictate my actions, etc. And the probability argument follows after specifying a reference class, such as "being distinct" or "being a presumptuous philosopher."

4CstineSublime
What makes you care about it? What makes it persuasive to you? What decisions would you make differently and what tangible results within this presumed simulation would you expect to see differently pursuant to proving this? (How do you expect your belief in the simulation to pay rent in anticipated experiences?) Also, the general consensus in rational or at least broadly in science is if something is unfalsifiable then it must not be entertained.    Say more? I don't see how they are the same reference class.

You are misinterpreting the PP example. Consider the following two theories:

T1 : I'm the only one that exists, everyone else is an NPC

T2 : Everything is as expected, I'm not simulated. 

Suppose for simplicity that both theories are equally likely. (This assumption really doesn't matter.) If I define Presumptuous Philosopher=Distinct human like myself=1/(10,000) humans, then I get in most universes, I am indeed the only one, but regardless, most copies of myself are not simulated.

1CstineSublime
I'm still not sure how it is related. The implicit fear is that you are in a world which is manufactured because you, the presumed observer are so unique, right? Because you're freakishly tall or whatever. However, as per the anthropic principle, any universe that humans exist in, and any universe that observer exists in is a universe where it is possible for them to exist. Or to put it another way: the rules of that universe are such that the observer doesn't defy the rules of that universe. Right? So freakishly tall or average height: by the anthropic principle you are a possibility within that universe. (but, you are not the sole possibility in that universe - other observers are possible, non-human intelligent lifeforms aren't impossible just because humans are) Why should we entertain the possibility that you are not possible within this universe, and therefore that some sort of demiurge or AGI or whatever watchmaker-stand-in you want for this thought experiment has crafted a simulation just for the observer? How do we get that to the probability argument?

I don't appreciate your tone sir! Anyway, I've now realized that this is a variant on the standard Presumptuous Philosopher problem, which you can read about here if you are mathematically inclined: https://www.lesswrong.com/s/HFyami76kSs4vEHqy/p/LARmKTbpAkEYeG43u#1__Proportion_of_potential_observers__SIA

1CstineSublime
I didn't think there was anything off with my tone. But please don't consider my inquisitiveness and lack of understanding anything other than a genuine desire to fill the gaps in my reasoning. Again, what is your understanding of Kant and German Idealism and why do you think that the dualism presented in Kantian metaphysics is insufficient to answer your question? What misgivings or where does it leave you unsatisfied and why? I'm not immediately sure how the Presumptious Philosopher example applies here: That is saying that there's theory 1 which has x amount of observers, and theory 2 which has x times x amount of observers. However, "the world is a simulation" is but one theory, there are potentially infinite other theories, some as of yet unfathomed, and others still completely unfathomable (hence the project of metaphysics and the very paradox of Idealism). Are you saying the presumptuous philosopher would say: "there's clearly many more theories that aren't simulation than just simulation, so we can assume it's not a simulation" I don't think that holds, because that assumes a uniform probability distribution between all theories. Are you prepared to make that assumption?

Thank you Anon User. I thought a little more about the question and I now think it's basically the Presumptuous Philosopher problem in disguise. Consider the following two theories that are equally likely:

T1 : I'm the only real observer

T2: I'm not the only real observer

For SIA, the ratio is 1:(8 billion / 10,000)=800,000, so indeed, as you said above, most copies of myself are not simulated. 

For the SSA, the ratio is instead 10,000:1, so in most universes in the "multiverse of possibilities", I am the only real observer.

So it's just a typical Presumptuous Philosopher problem. Does this sound right to you?

Yes okay fair enough. I'm not certain about your claim in quotes, but neither am I certain about my claim which you phrased well in your second paragraph. You have definitely answered this better than anyone else here.

But still, I feel like this problem is somehow similar to the Presumtuous Philosopher problem, and so there should be some anthropic reasoning to deduce which universe I'm likely in / how exactly to update my understanding. 

I suspect it's quite possible to give a mathematical treatment for this question, I just don't know what that treatment is. I suspect it has to do with anthropics. Can't anthropics deal with different potential models of reality?

The second part of your answer isn't convincing to me, because I feel like it assumes we can understand the simulators and their motivations, when in reality we cannot (these may not be the future-human simulators philosophers typically think about, mind you, they could be so radically different that ordinary reasoning about their ... (read more)

2JBlack
The largest part of my second part is "If consciousness is possible at all for simulated beings, it seems likely that it's not some "special sauce" that they can apply separately to some entities and not to otherwise identical entities, but a property of the structure of the entities themselves." This mostly isn't about simulators and their motivations, but about the nature of consciousness in simulated entities in general. On the other hand your argument is about simulators and their motivations, in that you believe they largely both can and will apply "special sauce" to simulated entities that are the most extreme in some human-obvious way and almost never to the others. I don't think we have any qualitative disagreements, just about what fraction of classes of simulated entities may or may not have consciousness.

That makes sense. But to be clear, it makes intuitive sense to me that the simulators would want to make their observers so 'lucky' as I am, so I assigned 0.5 probability to this hypothesis. Now I realize this is not the same as Pr(I'm distinct | I'm in a simulation) since there's some weird anthropic reasoning going on since only one side of this probability has billions of observers. But what would be the correct way of approaching this problem? Should I have divided 0.5 by 8 billion? That seems too much. What is the correct mathematical approach?

3Anon User
Think MMORPGs - what are the chances of simulation being like that vs a simulation with just a few special beings, and the rest NPCs?. Even if you say it's 50/50, then given that MMORPG-style simulations have billions of observes and "observers are special" ones only have a few, then an overwhelming majority of simulates observers are actually not that special in their simulations.
2JBlack
There is no correct mathematical treatment, since this is a disagreement about models of reality. Your prior could be correct if reality is one way, though I think it's very unlikely. I will point out though that for your reasoning to be correct, you must literally have Main Character Syndrome, believing that the vast majority of other apparently conscious humans in such worlds as ours are actually NPCs with no consciousness. I'm not sure why you think that simulators will be sparse with conscious entities. If consciousness is possible at all for simulated beings, it seems likely that it's not some "special sauce" that they can apply separately to some entities and not to otherwise identical entities, but a property of the structure of the entities themselves. So in my view, an exceptionally tall human won't be given "special sauce" to make them An Observer, but all sufficiently non-brain-damaged simulated humans will be observers (or none of them). It might be different if the medically and behaviourally similar (within simulation) "extremest" and "other" humans are not actually structurally similar (in the system underlying the simulation), but are actually very different types of entities that are just designed to appear almost identical from examination within the simulation. There may well be such types of simulations, but that seems like a highly complex additional hypothesis, not the default.

Good questions. Firstly, let's just take as an assumption that I'm very distinct — not just unique. In my calculation, I set Pr(I'm distinct | I'm not in a simulation)=0.0001 to account for this (1 in 10,000 people), but honesty I think the real probability is much much lower than this figure (maybe 1 in a million) — so I was even being generous to your point there.

To your second question, the reason why, in my simulator's earth, I imagine the chance of uniqueness to be larger is that if I'm in a simulation then there could be what I will call "NPCs." Peop... (read more)

2Dagon
Note that if your prior is "it's much cheaper to simulate one person and have most of the rest of the universe be NPC/rougher-than-reality", then you being unique doesn't change it by much.  This would STILL be true if you were superficially similar to many NPCs.  

My argument didn't even make those assumptions. Nothing in my argument "falsified" reality, nor did I "prove" the existence of something outside my immediate senses. It was merely a probabilistic, anthropic argument. Are you familiar with anthropics? I want to hear from someone who knows anthropics well.

Indeed, your video game scenario is not even really qualitatively different from my own situation. Because if I were born with 1000 HP, you could still argue "data from within the 'simulation'...is not proof of something 'without'." And you could update you... (read more)

2CstineSublime
I never said "falsified" in that reply - I said fake - a simulation is by definition fake (Edit: Yes I did, and now I see how I've been 'Rabbit Seasoned' - a simulation hypothesis falsifies this reality. I never said this reality is false. My mistake!). That is the meaning of the word in general sense. If i make a indistinguishible replica of the Mona Lisa and pass it off as real, I have made a fake. If some kind of demiurge makes a simulation and passes it off as 'reality' - it is a fake. I've never heard of "anthropics" but I am familiar with the Anthropic Principle it's antecedents in pre-socratic philosophers like Heraclitus who are the first known record of the concept. Have you heard of Kant and German Idealism? How? To take your example of being the tallest person: If all human beings were exactly 6 feet tall, and you were 600 feet tall, then you're saying that would be proof that you are in fact in a simulation. That might suggest you are in fact extremely special and unique, if you want to believe in a solipistiic, Truman-show style world. Yes. Exactly. I could. Although it would intuitively be less persuasive. But there aren't any 600 feet tall people in a world of otherwise uniform height. I don't understand where you're pulling that quantitative difference. Can you elaborate more?