In response to comment by ata on The Backup Plan
Comment author: Luke_A_Somers 14 October 2011 08:23:36PM 1 point [-]

Arguments are arguments. She's welcome to search for opposite arguments.

Comment author: ata 14 October 2011 09:04:22PM *  3 points [-]

A well-designed optimization agent probably isn't going to have some verbal argument processor separate from its general evidence processor. There's no rule that says she either has to accept or refute humans' arguments explicitly; as Professor Quirrell put it, "The import of an act lies not in what that act resembles on the surface, but in the states of mind which make that act more or less probable." If she knows the causal structure behind a human's argument, and she knows that it doesn't bottom out in the actual kind of epistemology that would be neccessary to entangle it with the information that it claims to provide, then she can just ignore it, and she'd be correct to do so. If she wants to kill all humans, then the bug is her utility function, not the part that fails to be fooled into changing her utility function by humans' clever arguments. That's a feature.

In response to comment by ata on The Backup Plan
Comment author: Luke_A_Somers 14 October 2011 03:35:33PM 0 points [-]

At least, at a first naive view. Hence a search for reasons that might overcome that argument.

Comment author: ata 14 October 2011 08:11:22PM 2 points [-]

But she won't be searching for reasons not to kill all humans, and she knows that any argument on our part is filtered by our desire not to be exterminated and therefore can't be trusted.

Comment author: Grognor 14 October 2011 04:05:32AM *  4 points [-]

A wise man proportions his belief to the evidence.

David Hume

Comment author: ata 14 October 2011 04:39:45AM *  9 points [-]

Roughly true, but downvoted for being basic (by LW standards) to the point of being an applause light. Good Rationality Quotes are ones we can learn from, not just agree with.

In response to The Backup Plan
Comment author: ata 14 October 2011 04:26:45AM *  8 points [-]

For starters, if she can prove she's friendly, then she can operate openly without causing nearly as much justified concern - which, in the early stages, will be helpful. Whatever her purposes are, if the restrictions of being friendly don't interfere as much as they help, that's a win.

If her current utility function is even a little bit different from Friendliness, and she expects she has the capacity to self-modify unto superintelligence, then I'd be very surprised if she actually modified her utility function to be closer to Friendliness; that would constitute a huge opportunity cost from her perspective. If she understands Friendliness well enough to know how to actually adjust closer to it, then she knows a whole lot about humans, probably well enough to give her much better options (persuasion, trickery, blackmail, hypnosis, etc.) than sacrificing a gigantic portion of her potential future utility.

Comment author: Alicorn 08 October 2011 12:55:26AM 7 points [-]

New batch of ponies up. I had to come up with names for all of these because nobody named their ponies ahead of time; I will change names accordingly if people don't like their assigned pony names.

Calamus, a unicorn pony of Jaime Astorga.

Equilibria, an earth pony of badger.

Pa'li, a My Little Na'vi of Eliza.

Pianissimo, a unicorn pony of Emil.

Pomme, an earth pony of MixedNuts.

Query, a unicorn pony of ata.

Seafoam, a hippocampus pony of my friend Gwen.

Strange Loop, a unicorn pony of Douglas Hofstadter.

Tempus, a unicorn pony of Leonhart.

Comment author: ata 09 October 2011 04:49:11AM 1 point [-]

Query, a unicorn pony of ata.

Awesome, thank you!!

Could I use that as my Facebook profile picture?

Comment author: PhilGoetz 29 September 2011 09:20:15PM *  6 points [-]

Down voted for not taking a few moments to explain how to answer these questions with a bit of research.

I already checked internic, and lesswrong.com is registered to enom, a domain name hosting company. I don't know how to find out who really owns it.

Comment author: ata 30 September 2011 03:46:51AM *  2 points [-]

lesswrong.{com,net,org} are registered to Trike, and I seem to recall that they manage its hosting and technical administrative aspects as well.

Comment author: ata 28 September 2011 03:07:16AM 7 points [-]

"No. You have just fallen prey to the meta-Dunning Kruger effect, where you talk about how awesome you are for recognizing how bad you are."

Horatio__Caine on reddit

Comment author: Eliezer_Yudkowsky 14 September 2011 06:26:53AM *  10 points [-]

I don't think people really understood what I was talking about in that thread. I would have to write a sequence about

  • the difference between first-order and second-order logic
  • why the Lowenheim-Skolem theorems show that you can talk about integers or reals in higher-order logic but not first-order logic
  • why third-order logic isn't qualitatively different from second-order logic in the same way that second-order logic is qualitatively above first-order logic
  • the generalization of Solomonoff induction to anthropic reasoning about agents resembling yourself who appear embedded in models of second-order theories, with more compact axiom sets being more probable a priori
  • how that addresses some points Wei Dai has made about hypercomputation not being conceivable to agents using Solomonoff induction on computable Cartesian environments, as well as formalizing some of the questions we argue about in anthropic theory
  • why seeing apparently infinite time and apparently continuous space suggests, to an agent using second-order anthropic induction, that we might be living within a model of axioms that imply infinity and continuity
  • why believing that things like a first uncountable ordinal can contain reality-fluid in the same way as the wavefunction, or even be uniquely specified by second-order axioms that pin down a single model up to isomorphism the way that second-order axioms can pin down integerness and realness, is something we have rather less evidence for, on the surface of things, than we have evidence favoring the physical existability of models of infinity and continuity, or the mathematical sensibility of talking about the integers or real numbers.
Comment author: ata 14 September 2011 10:15:57AM *  6 points [-]

I would like very very much to read that sequence. Might it be written at some point?

Comment author: Logos01 02 September 2011 08:14:37PM 2 points [-]

In short, I find this trope to be a fallacy. I'd expect an advanced civilisation to have a greater, not lesser, understanding of how intelligence works, its limitations, and failure modes in general.

Have you never looked at something someone does and asked yourself, "How can they be so stupid?"

It's not as though you literally cannot conceive of such limitations; just that you cannot empathize with them.

Comment author: ata 02 September 2011 08:53:01PM *  1 point [-]

It's anthropomorphism to assume that it would occur to advanced aliens to try to understand us empathetically rather than causally/technically in the first place, though.

Comment author: ata 26 August 2011 06:27:21PM *  2 points [-]

P(M) > 1

Typo?

If observing a dead cat causes the waveform to collapse such that the cat is dead, then P(D) = P(D) + P(M)(1-P(D)).  This is possible only if P(D) = 1.

Sorry if I'm missing something, but are you implying that the Copenhagen interpretation implies that the waveform collapse happens so as to retroactively make the cat dead if Schrödinger would have mistaken the cat for dead? Why would the sort of model that forms in Schrödinger's brain after the fact control what did in fact happen, even given the Copenhagen interpretation? (I didn't think it was quite that silly.)

View more: Prev | Next