"Wait for the opponents to catch up a little, stage some nice space battles... close the game window at some point. What if our universe is like that?"
Wow, what a nice elegant Fermi paradox solution:)
"Wait for the opponents to catch up a little, stage some nice space battles... close the game window at some point. What if our universe is like that?"
Wow, what a nice elegant Fermi paradox solution:)
"because you don't actually want to wake up in an incomprehensible world"
Is not it what all people do each morning anyway?
"Errr.... luzr, why would I assume that the majority of GAIs that we create will think in a way I define as 'right'?"
It is not about what YOU define as right.
Anyway, considering that Eliezer is existing self-aware sentient GI agent, with obviously high intelligence and he is able to ask such questions despite his original biological programming makes me suppose that some other powerful strong sentient self-aware GI should reach the same point. I also *believe* that more general intelligence make GI converge to such "right thinking".
What makes me worry most is building GAI as non-sentient utility maximizer. OTOH, I *believe* that 'non-sentient utility maximizer' is mutually exclusive with 'learning' strong AGI system - in other words, any system capable of learning and exceeding human inteligence must outgrow non-sentience and utility maximizing. I migh be wrong, of course. But the fact that universe is not paperclipped yet makes me hope...
Phil:
"If we are so unfortunate as to live in a universe in which knowledge is finite, then conflict may serve as a substitute for ignorance in providing us a challenge."
This is inconsistent. What would conflict really do is to provide new information to process ("knowledge").
I guess I can agree with the rest of post. What IMO is worth pointing out that the most pleasures, hormones and insticts excluded, are about processing 'interesting' infromations.
I guess, somewhere deep in all sentient beings, "interesting informations" are the ultimate joy. This has dire implications for any strong AGI.
I mean, the real pleasure for AGI has to be about acquiring new information patterns. Would not it be a little bit stupid to paperclip solar system in that case?
"But considering an unlimited amount of ice cream forced me to confront the issue of what to do with any of it."
"If you invoke the unlimited power to create a quadrillion people, then why not a quadrillion?"
"Say, the programming team has cracked the "hard problem of conscious experience" in sufficient depth that they can guarantee that the AI they create is not sentient - not a repository of pleasure, or pain, or subjective experience, or any interest-in-self - and hence, the AI is only a means to an end, and not an end in itself."
"What is individually a life worth living?"
Really, is not the ultimate answer to the whole FAI issue encoded there?
IMO, the most important thing about AI is to make sure IT IS SENTIENT. Then, with very high probability, it has to consider the very same questions suggested here.
(And to make sure it does, make more of them and make them diverse. Majority will likely "think right" and supress the rest.)
"real world is deterministic on the most fundamental level"
Is it?
http://en.wikipedia.org/wiki/Determinism#Determinism.2C_quantum_mechanics.2C_and_classical_physics
Tim:
Well, as off-topic recourse, I see only cited some engineering problems in your "Against Cyborgs" essay as contraargument. Anyway, let me to say that in my book:
"miniaturizing and refining cell phones, video displays, and other devices that feed our senses. A global-positioning-system brain implant to guide you to your destination would seem seductive only if you could not buy a miniature ear speaker to whisper you directions. Not only could you stow away this and other such gear when you wanted a break, you could upgrade without brain surgery."
is pretty much equivalent of what I had in mind with cyborging. Brain surgery is not the point. I guess it is even today pretty obvious that to read thoughts, you will not need any surgery at all. And if information is fed back into my glasses, that is OK with.
Still, the ability to just "think" the code (yep, I am a programmer), then see the whole procedure displayed before my eyes already refactored and tested (via weak AI augmentation) sound like nice productivity booster. In fact, I believe that if thinking code is easy, one, with the help of some nice programming language, could learn to use coding to solve much more problems in normal live situations, gradually building personal library of routines..... :)
Eliezer:
"Will, your example, good or bad, is universal over singletons, nonsingletons, any way of doing things anywhere."
I guess there is significant difference - for singleton, each mistake can be fatal (and not only for it).
I believe that this is the real part I dislike about the idea, except the part where singleton either cannot evolve or cannot stay singleton (because of speed of light vs locality issue).
Eliezer:
"Tim, your page doesn't say anything about FOR loops or self-optimizing compilers not being able to go a second round, which is the part you got from me and then thought you had invented."
Well, it certainly does:
"Today, machines already do a lot of programming. They perform refactoring tasks which would once have been delegated to junior programmers. They compile high-level languages into machine code, and generate programs from task specifications. They also also automatically detect programming errors, and automatically test existing programs."
I guess your claim is only a misunderstaning caused by not understaning CS terminology.
Find a new way how to optimize loops is application of automated refactoring and automated testing and benchmarking.
Abigail:
"The "Culture" sequence of novels by Iain M. Banks suggests how people might cope with machines doing all the work."
Exactly, I think Culture is highly relevant to most topics discussed here. Obviously, it is just a fictional utopia, but I believe it gives plausible answer to "unlimited power future".
For the reference: http://en.wikipedia.org/wiki/The_Culture