Houshalter comments on The AI That Pretends To Be Human - Less Wrong Discussion
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (69)
Yes, though you could condition it on a good human, or even yourself if you like.
The point is the human isn't superintelligent. Hitler wouldn't be able to hack a human brain, or invent an AGI. So an AI mimicking Hitler wouldn't have those abilities either, and would have to solve the problem the hard way.
But with Hitler's mass hysteria skills, the AI could lead others to very crazy ends.
Yes, you do not need to be superintelligent to kill everybody. The later is simplier task.
He was still just a man. He didn't have super powers. He was a skilled orator and leader, with the right ideology, at the right time and place in history. But he wouldn't be able to break out of an AI box. He wouldn't be able to design a self replicating machine to destroy the world. A Hitler-in-a-box could be safely contained.
Isn't the skill of oratory precisely the skill that gets you unboxed?
Enough skill in oratory (or something closely related) gets you unboxed. The question is how plausible it is that a superintelligent AI would have enough. (A related question is whether there's such a thing as enough. There might not be, just as there's no such thing as enough kinetic energy to let you escape from inside a black hole's horizon, but the reported results of AI-Box games[1] suggest -- though they certainly don't prove -- that there is.)
[1] The term "experiments" seems a little too highfalutin'.
[EDITED to add: I take it Houshalter is saying that Hitler's known oratorical skills aren't enough to convince him that H. would have won an AI-Box game, playing as the AI. I am inclined to agree. Hitler was very good at stirring up a crowd, but it's not clear how that generalizes to persuading an intelligent and skeptical individual.]
Well for one, the human isn't in a box trying to get out. So an AI mimicking a human isn't going to say weird things like "let me out of this box!" This method is equivalent to writing Hitler a letter asking him a question, and him sending you an answer. That doesn't seem dangerous at all.
Second, I really don't believe Hitler could escape from a box. The AI box experiments suggest a human can do it, but the scenario is very different than a real AI box situation. E.g. no back and forth with the gatekeeper, and the gatekeeper doesn't have to sit there for 2 hours and listen to the AI emotionally abuse him. If Hitler says something mean, the gatekeeper can just turn him off or walk away.
He could order such machine without leaving a box. He could grab power and start WW3 without leaving the box. The ability to leave the box an make own copies is not necessary condition for dangerous AI.
If Hitler appeared today, even without being stuck inside a box, he would not be able to start WW3 or grab power. Hitler is not that powerful. And being stuck in a box would be a significant constraint on his power.
He was powerful enough to grab power in Germany in 13 years, even being inside the box of his cranial. If he will be a little bit better now (but not a superintelligence), just ideal psychopath, he (or his equivalent) could be even more effective.
We can't judge based on behaviour that some one is superintelligent or not.
Yes we can. Superintelligences have abilities that normal intelligences do not.
Imagine a game of chess. A good AI will make vastly different moves than a bad AI or a human. More skilled players would easily be detectable. They would make very different moves.
But in some games it is better to look more stupid in the begging. Like poker, espionage and AI box experiment.
An AI that is indistinguishable from a human (to an even greater superintelligent AI) is not dangerous, because humans are not dangerous. Just like a chess master that is indistinguishable from a regular player wouldn't win many games.
It may be indistinguishable until it gets our of the building. Recent movie Ex Machine had such plot.
The AI doesn't want to escape from the building. It's utility function is basically to mimic humans. It's a terminal value, not a subgoal.
But most humans would like to escape from any confinement
I wonder if this is true in general. Have you read a good discussion on detecting superintelligence?
Can't remember ad hoc; but if superintelligence is able to do anything, it could easily pretend to be more stupid than it is. May be only "super super intelligence" could solve him. But it also may depend of the length of the conversation. If it say just Yes or No once, we can't decide, if it say longer sequences we could conclude something, but for any length of sentences is maximum level of intelligence that could be concluded from it.
The opportunities for detecting superintelligence would definitely be rarer if the superintelligence is actively trying to conceal the status.
What about in the case where there is no attempted concealment? Or even weaker, where the AI voluntary submits to arbitrary tests. What tests would we use?
Presumably we would have a successful model of human intelligence by that point. It's interesting to think about what dimensions of intelligence to measure. Number of variables simultaneously optimized? Optimization speed? Ability to apply nonlinear relationships? Search speed in a high dimensional, nonlinear solution space? I guess it is more the ability to generate appropriate search spaces in the first place. Something much simpler?
Probably winning humans in ALL known domains, including philosophy, poetry, love, power.
Although we use limited proxies (e.g., IQ test questions) to estimate human intelligence.
limited proxies - yes, well said. also I would add solving problems which humans were unable to solve for long: aging, cancer, star travel, word peace, resurrection of dead.
I mean, the ability to estimate the abilities of superintelligences appears to be an aspect of reliable Vingean reflection.
Or we could ask these AI to create the scale. We could use also its size to estimate power, like number on neurons. But real test needs to be powerful as well as universal optimization problem, something like ability to crack complex encryption or Go game.
I created a list of steps or milestones of the future AI and we could use similar list to estimate level of current super AI.
5a Self-improving threshold. AI becomes able to self-improve independently and quicker than all humanity 5b Consciousness and qualia threshold. AI is able not only pass Turing test in all cases, but has experiences and has understanding why and what it is.
Mankind-level AI. AI possessing intelligence comparable to that of the whole mankind. 2040-2100
AI with the intelligence 10 – 100 times bigger than that of the whole mankind. It will be able to solve problems of aging, cancer, solar system exploration, nanorobots building, and radical improvement of life of all people. 2050-2100
Jupiter brain – huge AI using the entire planet’s mass for calculations. It can reconstruct dead people, create complex simulations of the past, and dispatch von Neumann probes. 2100-3000
Galactic kardashov level 3 AI. Several million years from now.
All-Universe AI. Several billion years from now