since security is disjunctive and if AIs have a just 1% chance of talking themselves out, now you have a insecurity floor of 1%
If a particular situation poses a 1% risk if it comes up, one can lower the total risk by making that situation less likely: other security measures could stop the chance from being tried, e.g. a human facing the death penalty for a failed escape from a prison and a 1% success chance would not obviously try. Even an autonomous AI with interests in conflict with humanity to some degree might be designed without such a risk-loving decision algorithm as to try an improbable escape attempt in the face of punishment for failure or reward for non-attempt.
If a particular situation poses a 1% risk if it comes up, one can lower the total risk by making that situation less likely
You only do that by changing the problem; a different problem will have different security properties. The new risk will still be a floor, the disjunctive problem hasn't gone away.
a human facing the death penalty for a failed escape from a prison and a 1% success chance would not obviously try.
Many do try if the circumstances are bad enough, and the death penalty for a failed escape is common throughout history and in totalitari...
Update 2013-09-05.
I have since played two more AI box experiments after this one, winning both.
Update 2013-12-30:
I have lost two more AI box experiments, and won two more. Current Record is 3 Wins, 3 Losses.