morganism comments on [LINK] Concrete problems in AI safety - Less Wrong

15 Post author: Stuart_Armstrong 05 July 2016 09:33PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (6)

You are viewing a single comment's thread.

Comment author: morganism 07 July 2016 08:27:10PM 0 points [-]

these folks say that you won't be able to sandbox a AGI, due to the nature of computing itself.

Assuming that a superintelligence will contain a program that includes all the programs that can be executed by a universal Turing machine on input potentially as complex as the state of the world, strict containment requires simulations of such a program, something theoretically (and practically) infeasible.

http://arxiv.org/abs/1607.00913v1

But perhaps we could fool it, by poisoning some crucial databases it uses in subtle ways.

DeepFool: a simple and accurate method to fool deep neural networks

http://arxiv.org/abs/1511.04599v3

Comment author: Stuart_Armstrong 09 July 2016 08:13:59AM 1 point [-]

strict containment requires simulations of such a program, something theoretically (and practically) infeasible.

Sandboxing just requires that you be sure that the sandboxed entity can't send bits outside the system (except on some defined channel, maybe), which is perfectly feasible.

Comment author: Viliam 11 July 2016 02:31:59PM 1 point [-]

perfectly feasible

Citation needed.

Comment author: Stuart_Armstrong 11 July 2016 05:55:18PM 1 point [-]

In software, it's trivial: create a subroutine with only a very specific output, include the entity inside it. Some precautions are then needed to prevent the entity from hacking out through hardware weaknesses, but that should be doable (using isolation in faraday cage if needed).