"a safe AI design is possible and simple"
Excuse me, but of course it is. To believe that it's neither possible nor simple is to believe that human minds are so needlessly complex, viewed from the outside, that bottling up our model-forming and evaluation-forming processes for artificial processing is impossible or intractable.
The problem only looks hard because people are applying the wrong maps. They come at it with maps based on pure logic, microeconomic utility theory, and a priori moral philosophizing, when they really need the maps of statistical learning theory, computational cognitive science, and evaluative psychology.
Sometimes when things look really improbable, it's because you've got a very biased prior.
This thread is intended to provide a space for 'crazy' ideas. Ideas that spontaneously come to mind (and feel great), ideas you long wanted to tell but never found the place and time for and also for ideas you think should be obvious and simple - but nobody ever mentions them.
This thread itself is such an idea. Or rather the tangent of such an idea which I post below as a seed for this thread.
Rules for this thread:
If this should become a regular thread I suggest the following :