SoullessAutomaton comments on The Strangest Thing An AI Could Tell You - Less Wrong

81 Post author: Eliezer_Yudkowsky 15 July 2009 02:27AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (574)

You are viewing a single comment's thread. Show more comments above.

Comment author: SoullessAutomaton 18 July 2009 06:12:09PM *  6 points [-]

I suspect that by Eliezer's standards, "beat it" would be defined as "they would be forced to shut down enough subsystems to no longer have any semblance of a functioning intelligence".

Of course, I doubt that this is possible on the human cognitive architecture, but it would be a nice property of a fault-tolerant AI.

Comment author: JulianMorrison 18 July 2009 08:19:54PM 4 points [-]

Highly unlikely in a human. I don't know, but I'd guess that self-checking is one subsystem. Lose that, and the plainest contradictions pass as uninteresting.

On the other hand, the engineering challenge catches my interest. Might there be any way to train other parts of your mind, parts that normally don't do checking, to sync up if everything is working OK or intervene if the normal part is out of action? Get the right brain in on the act, perhaps. That might give you something like Dune "truth sense" but turned inward. It would certainly feel very different from normal reason.

Conscious-mind rationality is good - might unconscious-mind rationality be better? You could self-monitor even on autopilot.

How many subsystems can be made rational?

Comment author: SoullessAutomaton 18 July 2009 11:11:50PM *  3 points [-]

There are many reasons to expect that the non-conscious part of the mind is largely arational, in the LW senses of rationality. My impression is that it seems to operate mostly on trained responses and associative connections/pattern matching, mediated by emotional responses. In practice this means it can often actually be more rational in certain ways than the conscious mind, because it seems to be better at collecting and correlating information, cf. people who have a non-rational aversion to certain foods for reasons they don't consciously understand, then years later discover they're actually allergic to it.

I expect the better approach would be to deliberately train the non-conscious mind to use associations and heuristics derived by the rational conscious mind, and I mean "train" in the sense of "training a dog".

Any sort of high-level self-monitoring is probably beyond its capabilities, though perhaps recognizing warning signs and alerting the conscious mind would work. Some sort of "panic on unexpected input" type heuristic, I guess.