You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Luke_A_Somers comments on Superintelligence 9: The orthogonality of intelligence and goals - Less Wrong Discussion

8 Post author: KatjaGrace 11 November 2014 02:00AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (78)

You are viewing a single comment's thread. Show more comments above.

Comment author: Luke_A_Somers 11 November 2014 05:22:18PM 2 points [-]

How would they interact such that it's not simply adding over them, and they don't end up being predictably Dutch-bookable?

Comment author: IlyaShpitser 11 November 2014 06:00:02PM *  1 point [-]

In the same way people's minds do. They are inconsistent but will notice the setup very quickly and stop. (I don't find Dutch book arguments very convincing, really).

Comment author: Luke_A_Somers 11 November 2014 07:53:49PM 1 point [-]

Seems like a layer of inefficiency to have to resist temptation to run in circles rather than just want to go uphill.

Comment author: IlyaShpitser 12 November 2014 11:55:22AM *  1 point [-]

There are two issues:

(a) In what settings do you want an architecture like that, and

(b) Ethics dictate we don't just want to replace entities for the sake of efficiency even if they disagree. This leads to KILL ALL HUMANS. So, we might get an architecture like that due to how history played out. And then it's just a brute fact.

I am guessing (a) has to do with "robustness" (I am not prepared to mathematise what I mean yet, but I am thinking about it).


People that think about UDT/blackmail are thinking precisely about how to win in settings I am talking about.

Comment author: Luke_A_Somers 12 November 2014 01:52:15PM *  0 points [-]

Pick a side of this fence. Will AI resist running-in-circles trivially, or is its running in circles all that's saving us from KILL ALL HUMANS objectives like you say in part b?

If the latter, we are so utterly screwed.