You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

rcadey comments on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities - Less Wrong Discussion

25 Post author: KatjaGrace 16 September 2014 01:00AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (232)

You are viewing a single comment's thread. Show more comments above.

Comment author: rcadey 21 September 2014 08:19:55PM 1 point [-]

"How much smarter than a human could a thing be?" - almost infinitely if it consumed all of the known universe

"How about the same question, but using no more energy than a human?" -again the same answer - assuming we assume intelligence to be computable, then no energy is required (http://www.research.ibm.com/journal/rd/176/ibmrd1706G.pdf) if we use reversible computing. Once we have an AI that is smarter than a human then it would soon design something that is smarter but more efficient (energy wise)?

Comment author: leplen 22 September 2014 05:25:33PM 3 points [-]

This link appears not to work, and it should be noted that "zero-energy" computing is at this point predominantly a thought experiment. A "zero-energy" computer would have to operate in the adiabatic limit, which is the technical term for "infinitely slowly."

Comment author: KatjaGrace 22 September 2014 03:36:25AM 1 point [-]

Anders Sandberg has some thoughts on physical limits to computation which might be relevant, but I admit I haven't read them yet: http://www.jetpress.org/volume5/Brains2.pdf