SimonF comments on The Irrationality Game - Less Wrong

38 Post author: Will_Newsome 03 October 2010 02:43AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (910)

You are viewing a single comment's thread. Show more comments above.

Comment author: SimonF 05 October 2010 05:13:49PM *  3 points [-]

Yes!

Humans are "designed" to act intelligently in the physical world here on earth, we have complex adaptations for this environment. I don't think we are capable of acting effectively in "strange" environments, e.g. we are bad at predicting quantum mechanical systems, programming computers, etc.

Comment author: RomanDavis 06 October 2010 06:31:46AM 1 point [-]

But we can recursively self optimize ourselves for understanding mechanical systems or programming computers, not infinitely, of course, but with different hardware, it seems extremely plausible to smash through whatever ceiling a human might have.with the brute force of many calculated iterations of whatever humans are using,

And this is before the computer uses it's knowledge to reoptimize it's optimization process.

Comment author: SimonF 06 October 2010 09:24:17AM *  1 point [-]

I understand the concept of recursive self-optimization und I don't consider it to be very implausible.

Yet I am very sceptical, is there any evidence that algorithm-space has enough structure to allow for effective search to allow such an optimization?

I'm also not convinced that the human mind is good counterexample, e.g. I do not know how much I could improve on a the sourcecode of a simulation of my brain once the simulation itself runs effectively.

Comment author: wedrifid 06 October 2010 10:25:40AM *  2 points [-]

Yet I am very sceptical, is there any evidence that algorithm-space has enough structure to allow for effective search to allow such an optimization?

I count "algorithm-space is really really really big" as at least some form of evidence. ;)

Mind you by "is there any evidence?" you really mean "does the evidence lead to a high assigned probability?" That being the case "No Free Lunch" must also be considered. Even so NFL in this case mostly suggests that a general intelligence algorithm will be systematically bad at being generally stupid.

Considerations that lead me to believe that a general intelligence algorithm are likely include the observation that we can already see progressively more general problem solving processes in evidence just by looking at mammals. I also take more evidence from humanity than you do. Not because I think humans are good at general intelligence. We suck at it, it's something that has been tacked on to our brains relatively recently and it far less efficient than our more specific problem solving facilities. But the point is that we can do general intelligence of a form eventually if we dedicate ourselves to the problem.

Comment author: Risto_Saarelma 06 October 2010 06:12:10AM 1 point [-]

I don't think we are capable of acting effectively in "strange" environments, e.g. we are bad at predicting quantum mechanical systems, programming computers, etc.

You're putting 'effectively' here in place of 'intelligently' in the original assertion.

Comment author: SimonF 06 October 2010 08:56:01AM 0 points [-]

I understand "capable of behaving intelligently" to mean "capable of achieving complex goals in complex environments", do you disagree?

Comment author: Risto_Saarelma 06 October 2010 09:04:42AM *  0 points [-]

I don't disagree. Are you saying that humans aren't capable of achieving complex goals in the domains of quantum mechanics or computer programming?

Comment author: SimonF 06 October 2010 09:16:22AM 1 point [-]

This is of course a matter of degree, but basically yes!

Comment author: Risto_Saarelma 06 October 2010 09:34:37AM 0 points [-]

Can you give any idea what these complex goals would look like? Or conversely, describe some complex goals humans can achieve, which are fundamentally beyond an entity with a similar abstract reasoning capabilities as humans have, but lack some of humans' native capabilities for dealing more efficiently with certain types of problems?

The obvious examples are problems where a slow reaction time will lead to failure, but these don't seem to tell that much about the general complexity handling abilities of the agents.

Comment author: SimonF 06 October 2010 09:53:07AM *  2 points [-]

I'll try to give examples:

For computer programming: Given a simulation of a human brain, improve it so that the simulated human is significantly more intelligent.

For quantum mechanics: Design a high-temperature superconductor from scratch.

Are humans better than brute-force at a multi-dimensional version of chess where we can't use our visual cortex?

Comment author: wedrifid 06 October 2010 10:03:03AM 0 points [-]

Are humans better than brute-force at a multi-dimensional version of chess where we can't use our visual cortex?

We have a way to use brute force to achieve general optimisation goals? That seems like a good start to me!

Comment author: SimonF 06 October 2010 10:08:39AM 0 points [-]

Not a good start if we are facing exponential search-spaces! If brute-force would work, I imagine the AI-problem would be solved?

Comment author: wedrifid 06 October 2010 10:23:11AM 0 points [-]

Not a good start if we are facing exponential search-spaces!

Not particularly. :)

But it would constitute an in principle method of bootstrapping a more impressive kind of general intelligence. I actually didn't expect you would concede the ability to brute force 'general optimisation' - the ability to notice the brute forced solution is more than half the problem. From there it is just a matter of time to discover an algorithm that can do the search efficiently.

If brute-force would work, I imagine the AI-problem would be solved?

Not necessarily. Biases could easily have made humans worse than brute-force.