timtyler comments on How to measure optimisation power - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (30)
Yudkowsky's attempt seems to be of little practical use - as I explain in the comments there and here. It's a combination of a not-very-useful concept and misleading terminology.
Legg's AIQ seems to be a much more reasonable approach. Mahoney has previously done something similar.
I discussed a third group's attempt: http://lesswrong.com/lw/42t/aixistyle_iq_tests/
Thanks for the links, but I don't understand your objections - when measuring the optimisation power of a system or an AI, costs such as cycles, memory use, etc... are already implicitly included in Eliezer's measure. If the AI spends all it's time calculating without ever achieving anything, or if it has too little memory to complete any calculations, it will achieve an optimisation of zero.
Can you make sense of Shane Legg's objection, then?
One of my criticisms was this:
I don't see the point of calling something "optimisation power" - and then using it to award a brain-dead algorithm full marks.
I think your objection shows that you failed to read (or appreciate) this bit:
No "limited resources", just "preference ordering".
I would say that that the simple algorithm he describes has immense optimisation power. If there were a competitive situation, and other competent agents were trying to derail its goal, then its optimisation power drops close to zero. If your objection is that it's wrong to define a single "optimisation power" floating platonically above the agent, then I agree.
His very next paragraph is:
"outcome achieved". Hence the optimisation is measuring how effective the agent is at implementing its agenda. An agent that didn't have the ressources to think well or fast enough would score low, because it wouldn't implement anything.
The article talks about "preference ordering". There's no mention of how long the preference ordering takes to output. Resource constraints are missing from the whole article. It's optimisation power without consideration of resource limitation. Exhaustive search wins that contest - with the highest possible score.
Even if you factor in resource constraints (for example by imposing time and resource limits) - this is still a per-problem metric - while the term "optimisation power" suggests some more general capabilities.
"outcome achieved", "did you hit an outcome", "optimization processes produce surprises", "relative improbability of 'equally good or better' outcomes" - he's talking about the outcome produced (and then using the preference orderings to measure optimisation power given that that outcome was produced).
The time taken is not explicitly modeled, but is indirectly: exhaustive search only wins if the agent really has all the time in the world to implement its plans. An AI due to get smashed in a year if it doesn't produce anything will have an optimisation of zero if it uses exhaustive search.
"An optimisation of zero" ?!? <fx: shakes head, despairingly>
Are you objecting to the phrasing or to the point?
That was the terminology - though it isn't just the terminology that is busted here.
Frankly, I find it hard to believe that you seem to be taking this idea seriously.
I haven't decided whether the idea is good or bad yet - I haven't yet evaluated it properly.
But as far as I can tell, your objection to it is incorrect. A naive search program would have very low optimisation power by Eliezer's criteria - is there a flaw in my argument?
Related to Legg's work: Ben Goertzel's paper "Toward a Formal Characterization of Real-World General Intelligence "
Abstract: