Kaj_Sotala comments on A forum for researchers to publicly discuss safety issues in advanced AI - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (73)
The argument is that AIXI and Bayes assume infinite computing power, and thus simplify the problem by allowing you to work on it without needing to consider computing power limitations. If you can't solve the easier form of the problem where you're allowed infinite computing power, you definitely can't solve the harder real-world version either, so you should start with the easier problem first.
But the difference between infinity and any finite value is infinity . Intelligence itself, or a substantial subset if it, is easy, given infinite resources, as AIXI shows. But that's been of no use in developing real world AI: tractable approximations to AIXI aren't powerful enough to be dangerous.
It would be embarrassing to MIRI if someone cobbled together AI smart enough to be dangerous, and came to the worlds experts on AI safety for some safety features, only to be told "sorry guys, we haven't got anything that's compatible with your system, because it's finite".
What's high value again?
It's arguably been useful in building models of AI safety. To quote Exploratory Engineering in AI:
I feel as though you're engaging in pedantry for pedantry's sake. The point is that if we can't even solve the simplified version of the problem, there's no way we're going to solve the hard version--effectively, it's saying that you have to crawl before you can walk. Your response was to point out that walking is more useful than crawling, which is really orthogonal to the problem here--the problem being, of course, the fact that we haven't even learned to crawl yet. AIXI and Bayes are useful in that solving AGI problems in the context provided can act as a "stepping stone" to larger and bigger problems. What are you suggesting as an alternative? That MIRI tackle the bigger problems immediately? That's not going to work.
You are still assuming that infinite systems count as simple versions of real world finite systems, but that is the assumption I am challenging: our best real world AIs aren't cut down AIXI systems, they are something different entirely, so there is no linear progression from crawling to walking in your terms,
That's not just an assumption; that's the null hypothesis, the default position. Sure, you can challenge it if you want, but if you do, you're going to have to provide some evidence why you think there's going to be a qualitative difference. And even if there is some such difference, it's still unlikely that we're going to get literally zero insights about the problem from studying AIXI. That's an extremely strong absolute claim, and absolute claims are almost always false. Ultimately, if you're going to criticize MIRI's approach, you need to provide some sort of plausible alternative, and right now, unfortunately, it doesn't seem like there are any. As far as I can tell, AIXI is the best way to bet.
I have already pointed out that the best AI systems currently existing are not cut down infinite systems.
Something doesn't have to be completely worthless to be sub optimal.