But the bottleneck on organizational intelligence is either human intelligence or machine intelligence.
I disagree. I think there are lots of gains to intelligence that can happen at the point of human-computer interaction, or in the facilitation of human intelligence by machine intelligence, or vice versa.
For example, collaborative filtering technology. Or, internet message boards.
If we're very lucky, those computers will directly inherit the corporation's purported goal structure ("to enhance shareholder value"). Not that shareholder value is a good goal -- just that it's much less bad than a lot of the alternatives.
I'm curious why you think that an aritifical intelligence system built by Google would by likely to not meet the corporations goal structure (or some sub-goal).
In practice, AI programming tends to be about building expert systems for particular functions. It's difficult (and expensive) just to do that. So, building up an intelligent system that just goes crazy and kills people doesn't seem to be in, say, Google's interest.
That said, I'd be curious to follow the thread of whether maximizing shareholder value is a 'friendly' or 'mean' goal structure. Since that seems to be one of the predominant goal structures that it's likely for a superintelligence to have, it seems like that would be of particular interest. (Another one might be "win elections", since political parties are increasingly using machine intelligence to augment their performance.)
I disagree. I think there are lots of gains to intelligence that can happen at the point of human-computer interaction, or in the facilitation of human intelligence by machine intelligence, or vice versa.
For example, collaborative filtering technology. Or, internet message boards.
There are some gains, sure, but not lots and not, so far, recursive gains.
I'm curious why you think that an aritifical intelligence system built by Google would by likely to not meet the corporations goal structure (or some sub-goal).
I think that many AI systems presently...
If I understand the Singularitarian argument espoused by many members of this community (eg. Muehlhauser and Salamon), it goes something like this:
I'm in danger of getting into politics. Since I understand that political arguments are not welcome here, I will refer to these potentially unfriendly human intelligences broadly as organizations.
Smart organizations
By "organization" I mean something commonplace, with a twist. It's commonplace because I'm talking about a bunch of people coordinated somehow. The twist is that I want to include the information technology infrastructure used by that bunch of people within the extension of "organization".
Do organizations have intelligence? I think so. Here's some of the reasons why:
I talked with Mr. Muehlhauser about this specifically. I gather that at least at the time he thought human organizations should not be counted as intelligences (or at least as intelligences with the potential to become superintelligences) because they are not as versatile as human beings.
...and then...
I think that Muehlhauser is slightly mistaken on a few subtle but important points. I'm going to assert my position on them without much argument because I think they are fairly sensible, but if any reader disagrees I will try to defend them in the comments.
Mean organizations
* My preferred standard of rationality is communicative rationality, a Habermasian ideal of a rationality aimed at consensus through principled communication. As a consequence, when I believe a position to be rational, I believe that it is possible and desirable to convince other rational agents of it.