If Strong AI turns out to not be possible, what are our best expectations today as to why?
I'm thinking of trying myself at writing a sci-fi story, do you think exploring this idea has positive utility? I'm not sure myself: it looks like the idea that intelligence explosion is a possibility could use more public exposure, as it is.
I wanted to include a popular meme image macro here, but decided against it. I can't help it: every time I think "what if", I think of this guy.
Comparing megaflops performed by the silicon hardware with symbolic operation by the human brain is comparing apples and oranges. If you measure the number of additions and multiplications performed by the neurons (yes, less precise but more fault tolerant) you will arrive at a much higher number of flops. Think about mental addition more like editing a spreadsheet cell. That includes lots of operation related to updating, display, IO, ... and the addition itself is an insignicant part of it. Same if you juggle number which actually represent anything in your head. The representing is the hard part. Not the arithmetic itself.
You can see the teraflops of the human brain at work if you consider the visual cortex where it is easy to compare and amp the image transforms to well known operations (at least for the first processing stages).
OK like comparing apples and oranges. We wind up with apples AND oranges through similar mechanisms ni carbon and oxygen after 100s of millions of years of evolution, but we seriously consider we can't get there with design in silicon after less than 100 years of trying while watching the quality of our tools for getting there doubling every 5 years or so?
I'm not saying it HAS to happen. I'm just saying the smart bet is not against it happening.