The outside view on this is that such predictions have been made since the start of A(G)I 50 or 60 years ago, and it's never panned out. What are the inside-view reasons to believe that this time it will? I've only looked through the table of contents of the Goertzel book -- is it more than a detailed survey of AGI work to date and speculations about the future, or are he and his co-workers really onto something?
My prediction / contrarian belief is that they are really onto something, with caveats (did you look at the second book? that's where their own design is outlined).
At the very highest level I think their CogPrime design is correct in the sense that it implements a human-level or better AGI that can solve many useful categories of real world problems, and learn / self-modify to solve those categories it is not well adapted to out of the box.
I do take issue with some of the specific choices they made in both fleshing out components and the current implementa...
As per a recent comment this thread is meant to voice contrarian opinions, that is anything this community tends not to agree with. Thus I ask you to post your contrarian views and upvote anything you do not agree with based on personal beliefs. Spam and trolling still needs to be downvoted.