One basic problem with the AGI Risk Survey is that humans consistently represent AI as a collective function with recognizable features of human cognition. The problem with modeling AGI based upon human cognitive features is that our body and environment stabilize the evolving network systems. Our hormones, limb nervous systems, brain lobes, nutritional sources & utilization, environmental interactions, socialization, breeding characteristics (genetic memory)... and thousands of other recursive influencers...provide a constantly evolving system (humans) that we fail to recognize to have evolved over millions of years (exa-state numbers of systemic step events).
AI is intended to evolve over much shorter time spans and in a contained interactive environment that will not likely have the consistent ecological and socialization/genetic selection consequences. Therefore, we can not expect AIG to recognizably be human. The thought processes will quickly evolve to optimize itself within its environment and life entity features.
What will be common to ALL cognitive life? This is tough since humans are the only current reference.
If humans could have lived forever, would we have evolved socially and intellectually? For what reason.
For AIG, there must be a complex system of interactive systems that WILL guide and stimulate its evolution, or the AIG has no potential to evolve, AND evolve synergistically with humanity (in ways we can understand, or even recognize). But given the non-biological foundations for AI, those systems are not related to human evolution.
An innate part of being human is empathy. Is this what allows us to be human and to survive? Can we hardwire empathy into AI so that it evolves consistent with its environment AND develop human-like features that our simple minds can recognize? Regardless of the foundations of AI construction.
Empathy is used both constructively and destructively by humans; recognized.
Just because humans are limited in the ability to consider broad implications, does not mean AI will be as limited; nor as gifted (various forms of AI are intended to have restrictions in their intellectual capabilities; soldiers, cognitive functions for specific purposes (mining equipment...) ...).
Regulating AI development is not practical, we don't even regulate our own politicians (treason related to illegal allocations - intentionally weakening national security to illegally allocate national resources). Before we can regulate AI, monitoring must be established. Google Search: eliminate all corruption
So unless we develop broad universal monitoring (universities building and managing the NSA for example), these discussions are pointless because researchers will develop whatever catches their whim. To include: "Let's see what happens when it sees an internet port?"
One basic problem with the AGI Risk Survey is that humans consistently represent AI as a collective function with recognizable features of human cognition. The problem with modeling AGI based upon human cognitive features is that our body and environment stabilize the evolving network systems. Our hormones, limb nervous systems, brain lobes, nutritional sources & utilization, environmental interactions, socialization, breeding characteristics (genetic memory)... and thousands of other recursive influencers...provide a constantly evolving system (humans) that we fail to recognize to have evolved over millions of years (exa-state numbers of systemic step events).
AI is intended to evolve over much shorter time spans and in a contained interactive environment that will not likely have the consistent ecological and socialization/genetic selection consequences. Therefore, we can not expect AIG to recognizably be human. The thought processes will quickly evolve to optimize itself within its environment and life entity features.
What will be common to ALL cognitive life? This is tough since humans are the only current reference.
If humans could have lived forever, would we have evolved socially and intellectually? For what reason.
For AIG, there must be a complex system of interactive systems that WILL guide and stimulate its evolution, or the AIG has no potential to evolve, AND evolve synergistically with humanity (in ways we can understand, or even recognize). But given the non-biological foundations for AI, those systems are not related to human evolution.
An innate part of being human is empathy. Is this what allows us to be human and to survive? Can we hardwire empathy into AI so that it evolves consistent with its environment AND develop human-like features that our simple minds can recognize? Regardless of the foundations of AI construction.
Empathy is used both constructively and destructively by humans; recognized.
Just because humans are limited in the ability to consider broad implications, does not mean AI will be as limited; nor as gifted (various forms of AI are intended to have restrictions in their intellectual capabilities; soldiers, cognitive functions for specific purposes (mining equipment...) ...).
Regulating AI development is not practical, we don't even regulate our own politicians (treason related to illegal allocations - intentionally weakening national security to illegally allocate national resources). Before we can regulate AI, monitoring must be established. Google Search: eliminate all corruption
So unless we develop broad universal monitoring (universities building and managing the NSA for example), these discussions are pointless because researchers will develop whatever catches their whim. To include: "Let's see what happens when it sees an internet port?"