I think there’s lots of specific internal reasons why people make bad choices: sometimes it’s just pure selfishness of sadism.
But as for why some people are delusional, selfish, sadistic. As for why some people “succumb to evolved default behaviors like anger, instead of using their freedom of thought.” I’m not really seeing an alternate explanation here other than some people where unlucky enough to have genes and environment that built a brain that followed the laws of physics until it they did something bad. And from an internal perspective, maybe the p...
We can simulate the brain of C. elegans, I see no reason why it couldn’t theoretically be scaled up to a human brain. I guess technically you need computation AND a full map of the human brain not just computation for that.
I think the atoms in my brain will follow the laws of physics until a choice is made. And to me that process feels like I’m deciding something, because that’s what computation feels like from the inside. But actually the outcome is predetermined.
No, but only because I lack the computing power to do so. I very powerful AI could.
So why do some people choose to do good while others choose to do evil? I think genes and environment are fully sufficient to explain why people make different choices, but if you have an alternate hypothesis I’d be interested to hear it. But the answer can’t be something like “because some people choose different intentions” because then you’d have to explain why some people have different intentions.
To put it another way, you may choose your intentions deliberately, but did you make the choice to be the kind of person who chooses intentions deliberately?...
You raise two very valid concerns. That Hitler might hurt others if you allow him to interact with them, and that Hitler might find a way to escape the box.
Even if Hitler was willing to reflect on his actions and change, his presence in the network (B) would likely make other people unhappy.
So while I think (A) is ethically mandatory if you can contain him, (B) comes with a lot of complex problems that might not be solvable.
The bit of your brain that chooses to think nice thoughts (“I”/“me”) is just as much a product of your genes and environment as the bit of your brain that wants to think bad thoughts.
You didn’t choose to have a brain that tries not to think bad thoughts and Hitler didn’t choose to have a brain that outputs genocide when given some specific environmental conditions. The only way Hitler could have realised that his actions were bad and choose to be good would be if his genes and environment built a brain that would do so given some environmental input.
Hitler’s evil actions were determined by the physical structure of his brain. His brain was built by genes (which he didn’t choose), and modified by his environment (which didn’t choose), and then certain environmental inputs (which he didn’t choose) caused his brain to output genocide. If you had Hitler’s genes and Hitler’s environment, you would have Hitler’s brain and so you would do as Hitler did.
To punish someone, or in this case withhold high resolution paradise, can only be useful and good in so far as it changes behaviour or acts as a deterrent to ...
I think more exposition is needed. For example, one episode could have someone who knows how dangerous AI is, warns the other characters about it, and explains toward the end why things are going wrong. In other episodes, the characters could realise their own mistake, far too late, but in time to explain what's going on with a bit of dialogue. Alternatively, the AI explains its own nature before killing the characters.
For example, at the end of Cashbot, as nukes are slowly destroying civilisation, someone could give a short monologue about how AIs don't have human values, ethics, empathy or restraint, and that they will follow their goals to the exclusion of all else.
“Suffering is a mainly measure, rather than the target metric.”
What is the metric? What could be more important than reducing suffering and increasing happiness? Things are only bad if they cause suffering or reduce happiness, and only good if they increase happiness or decrease suffering.
“Either what you are doing was fine before, or you did not hereby make it fine.”
The bad things we do to animals (cages, slaughter, etc) are bad because it causes them suffering. If we find a way to prevent them from suffering, these bad things are no longer bad.
“but expec... (read more)