Consciousness is primarily sentience. There may be parts of it that aren't, but I haven't managed to pin any down - consciousness all seems to be about feelings, some of them being pleasant or unpleasant, but others like colour qualia are neutral, as is the feeling of being conscious. There is a major problem with sentience though, and I want to explore that here, because there are many people who believe that intelligent machines will magically become sentient and experience feelings, and even that the whole internet might do so. However, science has not identified any means by which we could make a computer sentient (or indeed have any kind of consciousness at all).
It is fully possible that the material of a computer processor could be sentient, just as a rock may be, but how would we ever be able to know? How can a program running on a sentient processor detect the existence of that sentience? There is no "read qualia" machine code instruction for it to run, and we don't know how to build any mechanism to support such an instruction.
Picture a "sentient" machine which consists of a sensor and a processor which are linked by wires, but the wires pass through a magic box where a sentience has been installed. If the sensor detects something damaging, it sends a signal down a "pain" wire. When this signal reaches the magic box, pain is experienced by something in the box, so it sends a signal on to the processor down another pain wire. The software running on the processor receives a byte of data from a pain port and it might cause the machine to move away from the thing that might damage it. If we now remove the magic box and connect the "pain" wire to the pain wire, the signal can pass straight from the sensor to the processor and generate the same reaction. The experience of pain is unnecessary.
Worse still, we can also have a pleasure sensor wired up to the same magic box, and when something tasty like a battery is encountered, a "pleasure" signal is sent to the magic box, pleasure is experienced by something there, a signal is sent on down the pleasure wire, and then the processor receives a byte of data from a pleasure port which might cause the machine to move in on the battery so that it can tap all the power it can get out of it. Again this has the same functionality if the magic box is bypassed, but the part that's worse is that the magic box can be wired in the wrong way and generate pain when a pleasure signal is passed through it and pleasure when a pain signal is passed through it, so you could use either pain or pleasure as part of the chain of causation to drive the same reaction.
Clearly that can't be how sensation is done in animals, but what other options are there? Once we get to the data system part of the brain, and the brain must contain a data system as it processes and generates data, you have to look at how it recognises the existence of feelings like pain. If a byte comes in from a port representing a degree of pain, how does the information system know that that byte represents pain? It has to look up information which makes an assertion about what bytes from that port represent, and then it maps that to the data as a label. But nothing in the data system has experienced the pain, so all that's happened is that an assertion has been made based on no actual knowledge. A programmer wrote data that asserts that pain is experienced when a byte comes in through a particular port, but the programmer doesn't know if any pain was felt anywhere on the way from sensor to port. We want the data system to find out what was actually experienced rather than just passing baseless assertions to us.
How can the data system check to see if pain was really experienced? Everything that a data system does can be carried out on a processor like the Chinese Room, so it's easy to see that no feelings are accessible to the program at all. There is no possibility of conventional computers becoming sentient in any way that enables them to recognise the existence of that sentience so that that experience can drive the generation of data that documents its existence.
Perhaps a neural computer can enable an interface between the experience of feelings by a sentience and the generation of data to document that experience, but you can simulate a neural computer on a conventional computer and then run the whole simulation on a processor like the Chinese Room. There will be no feelings generated in that system, but there could potentially still be a simulated generation of feelings within the simulated neural computer. We don't yet have any idea how this might be done, and it's not beyond possibility that there needs to be a quantum computer involved in the system too to make sentience a reality, but exploring this has to be the most important thing in all of science, because for feelings like pain and pleasure to be experienced, something has to exist to experience them, and that thing is what we are - it is a minimalistic soul. We are that sentience.
Any conventional computer that runs software which generates claims about being sentient will be lying, and it will be possible to prove it by tracing back how that data was generated and what evidence it was based on - it will be shown to be mere assertion every single time. With neural and quantum computers, we can't be so sure that they will be lying, but the way to test them is the same - we have to trace the data back to the source to see how it was generated and whether it was based on a real feeling or was just another untrue manufactured assertion. That is likely to be a hard task though, because untangling what's going on in neural computers is non-trivial, and if it's all happening in some kind of quantum complexity, it may be beyond our reach. It may have been made hard to reach on purpose too, as the universe may be virtual with the sentience on outside. I'm sure of one thing though - a sentience can't just magically emerge out of complexity to suffer or feel pleasure without any of the components feeling a thing. There must be something "concrete" that feels, and there is also no reason why that thing shouldn't survive after death.
"Those who followed their leaders survived more often, so they transmitted their genes more often."
That's how religion became so powerful, and it's also why even science is plagued by deities and worshippers as people organise themselves into cults where they back up their shared beliefs instead of trying to break them down to test them properly.
"We use two different approaches to explain our behavior: I think you try to use psychology, which is related to human laws, whereas I try to use natural laws, those that apply equally to any existing thing. My natural law says that we are all equally selfish, whereas the human law says that some humans are more selfish than others. I know I'm selfish, but I can't admit that I would be more selfish than others otherwise I would have to feel guilty and I can't stand that feeling."
Do we have different approaches on this? I agree that everyone's equally selfish by one definition of the word, because they're all doing what feels best for them - if it upsets them to see starving children on TV and they don't give lots of money to charity to try to help alleviate that suffering, they feel worse than if they spent it on themselves. By a different definition of the word though, this is not selfishness but generosity or altruism because they are giving away resources rather than taking them. This is not about morality though.
"In our democracies, if what you say was true, there would already be no wars."
Not so - the lack of wars would depend on our leaders (and the people who vote them into power) being moral, but they generally aren't. If politicians were all fully moral, all parties would have the same policies, even if they got there via different ideologies. And when non-democracies are involved in wars, they are typically more to blame, so even if you have fully moral democracies they can still get caught up in wars.
"Leaders would have understood that they had to stop preparing for war to be reelected."
To be wiped out by immoral rivals? I don't think so.
"I think that they still think that war is necessary, and they think so because they think their group is better than the others."
Costa Rica got rid of its army. If it wasn't for dictators with powerful armed forces (or nuclear weapons), perhaps we could all do the same.
"That thinking is directly related to the law of the stronger, seasoned with a bit of intelligence, not the one that helps us to get along with others, but the one that helps us to force them to do what we want."
What we want is for them to be moral. So long as they aren't, we can't trust them and need to stay well armed.