Pronouns: Any / All
Computer Science Scholar (finishing my Masters Degree and looking for a PHD), currently working in AI (Generative Modeling, Score Modeling), and Scentientist.
I am currently convinced that Sentience is not a function of intelligence, but a side effect of the machinery facilitating said intelligence. This would mean that it cannot be inferred from behaviour. I hope to find existing theories and counter arguments here, that I didn't find anywhere else yet.
This resonated a lot with me, but I had a different mental model of this kind of power, which I think may interest you: 'Real' Power vs 'Make Believe' Power.
The 'Kings' power you described relies heavily on a social group collectively believing in your power. If every subject of a King stopped believing in Monarchy at the same time, you would lose your power immediately. If the whole of America were to stop believing in student loan debt, banks would lose a lot of money. And if people stopped believing the the trading value of paper bills, money would lose its value and every rich person would lose a lot of their influence. And if you stop believing that homoeopathic sugar pills will stop your pain, their effect can become a lot lower. This kind of power is defined by the requirement that a critical mass within the social group has to continue believing in it.
What you described as 'Wizard' Power will not vanish in the same way, no matter who believes in it. That is real power. No matter who stops believing in your abilities, they still work.
You could also put the kinds of power on a spectrum. Something is closer to real power the less likely it is to vanish at a moments notice. Money is weaker than the possession of physical materials, for instance. "You cant eat money". Social status is weaker than abilities.
I must disagree and would like to suggest a confounding variable that seems to explain the Feynman example better (I will come to the 'Invention of Lying Scene later'): Intellectual Consistency.
Bohr wanted to talk to Feynman, but not because he was rude or unable to be socially graceful. (It may even be argued that Feynmans fame comes in part from his social abilities). According to your quote, Bohr wanted to talk to him because he didn't disregard his own intuition and knowledge for the sake of being engraciating. All the others doing so, did not recognize Bohr as a truth seeker, but as someone wanting to hear a 'Yes Mr Bohr' to every claim. To a good scientist, this can be insulting, showing lack of social grace.
There is quite a lot of evidence that intellectual consistency increases your chances of exerting minority influence (i. e. being outnumbered in views, but turning the groups view over in the end) greatly. However, if you fail to package this information properly, you might be shunned out of the group or seen as unreasonable before your consistency can have its effect on the group. Let me give an Example:
A common problem for any leftist groups in America is, that they interpret the word 'socialism' differently than the majority of the population. If they critique free market capitalism (and call it by its name) e. g.: 'Free market capitalism lead to a rise of homelessness. Homeless people should be helped by the state to get back on their feet', they will quickly be called a 'commie' and silenced. Note that this is not a reaction to the issue itself, but just to the word 'capitalism'. A strategy to counteract this reaction is to rephrase the problem to 'We need a system that helps homeless people get back into the labour market', and virtually the same claim may be seen as better aligned with american values.
The crux is that humans, even if they prefer honesty, have emotional reactions to what you say. Being challenged sparks cognitive dissonance, but having your self image negatively affected can be too much at once. In short, be consistent, but not insulting.
Back to the 'Invention of Lying' Scene. Firstly, the waiter lacks crucial information to say something like 'She is out of your league'. He does not know who asked whom out. He does not know if 'Leagues' even concern any of them, or if he has merits beyond his looks. For all he knows, the guy could be a famous author she admires. By keeping his mouth shut, he would not omit the truth, but wait until he is sufficiently sure to make any statement.
Secondly, the reason this clip seems so funny to us is that any human with emotions would have a negative emotional reaction to such a remark, which is just missing here. We can wish all we want that humans were like this, but as it stands right now, even the most blunt people have a sore spot, and scratching it will only lead to discord if they are not in the proper mood for it. For the truth to have any effect on the others behaviour, you have to make sure others are ready to hear it
Hi! I am a master student in Computer Science majoring in — yes you guessed it — AI, but also with a background in psychology.
I was always interested in thinking and optimizing my own thought patterns. It's probably half my urge to be 'more intelligent' and half the necessity to overcome my ADHD challenges. Through my studies, I already learned about many of the things taught here or in Eliezers works, but HPMOR and the Sequences still had/have a lot to teach me, or at least help my knowledge affect my actions.
I pursue AI due to my interest in cognition. I would like to know how intelligence and reasoning work, and what heuristics make them better or worse. A proof of understanding, after all, is the ability to recreate it using your beliefs about it.
I also have a deep interest in ethics. My current stance is a form of sentientism. In short, I currently believe most (human or non-human) animals have a certain ability to suffer, which is correlated with, but not caused by, intelligence. I want to grant these beings rights for protection from unnecessary harm and the likes. Intelligence just gives you the additional rights; e. g. to pursue a purpose.
I don't really have a lot of people whom I can discuss my thoughts and Ideas with in real life. Most non-AI people hate the work I am doing or are just disinterested, and many AI people hope for salvation from AGI or riches from their jobs. I just want to understand how thinking works and can be optimized, and I have the feeling there are people here who think like me. Of course, I would also like to learn more strategies for system 2 thinking, and improvmymy heuristics for system 1. I also hope I can contribute to this collection of knowledge at some point! :D
One thing I would like to add in terms of a life-goal or life strategy. Learning how to make vaccines or microchips is cool, but it requires you to stay within some amount of financial stability. Using your metaphor, it is wizardry that requires a wand. If you sell your lithography machine for your next month's rent (inequality is rising), the usefulness of your skill will vanish as quickly as the kings power during the French revolution.
I therefore find it prudent to start at the bottom assuming no possessions. One would hence start out with basic survival skills. Procuring calories, making necessary tools from trash or natural materials, solid first aid, making crude medicines, constructing shelter. One can then go on to other things like producing electricity and so on. Power becomes more real if it depends on less outside conditionals.
Learn being a human first, then go on to becoming a wizard.