It would be surprising, if bad intent were so rare in the relevant sense, that people would be so quick to jump to the conclusion that it is present. Why would that be adaptive?
You may not be wrong but I don't think it would necessarily be surprising. We adapted under social conditions that are radically different than exist today. It may no longer be adaptive.
Hypothesis: In small tribes and family groups assumptions of bad faith may have served to help negotiate away from unreasonable positions while strong familial ties and respected third parties mos...
I really liked this post. I thought it was well written and thought provoking.
I do want to push back a bit on one thing though. You write:
What makes for a crony belief is how we're rewarded for it. And the problem with beliefs about climate change is that we have no way to act on them — by which I mean there are no actions we can take whose payoffs (for us as individuals) depend on whether our beliefs are true or false.
It is true that most of us probably won't take actions whose payoffs depend on beliefs about global warming, but it is not true that th...
Thanks, I did end up figuring out my error.
Maybe I'm confused, in the 'muddy children puzzle' it seems it would be common knowledge from the start that at least 98 children have muddy foreheads. Each child sees 99 muddy foreheads. Each child could reason that every other child must see at least 98 muddy foreheads. 100 minus their own forehead which they cannot see minus the other child's forehead which the other child cannot see equals 98.
What am I missing?
Desire is a contract you make with yourself to be unhappy until you get what you want.
You can see more results here: Image Annotation Viewer
Judging generously, but based on only about two dozen or so image captions, I estimate it gives a passably accurate caption about one third of the time. This may be impressive given the simplicity of the model, but it doesn't seem unreasonably effective to me, and I don't immediately see the relevance to strong AI.
Let's say you precommit to never paying off blackmailers. The advantage of this is that you are no longer an attractive target for blackmailers since they will never get paid off. However if someone blackmails you anyway, your precommitment now puts you at a disadvantage, so now (NDT)you would act as if you had a precommitment to comply with the blackmailers all along since at this point that would be an advantageous precommitment to have made.
It's a funny joke but beside the point. Knowing that he is in a balloon about 30 feet above a field is actually very useful. It's just useless to tell him what he clearly already knows.
I recall a SF story that took place on a rotating space station orbiting Earth that had several oddities. The station had greater than Earth gravity. Each section was connected to the next by a confusing set of corridors. The protagonist did some experiments draining water out of a large vat and discovered a coriolis effect.
So spoiler alert it turned out that the space station was a colossal fraud. It was actually on a massive centrifuge on Earth.
Due to the finite speed of sound, the explosion would have had to occur approximately 20 seconds before they heard it. So if Voldemort's death was coincident with the explosion it would had to have happened about 20 seconds before Harry said it did.
...She'd just about decided that this had to all be a prank in unbelievably poor taste, when a distant but sharp CRACK filled the air. [...] "It worked," Harry Potter gasped aloud, "she got him, he's gone." [...] "I think it's in that direction." Harry Potter pointed in the rough dir
Why would it backtrack (or what do you mean by backtrack)? Eventually, it observes that w = false (that "ON" went through unchanged) and that its actions are no longer beneficial, so it just stops doing anything, right? The process terminates or it goes to standby?
I think the presumption is that the case where the "ON" signal goes thru normally and the case where the "ON" signal is overwritten by a thermodynamic miracle... into exactly the same "ON" signal are equivalent. That is that after the "ON" sign...
I was not aware of Tuxedage's ruleset. However any ruleset that allows for the AI to win without being explicitly released by the gatekeeper is problematic.
If asd had won due to the gatekeeper leaving it would only have demonstrated that being unpleasant can cause people to disengage from conversation, which is different from demonstrating that it is possible to convince a person to release a potentially dangerous AI.
That's not really in the spirit of the experiment. For the AI to win the gatekeeper must explicitly release the AI. If the gatekeeper fails to abide by the rules that merely invalidates the experiment.
Everything is actually about signalling.
Counterclaim: Not everything is actually about signalling.
Almost everything can be pressed into use as a signal in some way. You can conspicuously overpay for things to signal affluence or good taste or whatever. Or you can put excessive amounts of effort into something to signal commitment or the right stuff or whatever. That almost everything can be used as a signal does not mean that almost everything is being used primarily as a signal all of the time.
Signalling only makes sense in a social environment, so thi...
That may well be true, but I should clarify that neither of my hypotheticals require or suggest that bad faith communication was more common in the past. They do suggest that assumptions of bad faith may have been significantly more common than actual bad faith, and that this hypersensitivity may have been adaptive in the ancestral environment but be maladaptive now.