Why, then, is it that people continue to mistrust algorithms, in spite of their superiority, and instead cling to human advice? A recent paper by Dietvorst, Simmons and Massey suggests it is due to a cognitive bias which they call algorithm aversion.
Why is it that opium puts people to sleep? A recent paper by Molière suggests it is due to a property which he calls its dormitive principle.
Haha yes that did strike me too. However, I suppose there could have been other explanations of people's unwillingness to trust algorithms than a cognitive bias of this sort. For instance, the explanation could have been that experts conspire to fool people that they are in fact better than the algorithms. The fact that people mistrust algorithms even in this case, where there clearly wasn't an expert conspiracy going on, suggests that that probably isn't the explanation.
For some background on /u/RichardKennaway's point, see:
My intuition was that it was fear of edge cases, even when the person couldn't articulate exactly what an edge case is.
I would loosely model my own aversion to trusting algorithms as follows: Both human and algorithmic forecasters will have blind spots, not all of them overlapping. (I.e. there will be cases "obvious" to each which the other gets wrong.) We've been dealing with human blind spots for the entire history of civilization, and we're accustomed to them. Algorithmic blindspots, on the other hand, are terrifying: When an algorithm makes a decision that harms you, and the decision is -- to any human -- obviously stupid, the resulting situation would best be described as 'Kafkaesque'.
I suppose there's another psychological factor at work here, too: When an algorithm makes an "obviously wrong" decision, we feel helpless. By contrast, when a human does it, there's someone to be angry at. That doesn't make us any less helpless, but it makes us FEEL less so. (This makes me think of http://lesswrong.com/lw/jad/attempted_telekinesis/ .)
But wait! If many of the algorithm's mistakes are obvious to any human with some common sense, then there is probably a process of algorithm+sanity check by a human, which will outperform even the algorithm. In which case, you yourself can volunteer for the sanity check role, and this should make you even more eager to use the algorithm.
(Yes, I'm vaguely aware of some research which shows that "sanity check by a human" often makes things worse. But let's just suppose.)
I do think an algorithm-supported-human approach will probably beat at least an unassisted human, and I think a lot of people would be more comfortable with it than algorithm-alone. (As long as the final discretion belongs to a human, the worst fears are ameliorated.)
Real world example.
Years ago, I worked at a company which made a machine to screen pap smear slides. Granted, that much of the insanity in health care is about regulatory power, law, and money, but even so, people were just weird about algorithmic screening.
The machine was much more accurate than the great mass of labs in the country.
But no matter how accurate automated screening was when compared to manual screening, there was always such a tizzy about any algorithmic faults. The fact that manual screening produces many more such faults was simply glossed over. Human faults were invisible and accepted, machine faults were a catastrophe.
From that Future of Life conference: if self-driving cars take over and cut the death rate from car accidents from 32000 to 16000 per year, the makers won't get 16000 thank-you cards -- they'll get 16000 lawsuits.
Here's an article in Harvard Business Review about algorithm aversion:
It’s not all egotism either. When the choice was between betting on the algorithm and betting on another person, participants were still more likely to avoid the algorithm if they’d seen how it performed and therefore, inevitably, had seen it err.
My emphasis.
The authors also have a forthcoming paper on this issue:
If showing results doesn’t help avoid algorithm aversion, allowing human input might. In a forthcoming paper, the same researchers found that people are significantly more willing to trust and use algorithms if they’re allowed to tweak the output a little bit. If, say, the algorithm predicted a student would perform in the top 10% of their MBA class, participants would have the chance to revise that prediction up or down by a few points. This made them more likely to bet on the algorithm, and less likely to lose confidence after seeing how it performed.
Of course, in many cases adding human input made the final forecast worse. We pride ourselves on our ability to learn, but the one thing we just can’t seem to grasp is that it’s typically best to just trust that the algorithm knows better.
Presumably another bias, the IKEA effect, which says that people prefer products they've partially created themselves, is at play here.
Possible counterexample:
My father is a professor of electrical engineering. The electronics lab courses involve using simulation software as well as using physical components. In one lab experiment, the students built a circuit that the software didn't simulate correctly (because of simplifications in the models the software used), and one of the questions the students had to answer was why they thought the computer simulation didn't match the measured values. All the students blamed experimental error, and none questioned the computer models...
Probably because humans who don't know much about algorithms basically have no way to observe or verify the procedure. The result of an algorithm has all the force of an appeal to authority, and we're far more comfortable granting authority to humans.
I think people have also had plenty of experience with machines that malfunction and have objections on those grounds. We can tell when a human goes crazy if his arguments turn into gibberish, but it's a bit harder to do with computers. If an algorithm outputs gibberish that's one thing, but there are cases when the algorithm produces a seemingly reasonable number that ends up being completely false.
It's a question of whether to trust a transparent process with a higher risk of error or a black box with a lower, but still non-negligible risk of error.
I'm not sure that explains why they judge the algorithm's mistakes more harshly even after seeing the algorithm perform better. If you hadn't seen the algorithm perform and didn't know it had been rigorously tested, you could justify being skeptical about how it works, but seeing its performance should answer that. Besides, a human's "expert judgment" on a subject you know little about is just as much of a black box.
If people see you as an authority and you make a mistake, they can accept that no one is perfect and mistakes happen. If they doubt the legitimacy of your authority, any mistakes will be taken as evidence of hubris and incompetence.
I think part of it is the general population just not being used to algorithms on a conceptual level. One can understand the methods used and so accept the algorithm, or one can get used to such algorithms over a period of time and come to accept them.
Besides, a human's "expert judgment" on a subject you know little about is just as much of a black box.
And such experts are routinely denounced by people who know little about the subject in question. I leave examples as an exercise for the reader.
And such experts are routinely denounced by people who know little about the subject in question. I leave examples as an exercise for the reader.
True, but that seems inconsistent with taking human experts but not algorithms as authorities. Maybe these tend to be different people, or they're just inconsistent about judging human experts.
It's worth thinking about what makes one an expert, and what convinces others of one's expertise. Someone has to agree that you're an expert before they take you as an authority. There's a social dynamic at work here.
Like other commenters already pointed out, algorithms are scary because they always fail hard. Humans fail, but can recover. Hofstadter's terms seem useful here. Humans can notice or be made to notice that something isn't right and jump out of the system of the basic diagnostic procedure. All the algorithms we currently have are sphexish and will forever remain in their initial framework even when things go wrong.
(I think sphexish is Dawkins, not Hofstadter.)
Hofstadter uses it heavily in Gödel, Escher, Bach in 1979 as the metaphor for things that are unable to Jump Out Of The System. Dawkins only had The Selfish Gene out by then, and The Selfish Gene wasn't really about algorithmic rigidity.
If people trust human forecasters over machine forecasters, but the machine forecasts are better, just use the machine secretly, and take all the credit.
Indeed. That is precisely what the so-called "closet index funds" are doing. They are said to be actively managed funds, but are in reality so-called index trackers, which just are tracking the stock market index.
The reason the managers of the fund are using index-tracking algorithms rather than human experts is, however, not so much that the former are better (as I understand they are roughly on par) but that they are much cheaper. People think that the extra costs that active management brings with it are worth it, however, since they erroneously believe that human experts can consistently beat the index.
The reason the managers of the fund are using index-tracking algorithms rather than human experts is, however, not so much that the former are better (as I understand they are roughly on par)
Maybe human experts tend to track the index anyway?
My intuition was that participants assumed the human forecasters could use some knowledge the model couldn't, but that is definitely not the case in the xperiments. The model and the forecaster have exactly the same data and that is made clear in the setup. It is really that participants prefer the human despite 'better knowledge' - esp. if the comparison is explicit:
Interestingly, participants in the model-and-human conditions, most of whom saw the model outperform the human in the first stage of the experiment [...] were, across all studies, among those least likely to choose the model.
Page 6, emphasis mine.
I admit that I am surprised and I do not understand the cause of this algorithm aversion.
I wonder if this (distrusting imperfect algorithms more than imperfect people) holds for programmers and mathematicians. Indeed, the popular perception seems to be that such folks overly trust algorithms...
I was under the impression that mathematicians are actually too distrusting of imperfect algorithms (compared to their actual error rates). The three examples I ran into myself were:
In analysis, in particular in bifurcation analysis, a (small) parameter epsilon is introduced which determine the size of the perturbation. Analysts always loudly proclaim that 'there exists an epsilon small enough' such that their analysis holds (example values are often around 1/1000), but frequently the techniques are valid for values as large as epsilon = 1/2 (for example). Analysist who are unwilling to make statements about such large values of epsilon seem to be too mistrusting of their own techniques/algorithms.
Whether or not pi and e are normal are open questions in mathematics, but statistical analysis of the first couple of billion of digits (if I am not mistaken) suggests that pi might be normal whereas e is probably not. Still, many mathematicians seem to be agnostic about these questions, as only a few billion data points have been obtained.
In the study of number fields probabilistic algorithms are implemented to compute certain interesting properties such as the class group (algorithms that are guaranteed to give the right answer exist, but are too slow to be used in anything other than a few test cases). These algorithms generally have a guaranteed error rate of about 0.01% (sometimes this is a tune-able parameter), but I know of a few mathematicians in this field (which makes it a high percentage, since I only know a few mathematicians in this field) who will frequently doubt the outcome of such an algorithm.
Of course these are only my personal experiences, but I'd guess that mathematicians are on the whole too fond of certainty and trust imperfect algorithms too little rather than too much.
Are algorithms easier to exploit than humans? Consider the case of playing a video game against a deterministic AI opponent; if the AI makes a stupid mistake once, it'll make the same stupid mistake over and over, and an AI with a known exploit is easy to beat regardless of how well it plays in "normal" circumstances.
Sure, and when algorithms are used in adversarial situations, such as on the stock market, they usually have humans standing guard ready to hit the red button. But most situations are not adversarial, eg, medical diagnosis.
I think it is the issue with moral responsibility, same way as with self-driving cars. People don't want a decision who may negatively affect a person be based on an algorithm because an algorithm is not a moral agent. They want some expert to stand up and say "I risk my professional prestige and blame and all that and declare this person to be prone to violence and thus needs to have a restraining order issued, and if I am wrong it is my bad". As my dad used to say it is all about who is willing to put their dick in the cigar cutter? To accept responsibility, blame, even punishment for a decision made that affects others?
Part of it is rational: decision makers having skin in the game makes decisions better. See Taleb, Anti-Fragile.
Part of it is simply we being used to or evolved to thinking without responsibility there cannot be good decisions, which is true as long as humans make them. We are not evolved to deal with algorithms.
To the blind spots/edge cases objection; the article about SPRs linked in the post shows that even when human experts are given the results of an algorithm and asked to fix their errors, they still do worse than just the algorithm on it's own.
Because even though the human may recognize an obvious mistake, they will also recognize ten good predictions as "mistakes".
It has long been known that algorithms out-perform human experts on a range of topics (here's a LW post on this by lukeprog). Why, then, is it that people continue to mistrust algorithms, in spite of their superiority, and instead cling to human advice? A recent paper by Dietvorst, Simmons and Massey suggests it is due to a cognitive bias which they call algorithm aversion. We judge less-than-perfect algorithms more harshly than less-than-perfect humans. They argue that since this aversion leads to poorer decisions, it is very costly, and that we therefore must find ways of combating it.