It occurred to me that our biases might not matter very much lots of the time. They matter if you want to maximize your chances of finding the truth, but not if you are interested in maximizing the chances the chances of someone finding the truth.
Take science as an example, scientists aren't always free from biases when it comes to thinking about their own theories. They look for ways to confirm what they believe, not discomfirm it. But does that matter? As long as their rivals are looking for ways to falsify those theories then overall the system should work well.
In fact it might work better:
Imagine that there is a prize (a really big prize) at the end of a maze. Several of you are sent into the maze and the first person to find the prize gets to keep it. You all head down the corridor and come to three doors, there are some clues written on the wall, but before you can even begin to read them, someone dashes through door 1. Someone else follows then another through door 2 and a few take door 3. What do you do? You could be methodical and try to solve the clues. This would maximize your chance of finding the right path. However, it will not maximize your chances of being the first one to find the prize. For this, you need to pick a door and run.
Likewise, in science, if you want to get the prize (nobel prize, or a good job, or fame, or a best-selling book), then you might be better off coming up with a new theory and running with it (anecdotally, this seems to be what a lot of successful scientists have done). Having lots of people making leaps in different directions might also make science progress faster overall.
What does all this mean for biases? Are they best thought of on an individual level or group level? Is it really a bias if it is the best thing to do? Can you think of other examples where individual biases might produce better results for the group?
When the a hive of ants (specifically Temnothorax curvispinosus) need to find a new hive location, explorers will spread out looking for likely places. If an ant finds one, she will go back to the existing hive and bodily grab another ant to carry to the new location. If both ants still think it is a good place they each grab another ant, and if everyone still agrees the body-movers bring back 8, 16, 32, etc ants. If a place is marginal then only a small percent of ants will pitch in after being physically carried and the growth will be much slower. A social algorithm like this can select between competing possible hive locations, taking into account issues like distance (less distance meaning shorter doubling times) as well as suitability, and it requires some measure of behavioral variability among the ants. Their aggregate behavior appears to avoid some decision biases found in everything from birds and humans, partly because each ant only seems to judge the quality of one option at a time.
The trick is, this behavior is something that has had many generations to be tuned by evolution. Ants hives have been solving this problem for a long time over a large number of hives that compete with each other for food and territory, whereas there have not been many hundred-million-person-groups competitively solving scientific collective action problems for thousands of generations, such that group selection could have tuned us to do it particularly well. For example, we might be well tuned for rock and spear fights between platoon size groups and some of that might carry over to the effectiveness of groups smaller than 50 people, but I strongly doubt we are genetically tuned for anything that happens at the level of democratic nation states, like subsidized physics research or stock market regulation.
I can imagine one or more human planners who figured out a way for relatively uncultivated humans to be given small amounts of training and arranged in certain institutions that are structurally tuned to thrive on their expected random errors, but doing this would itself require substantial cognitive effort, and if the situation changed (like memory intensive biology turned out to deserve way more resources than working memory intensive theoretical physics) this might need to be recognized somehow and adjusted for via exogenous rationally calculated efforts. You'd be able to point to the person or group of people doing the modeling and optimization work implied, and such people might have existed in the U.S. in the 1890's and 1940's but if they exist for the English speaking internet using world right now, I'm not familiar with their work.
In your example, the people who ran off down each door could probably have produced a better outcome for the group if they had stopped to trade cell phone numbers, agreed on a scheme for dividing the prize, and coordinated their exploration, possibly with some people in reserve to do warm up stretches while waiting to be effectively deployed later based on discoveries relayed back from early explorers of each path. If the maze, represented as a tree structure, had fewer leaf nodes than people and wasn't lopsided, then maybe "everyone run along a path not taken by competitors" could work, but at the very least they should agree on some condition for deciding to come back to the entrance to strategize more effectively based on individual discoveries, or they might all end up simply getting lost.
Adaptive structure doesn't come from nowhere. If Azathoth has not sacrificed a ridiculous number of people on the altar of fitness to have made us into people who are "naturally inclined" towards success in some particular environment, then we have to sacrifice false theories on the altar of truth so that they may die in our stead.