Perhaps the most important concept in strategy is the importance of operating within the right paradigm. It is extremely important to orient towards the right style or the right doctrine before you begin implementing your plan - that could be "no known doctrine, we'll have to improvise", but if it is you need to know that! If you choose the wrong basic procedure or style, you will end up refining a plan or method that ultimately can't get you to where you want to, and you will likely find it difficult to escape.
This is one of the Big Deep Concepts that seem to crop up all over the place. A few examples:
- In software development, one form of this error is known as "premature optimization," where you focus on optimizing existing processes before you consider whether those processes are really what the final version of your system needs. If those processes end up getting cut, you've wasted a bunch of time; if you end up avoiding "wasting work" by keeping these processes, the sunk cost fallacy may have blocked you from implementing superior architecture.
- In the military, a common mistake of this type leads to "fighting the last war" - the tendency of military planners and weapons designers to create strategies and weapon systems that would be optimal for fighting a repeat of the previous big war, only to find that paradigm shifts have rendered these methods obsolete. For instance, many tanks used early in World War II had been designed based on the trench warfare conditions of World War I and proved extremely ineffective in the more mobile style of warfare that actually developed.
- In competitive gaming, this explains what David Sirlin calls "scrubs" - players who play by their own made-up rules rather than the true ones, and thus find themselves unprepared to play against people without the same constraints. It isn't that the scrub is a fundamentally bad or incompetent player - it's just that they've chosen the wrong paradigm, one that greatly limits their ability when they come into contact with the real world.
This same limitation is present in almost every field that I have seen, and considering it is critical. Before you begin investing heavily in a project, you should ask yourself whether this is really the right paradigm to accomplish your goals. Overinvesting in the wrong paradigm has a doubly pernicious effect - not only are your immediate efforts not as effective as they could be, but it also renders you especially vulnerable to the sunk cost fallacy. Keep in mind that even those who are aware of the sunk cost fallacy are not immune to it!
Therefore, when making big decisions, don't just jump into the first paradigm that presents itself, or even the one that seems to make the most sense on initial reflection. Instead, realy truly consider whether this approach is the best one to get you what you want. Look at the goal that you're aiming for, and consider whether there are other ways to achieve it that might be more effective, less expensive, or both.
Here are some sample situations that can be considered paradigm-selection problems:
- Do you really need to go and get a CS degree in order to become a computer programmer, or will a bootcamp get you started faster and cheaper?
- Does your organization's restructuring plan really hit the core problems, or is it merely addressing the most obvious surface-level issues?
- Will aircraft carrier-centric naval tactics be effective in a future large-scale conventional war, or is the aircraft carrier the modern equivalent of the battleship in WW2?
I don't necessarily know the answers to all these questions - note that only one is even framed as a clear choice between two options, and there are obviously other options available even in that case - but I do know that they're questions worth asking! When it comes time to make big decisions, evaluating what paradigms are available and whether the one you've chosen is the right one for the job can be critical.
My quite simplistic understanding is that (a) yes, there are already existing solutions, but (b) the people providing those solutions will charge you a lot of money, especially if you later become dependent on them; and you still need to check their work, and they may disagree with you on some details because at the end of the day they are optimizing for themselves, not for you.
Doing things yourself requires extra time and energy, but the money which would otherwise become someone else's profit now stays in your pockets. Essentially, as soon as you feel reasonably sure that the project will be successful, getting rid of each subcontractor means increasing your profit. You don't need to become an expert on everything, you can still hire the experts, but now they are your employees working for a salary, instead of a separate company optimizing for their own profit.
Not sure how realistic this is, but if you imagine that even a typical successful company somewhat resembles the Dilbert comic, then if you can build your own company better, you can just take over their people who do the actual work, and stop feeding the remaining ones.
EDIT: I don't have an experience running a company, but I am thinking about a friend who recently reconstructed his house. His original thoughts were "I am a software developer, this is my competitive advantage, so I will just pay the people who are experts on house construction", but it turned out that the real world doesn't work this way. Most of the so-called experts were quite incompetent, and he had to do a lot of research in their field of expertise just to be able to tell the difference. When the reconstruction was over, he already felt like he could start a new profession and do better than most of these experts. In this case, however, those experts were typically sole proprietors. If instead they would have been companies renting the experts, and if my friend would be in some kind of a business of repeatedly reconstructing houses, it would make sense for him to start optimizing the details seriously, and build replacements of what exists out there.