Strong upvote for the quality of suggestion.
I have seen first hand otherwise very intellectual folks devolve into child-like argumentation very rapidly, usually after the word 'surely' or some equivalent phrase.
Such as 'surely everyone is like that', 'surely they won't behave like that', etc.
I had trouble making the connection between the first two paragraphs and the rest. Are you introducing what you mean by an "alarm" and then giving a specific proposal for an alarm afterwards? Is there significance in how the example alarms are in response to specific words being misleading?
Yeah, all these "alarms" are supposed to warn you that the word (or something) might be misleading, and you should pay extra attention (unless it's already obvious) to avoid being misled. Or, pay extra attention because there is something you can do in response which is profitable.
Daniel Dennett argues for the idea of a "Surely alarm" which says that whenever you hear someone say "surely" in an argument, an alarm should sound in your head—*ding!*—and you should pay attention for weaknesses. When someone says "surely", it's often just meant to appeal to a sense of disbelief that anyone could possibly disagree with them.
Readers of the Sequences will be familiar with similar alarms upon hearing "emergence", "by definition", "scientific", "rational", "merely", and more. And if you're really unlucky, some words just explode without warning.
Anyway, when you notice that people aren't doing something to attain what they've explicitly expressed a preference for, you should be wondering what's going on. I've seen countless examples of people noticing the discrepancy only to brush it aside because "surely if they wanted to do it, they would have." People routinely underestimate the cost of some seemingly insignificant obstacles.
The fact that A) it's obvious that people really want Y, and that B) it's also obvious that there's almost no cost to doing X in order to attain Y, should not just cause you notice-it-as-plain-fact. There's a deep mystery here, and if you also believe both A and B, there's real epistemic profit to be had by paying close attention. The potential profit is proportional to the total strength of the beliefs you've just noticed are in direct conflict.
You might naively think that if the confusion is so profitable, surely you—with all your ability to reason and pay close attention to reality—would've already noticed and exploited it. But if acquired beliefs A and B from a pre-attentive stage of your life, this may also explain why you've skipped over this obvious opportunity until now.
Besides, it's really cheap to do a second look-over, so why-don't-you-just investigate?