For Military Officers

The Most Common AI Mistakes Military Officers Make

Military officers working with AI systems face a specific cognitive trap: the pressure of decision timelines makes AI recommendations feel authoritative, even when they rest on incomplete data or pattern-matching errors. The mistakes that follow degrade your command judgement, blur accountability in lethal operations, and create intelligence blind spots that no algorithm will warn you about.

These are observations, not criticism. Recognising the pattern is the first step.

Download printable PDF

Mistaking Speed for Certainty

Palantir systems surface correlated data patterns in seconds, and under operational time pressure you can treat speed as a proxy for accuracy. The system may have ingested corrupted signals, outdated intelligence feeds, or data poisoned by adversaries, and you will not see those weaknesses in the dashboard.

The fix

Before acting on a Palantir recommendation, spend 90 seconds asking your intelligence officer one question: what are the three oldest or weakest data sources feeding this conclusion.

Azure machine learning models rank targets by statistical likelihood of threat, but they cannot weight cultural context, civilian proximity assessments, or strategic consequence the way your experience can. You defer to the ranking because it arrives instantly and you have limited time.

The fix

Treat Azure rankings as a starting list only; manually re-order the top five targets yourself using your knowledge of the operational geography and mission objectives.

DARPA AI systems flag events that match trained threat patterns, but they will match false positives, sensor errors, and civilian activities that look like military ones. Your instinct is to respond fast because the cost of missing a real threat feels higher than the cost of false alarm.

The fix

Create a standing rule: no kinetic response to an AI alert until your forward observer or drone feed confirms the threat signature matches the alert.

AI systems excel at finding what they were trained to find. A novel threat, an adversary tactic you have not seen before, or an entirely new operational approach will not trigger an alert because it does not match the historical patterns in the training data. You believe the system is watching everything.

The fix

Reserve 15 minutes per week to review raw intelligence that did not trigger any AI flag, specifically looking for anomalies your experience tells you matter.

When a defence AI tool reports 94 percent confidence in an assessment, you read that as a 94 percent chance the assessment is correct. The confidence score actually measures how closely the data fits the statistical model, not whether the model itself is valid or whether it has missed something outside its training scope.

The fix

Ask your AI analyst what the confidence score actually measures in this specific tool, and what class of errors the tool is blind to.

Letting AI Erase Your Responsibility

When you tell your command that an airstrike followed an AI target recommendation, you are creating a narrative where the decision feels distributed rather than owned. You remain legally and morally accountable, but the AI language makes that accountability feel blurred to you, your team, and your superiors.

The fix

Document every lethal decision in your operations log using this sentence structure: I ordered a strike on [location] based on [specific intelligence sources], which Palantir ranked as priority one.

When your judgement contradicts what the AI system recommended, you may not articulate the reasoning to your team. Over time, your subordinates stop trusting their own assessment and start treating AI recommendations as authority rather than input. Your command culture shifts toward compliance with algorithms.

The fix

When you override an AI recommendation, tell your team explicitly why: This system did not account for [X factor], so I am ordering [different action].

You run operations with DARPA systems and Palantir without systematically recording whether their assessments were correct, partially correct, or wrong. You have no empirical feedback on whether the system is drifting or degrading. You cannot improve what you do not measure.

The fix

Create a simple three-column log: AI prediction, actual outcome, variance. Review it monthly to identify where the system consistently underperforms.

An Azure AI system flags a pattern and you receive an alert directly, sometimes without your intelligence officer seeing it first. You act on it or pass it to a subordinate who acts on it, and the pattern never goes through your organisation's validation process. Errors propagate faster.

The fix

Require all AI-generated intelligence with operational implications to pass through your S2 or intelligence chief before it reaches the decision point.

You have developed instincts for when to trust and when to challenge your AI tools, but you have not documented or taught those instincts to your deputy or successor. When you leave, the next officer reverts to treating the systems as black boxes.

The fix

Spend one session with your replacement walking through three past decisions where you either accepted or rejected what the AI recommended, explaining your reasoning each time.

Atrophying Your Own Judgement

When Palantir surfaces a threat correlation before you have analysed the raw data, you skip the analytical step and go straight to decision. Your brain does not build the pattern-recognition muscle that comes from sitting with messy data. Over months, your tactical intuition weakens.

The fix

Once per week, analyse a threat scenario without consulting the AI tool first, then compare your assessment to what the system produced.

If an Azure Government AI system does not flag a concern, you assume there is no concern worth worrying about. You stop doing the independent checks that experienced officers do. When the gap is an actual vulnerability, you have no defence.

The fix

Explicitly ask your team each week: What would we care about that this AI system would not flag.

An intelligence AI ranks political factions, tribal relationships, or economic factors in an area, but these require deep knowledge of human networks and local history. You trust the AI ranking because it is systematic, not recognising that you have better intuition from operational experience.

The fix

When AI tools rank non-kinetic factors like political stability or local sentiment, consult your team members who have actually worked in that area before finalising your assessment.

A DARPA system generates three future scenarios based on historical patterns, and you treat those as the three scenarios you need to plan for. You skip the harder work of imagining what an adaptive adversary might do that breaks the historical pattern. Your contingency planning becomes narrow.

The fix

Generate your own set of scenarios first, then use the AI scenarios as a check on whether you missed something obvious.

Palantir or Azure produces an executive summary of intercepted communications or captured documents, and you read the summary instead of the source. You miss the tone, the uncertainty, the qualifier language, and the human context that the source contained. Your situational awareness becomes artificial.

The fix

Read the primary source material for at least one significant intelligence item per week, even though the AI summary exists.

Worth remembering

Related reads

The Book — Out Now

Cognitive Sovereignty: How To Think For Yourself When AI Thinks For You

Read the first chapter free.

No spam. Unsubscribe anytime.