For Cybersecurity Professionals
Cybersecurity analysts often hand over threat assessment and incident prioritisation to AI systems, then struggle to explain why real attacks slip past the alerts. This creates two dangers at once: you stop building the adversarial thinking that catches novel attacks, and you lose the ability to spot when AI has missed something critical.
These are observations, not criticism. Recognising the pattern is the first step.
Darktrace and CrowdStrike AI rank threats by statistical models that work well for known attack patterns but give no explanation of their reasoning. When you trust the score without checking, you miss attacks that don't fit the training data.
The fix
For every critical alert, write down what you see in the raw logs before checking what the AI says, then compare your reading to its verdict.
AI threat detection tools are tuned to catch suspicious behaviour, which means they generate hundreds of signals daily. Analysts often assume high alert counts mean better coverage, when really it means you are drowning in noise and missing the real threat in the middle.
The fix
Measure alert quality by false positive rate and time to resolution, not by raw alert count.
Microsoft Security Copilot and Splunk AI both apply pre-filtering to decide which alerts warrant your attention. You never see the ones the AI threw away, so you cannot judge if it made the right call or if it discarded something important.
The fix
Configure your tools to log all raw detections separately, and sample them weekly to check if the AI's filtering is hiding attack patterns you should know about.
When ChatGPT or Copilot rates a malware sample as low risk with high confidence, analysts often skip the manual checks that would have caught the variant. The tool's certainty feels authoritative even though it may have never seen this exact specimen before.
The fix
Threat hunt manually on every sample that touches critical systems, regardless of what the AI confidence score says.
Darktrace and CrowdStrike train on global attack data, not your specific network behaviour and threat model. The system may flag legitimate activity in your environment as abnormal because it doesn't match threats seen elsewhere.
The fix
Spend the first month after deployment tuning the AI to your actual baseline, and revisit those baselines every quarter.
When Darktrace or CrowdStrike identifies a suspicious connection, many analysts skip the manual packet inspection that would reveal what data actually moved. You lose the ability to spot data exfiltration that doesn't match known signatures.
The fix
Manually inspect packet payloads for every incident rated as high or critical, even if the AI already gave you a verdict.
Microsoft Security Copilot can generate a narrative of an incident in seconds, but it works from the detections the AI already made. You miss attack chains that require you to find and connect events the system never flagged.
The fix
Do your own timeline of events from raw logs before reading the AI summary, then compare what you found to what it found.
When Copilot or ChatGPT suggests isolating a host or blocking an IP, it is giving you the textbook response, not the response that fits your specific attack scenario. You might cut off your own visibility or tip off the attacker.
The fix
Before executing any AI-recommended action, write down what you expect to see if the recommendation is wrong, and what you'll do then.
Splunk AI and Copilot excel at matching known escalation patterns but often miss the subtle behaviour changes that an analyst who reads logs regularly would catch. The adversarial thinking required to spot new techniques atrophies.
The fix
Manually review privilege change logs for your critical systems weekly, and look for behaviour that confuses you, not just behaviour the AI flagged.
When you ask ChatGPT or Copilot to rank security priorities, it returns the most common vulnerabilities and standard fixes. It cannot model what your organisation actually looks like to an attacker or what your real attack surface is.
The fix
Before asking AI for recommendations, write down the three attack paths you think matter most to your environment, then use AI to test those specific paths.
Vulnerability assessment tools rate all exposures against generic severity scales. They do not know which systems support your critical business processes or which compromise would hurt you most.
The fix
For any vulnerability AI rates as critical, verify it against your actual threat model before treating it as critical to fix.
When you ask Copilot what attack scenarios to test, it suggests the ones it knows how to model. You stop thinking about the attacks that are specific to your organisation, your industry, or your recent threat intelligence.
The fix
Design your penetration testing scenarios before talking to AI, then use AI to help you execute those scenarios, not to choose them.
ChatGPT and other models synthesise threat data from many sources but have no idea if a particular threat is real for you. You might spend resources on threats that are not relevant while ignoring threats that actually target your sector.
The fix
Cross-check any AI-generated threat assessment against your actual incident logs and your sector-specific threat reports.
Worth remembering
The Book — Out Now
Read the first chapter free.
No spam. Unsubscribe anytime.