
“Monitor everything” is the conventional wisdom. Collect all the logs. Build all the alerts. Correlate all the events.
In practice, this produces a SIEM that generates hundreds of alerts daily, most of which are noise. Your security team spends hours triaging false positives. Actual attacks hide in the flood.
Smart monitoring isn’t about more alerts—it’s about better alerts. Detecting attacker behavior rather than just logging everything and hoping you’ll notice.
Most organizations approach detection backwards. They start with data sources—”we have Windows event logs, let’s write alerts against them”—and end up with detection based on what’s available rather than what matters.
Flip the approach. Start with attacker behavior. What do attackers actually do? Then ask: how would we detect that behavior? What data do we need?
This sounds obvious but is rarely practiced. The result of data-first approaches is alert fatigue, missed attacks, and detection that covers easy-to-detect behaviors while missing harder-to-detect ones.
Before exploitation comes reconnaissance. Attackers map your environment to find opportunities. Detecting this phase provides early warning—but it’s also where most monitoring fails.
What happens: Port scanning, service enumeration, vulnerability scanning, banner grabbing, DNS reconnaissance.
Why detection is hard: Internet background noise is massive. Automated scanners, researchers, bots—everyone is scanning everyone. Your attack surface receives constant probing. Distinguishing targeted reconnaissance from noise is nearly impossible at scale.
What to detect instead:
Resource-appropriate approach: Don’t try to alert on all scanning. Focus on scanning that targets your crown jewels specifically, or scanning followed by authentication attempts. Correlation is key.
Once inside (via phishing, compromised credentials, etc.), attackers map the internal network.
What happens: Network scanning, LDAP/AD enumeration, share discovery, internal DNS queries, service discovery.
What to detect:
Example detections:
Source IP X attempted connections to > 20 unique internal IPs on port 445 in 10 minutes
Account X queried >100 user or group objects from Active Directory in 5 minutes
Host X connected to >50 unique ports across internal network in 1 hour
After reconnaissance, attackers move. This is often the most detectable phase—movement creates logs.
What happens: Attackers use stolen credentials to authenticate to other systems—RDP, SSH, WinRM, SMB.
What to detect:
Example detections:
Account X logged into Host Y for the first time (never seen in 90-day baseline)
Account X normally authenticates from Workstation A; now authenticating from Workstation B
Service account S authenticated interactively (service accounts shouldn't log in interactively)
What happens: Attackers use credential hashes or Kerberos tickets without knowing passwords.
What to detect:
Resource-appropriate approach: These detections require baseline knowledge of normal authentication patterns. Start with service accounts and privileged accounts—they have more predictable patterns.
What happens: PsExec, WMI, WinRM, SSH—attackers execute commands on remote systems.
What to detect:
Example detections:
A service was created on Host Y via remote call from Host X
PowerShell remoting session initiated from Host X to Host Y (not typical admin activity)
Attackers access data. Often, this is the objective.
What to detect:
Example detections:
User X accessed >500 files on SharePoint in 1 hour (baseline: <100)
Account X accessed HR share for the first time
Privileged account activity at 3 AM (account has no prior 3 AM activity)
What to detect:
Resource-appropriate approach: You can’t inspect all traffic. Focus on:
Enterprise detection strategies don’t translate to smaller teams. Here’s a scaled approach:
These generate few false positives and catch high-impact activity. Start here.
These detections are low-tuning, high-value. Get these working first.
Focus on accounts where compromise would be catastrophic.
Build behavioral baselines for these accounts:
Alert on deviation. The limited scope makes tuning manageable.
Harder to tune, but valuable:
Modern NDR (Network Detection and Response) tools help here, but they’re expensive. Proxy and firewall logs with custom analysis can provide some of the same visibility.
Once Tiers 1-3 are solid and tuned, expand:
Every detection requires tuning. Alerts that work in lab conditions generate noise in production.
Expect this process:
This isn’t failure—it’s how detection engineering works. Budget time for tuning. A detection that alerts 100 times a day and gets ignored is worse than no detection.
Tuning heuristics:
Single events are often ambiguous. The authentication from a new device might be an attacker—or an employee got a new laptop. The large file access might be exfiltration—or someone preparing a presentation.
Correlation provides confidence:
Example chain:
Any single event might be noise. The chain is clearly malicious.
Build correlation rules that connect related events:
Measure your detection capability:
Coverage: What percentage of MITRE ATT&CK techniques relevant to your threat model do you have detection for?
Efficacy: When you test detection (via red team, purple team, or atomic tests), what percentage fires correctly?
Signal-to-noise: What percentage of alerts are true positives vs. false positives?
Time to detect: For true positives, how long between malicious activity and alert?
Time to triage: How long does it take analysts to determine if an alert is real?
Track these over time. Improvement should be visible.
You will not detect everything. Sophisticated attackers with time and resources can evade detection.
The goal isn’t perfect detection—it’s making detection reliable enough that attackers can’t operate freely, and fast enough that damage is limited when they do get in.
Focus on high-value, low-noise detections first. Build behavioral baselines for your most critical accounts. Invest in correlation and context. Tune relentlessly.
A few detections that work beat hundreds of alerts nobody investigates.