24 days
Median attacker dwell time in enterprise networks before detection, per Mandiant M-Trends 2025, meaning automated detection alone leaves a multi-week blind window
68%
Of threat hunters report finding active attacker presence during hunts, demonstrating that automated detection consistently misses live threats
5x
Faster mean time to detect (MTTD) for organizations with mature threat hunting programs compared to those relying solely on automated detection and alerting
MITRE ATT&CK
The primary framework used by 87% of threat hunting teams as the basis for hunt hypothesis generation, per SANS 2025 survey

Threat hunting is the proactive, human-led search for threats that have evaded automated detection. Detection rules fire on known indicators and signatures. Hunting looks for the behavioral patterns of adversaries who are operating within your environment, using techniques that do not match existing rules, credentials that appear legitimate, and persistence mechanisms that blend with normal administrative activity.

The median attacker dwell time of 24 days means that for most successful intrusions, the attacker spent nearly a month in your environment before being detected. Threat hunting exists to close that gap: finding adversaries during the dwell period, before they complete their objectives, rather than after the damage is done.

This playbook is a step-by-step guide for security teams beginning or maturing a threat hunting program. It covers the full hunt cycle from hypothesis generation through data collection, analysis, and converting hunt findings into permanent detection improvements.

Phase 1: Hypothesis Generation

Every hunt starts with a hypothesis: a testable statement about a threat actor technique that might be present in your environment. Hypotheses without threat intelligence grounding waste hunting time on low-probability scenarios. Hypotheses grounded in current threat intelligence direct hunting toward techniques that adversaries targeting your sector are actively using.

Four inputs should drive hypothesis generation. First, threat intelligence relevant to your sector and technology stack: if your organization is a financial institution and a recent Mandiant report documents a threat group targeting financial sector networks using a specific living-off-the-land technique, that technique is a high-priority hypothesis. Second, MITRE ATT&CK: ATT&CK provides a comprehensive taxonomy of adversary techniques with detection notes and data source requirements. For each tactic (Initial Access, Execution, Persistence, Privilege Escalation, Defense Evasion, Credential Access, Discovery, Lateral Movement, Collection, Exfiltration), ATT&CK lists specific techniques that can be translated directly into hunt hypotheses. Third, your own detection gaps: techniques for which you have no existing detection rules are the highest-value hunt targets because automated detection is provably absent. Fourth, prior hunt findings and red team results: if a previous hunt found evidence of credential dumping activity or a red team exercise demonstrated that a specific technique was undetected, follow-up hunting in those technique areas is high-yield.

A well-formed hunt hypothesis follows the structure: 'If [threat actor type] with [objective] has been present in our environment, I would expect to find evidence of [specific technique] in [specific data source].' This structure makes the hypothesis testable and defines the data required before the hunt begins.

Threat intelligence-driven

Base hypotheses on techniques documented for threat actors targeting your sector. Sector-specific CTI reports from vendors like Mandiant, CrowdStrike, Recorded Future, and CISA advisories are primary sources.

ATT&CK-mapped

Map each hypothesis to a specific ATT&CK technique ID. This enables tracking hunt coverage across the matrix and identifying persistent gaps in detection coverage.

Data source validated

Before beginning a hunt, confirm the required data source exists, has adequate retention, and covers the scope needed. A hypothesis that requires Windows Event ID 4688 is invalid if process creation logging is not enabled.

Falsifiable

A good hypothesis can be proven false. If the hunt finds no evidence of the technique in the available data, that is a valid result. Document it. An absence of evidence in a well-instrumented environment is meaningful intelligence.

Phase 2: Data Source Requirements and Collection

The quality of a threat hunt is bounded by the quality and completeness of available telemetry. Hunting for credential dumping with no EDR telemetry, no process creation logging, and no LSASS access events is not possible regardless of analyst skill. Data source validation must occur before the hunt begins.

For endpoint hunting, the minimum data requirements are: process creation events with full command lines (Windows Event ID 4688 with command line auditing enabled, or EDR process telemetry); network connection events from endpoints (source/dest IP, port, process); file creation, modification, and deletion events for sensitive paths; registry modification events for persistence-relevant keys; PowerShell script block logging (Event ID 4104); and module load events for detecting reflective loading and injection techniques.

For identity and authentication hunting, required data includes: all logon events (4624, 4625, 4648); Kerberos ticket requests and service ticket events (4768, 4769, 4771); NTLM authentication events (4776); and Active Directory changes (4720, 4728, 4732, 4756 for group membership changes).

For network hunting, required data includes: full DNS query logs (not just failed queries); proxy logs with full URLs and response codes; NetFlow or PCAP for east-west traffic; and firewall logs with connection state information.

Retention is as important as collection. Attackers with median 24-day dwell times may have established persistence two to three weeks before the hunt begins. Hunting with seven days of logs misses evidence from most of the dwell period. The minimum useful hunting retention is 90 days; 180 days or more is recommended for hunts targeting APT dwell periods.

Free daily briefing

Briefings like this, every morning before 9am.

Threat intel, active CVEs, and campaign alerts, distilled for practitioners. 50,000+ subscribers. No noise.

Phase 3: Hunt Execution and Analytic Techniques

Threat hunting uses a defined set of analytic techniques. Selecting the right technique for each hypothesis improves hunt efficiency and reduces the time spent processing irrelevant data.

Stack counting (frequency analysis) is the most widely applicable technique. Sort every instance of a value in a dataset by frequency. The items at the extremes (very common or very rare) often warrant investigation. Process creation events where a specific parent/child process combination appears only once across your entire fleet in 30 days is worth investigating. PowerShell commands that appear once in three months among 5,000 endpoints are worth examining.

Clustering groups similar events to identify outliers. Clustering endpoint behavior by process tree patterns identifies systems that exhibit unusual behavior compared to their peer group. A finance workstation running the same processes as a developer workstation is an outlier worth investigating.

Baselining establishes the normal distribution of a value and flags deviations. Time-based baselining (processes running at 3 AM that never run at 3 AM in baseline data) and volume-based baselining (DNS query volume from a single endpoint 10x above baseline) surface anomalies that rule-based detection misses because they are statistically defined rather than signature-defined.

Graph analysis maps relationships between entities (users, systems, IP addresses, processes) to identify unusual connection patterns. This is particularly effective for lateral movement hunting: visualizing which accounts authenticated to which systems in a time window reveals unusual access chains that tabular queries miss.

IOC-based pivoting starts from a known indicator (an IP address, domain, file hash, or username) and pivots through related data to identify additional activity from the same actor or campaign. IOC-based hunting is the most reactive technique but is useful when starting from threat intelligence that includes specific indicators.

TTP-based hunting maps a specific ATT&CK technique to the data patterns it produces and searches for those patterns. For example, hunting for T1003 (OS Credential Dumping) searches for LSASS memory reads (Sysmon Event ID 10), outbound network connections shortly after mimikatz-associated process execution, or SAM database access events from non-system accounts.

Phase 4: Investigation, Documentation, and Detection Engineering

When a hunt query returns suspicious results, the analyst transitions from hunting to investigation. The investigation goal is to determine: is this a true positive (actual malicious activity) or a false positive (legitimate but unusual behavior); if true positive, what is the scope and timeline of the activity; and what other evidence exists in other data sources that corroborates or contradicts the initial finding.

Investigation methodology follows the same patterns as alert-driven investigation: pivot from the initial finding through process trees, network connections, file system activity, and authentication events to build a timeline of the attacker's activity. Document every pivot, every query, and every finding, including negative findings. Hunt investigations frequently reveal that what appeared to be a single suspicious event is actually a pattern of activity spanning weeks and multiple systems.

Documentation is the most underinvested part of hunting programs. Every hunt should produce: the hypothesis tested, the data sources and time range queried, the analytic technique applied, all queries and their syntax, the findings (positive or negative), the disposition (confirmed true positive, false positive, or inconclusive), and recommendations for follow-up hunts or detection improvements.

Converting hunt findings into detection improvements is the mechanism by which hunting raises the overall security baseline. Every confirmed true positive finding, and every technique that was huntable but undetected by existing rules, should produce a new detection rule, a refinement to an existing rule, or a documented data source gap that needs to be addressed. A hunting program that does not systematically feed findings back into detection engineering produces the same value indefinitely rather than raising the baseline over time. Track hunt coverage against the ATT&CK matrix to measure which techniques are covered by automated detection, which are covered only by periodic hunting, and which have no coverage at all.

Building a Threat Hunting Program

A mature threat hunting program requires defined processes, dedicated analyst capacity, adequate tooling, and leadership commitment to an activity that produces value over time rather than through individual discrete incidents.

The minimal viable threat hunting program is one dedicated analyst running two to four structured hunts per month against a defined hypothesis backlog, with outputs systematically fed into detection engineering. This level of program produces measurable value: reduced MTTD, increased detection coverage, and periodic discovery of active threats that automated detection missed.

Hunting tooling is separate from alert-driven SOC tooling. Hunters need: a SIEM or data lake capable of long-retention, high-volume search with flexible query syntax (Splunk, Elastic, Google SecOps, Microsoft Sentinel); visualization tools for graph analysis (Maltego, Neo4j, or built-in SIEM graph features); threat intelligence feeds to drive hypothesis generation; and a hunt tracking system to manage hypothesis backlogs, document findings, and track ATT&CK coverage.

Hunting frequency should be calibrated to threat intelligence. After a major threat actor advisory relevant to your sector, run a targeted hunt within 48 hours. Structured hypothesis-driven hunts should run monthly at minimum. Unstructured threat intelligence analysis and data exploration should be a continuous background activity for hunters in between structured hunts.

Measuring hunting program effectiveness uses the same metrics as detection engineering: MTTD for threats discovered by hunting versus automated detection; percentage of ATT&CK techniques with automated detection coverage versus hunting-only coverage; number of new detections created per hunt; and confirmed true positive rate per hunt (a hunting team that never finds confirmed malicious activity is either operating in an environment with no threats, or hunting against hypotheses that are not threat-intelligence-grounded).

The bottom line

Threat hunting is not a substitute for strong automated detection, it is the capability that catches what automated detection misses. The 24-day median attacker dwell time represents three weeks of automated detection failure during which a hunter could be actively searching for adversary presence. Start with two structured hunts per month, ground every hypothesis in current threat intelligence and ATT&CK, confirm your data sources before beginning each hunt, and systematically convert findings into detection improvements. The hunting program that runs consistently for 12 months will have closed detection gaps and found threats that no rule-based system would have identified.

Frequently asked questions

What is the difference between threat hunting and alert investigation?

Alert investigation is reactive: an automated detection fires and an analyst investigates whether it is a true positive. Threat hunting is proactive: a human analyst develops a hypothesis about attacker behavior that automated detection might miss, then searches for evidence of that behavior without waiting for an alert to fire. Hunting specifically targets the detection gaps where rules do not exist or where attackers have evaded existing rules.

What skills does a threat hunter need?

Effective threat hunters combine: deep knowledge of attacker TTPs (MITRE ATT&CK proficiency, familiarity with common malware and living-off-the-land techniques); strong data analysis skills (SIEM query languages, statistical analysis, pattern recognition); knowledge of operating system internals (Windows process trees, registry, event logging; Linux syscalls, /proc, cron); network protocol knowledge (DNS, HTTP, SMB, Kerberos anomalies); and threat intelligence analysis skills to translate CTI reports into actionable hunt hypotheses.

How long should hunts take?

Hunt duration depends on hypothesis complexity and data volume. A focused technical hunt against a single ATT&CK technique with well-defined query logic can be completed in two to four hours. A broad hypothesis covering multiple techniques across a 90-day dataset may take two to three days of analyst time. Scope hunts to fit the analyst's available capacity: an incomplete hunt that runs out of time without producing a documented result wastes time. Better to complete a focused hunt than to start a broad one that never finishes.

What data sources are most important for threat hunting?

The highest-value data sources for most threat hunting are: endpoint process creation logs with command lines (provides visibility into what ran, launched by what, with what arguments); DNS query logs (reveals C2 beaconing, domain generation algorithms, and data exfiltration via DNS); authentication logs (reveals credential abuse, lateral movement, and privilege escalation); and network flow logs (reveals unusual connection patterns, beaconing timing, and data transfer volumes). Without these four sources, many ATT&CK technique hunts are not possible.

How do we prioritize which hypotheses to hunt first?

Prioritize hypotheses based on: threat relevance (is this technique documented for threat actors currently targeting your sector?); detection gap (is there any existing automated detection for this technique?); data availability (do you have the required data sources with adequate retention?); and impact if present (what would an attacker using this technique be able to do in your environment?). High-relevance, no-existing-detection, data-available, high-impact hypotheses are your top priorities.

What is ATT&CK Navigator and how is it used in threat hunting?

ATT&CK Navigator is a web-based tool from MITRE that visualizes coverage across the ATT&CK matrix. Threat hunting teams use it to map which techniques have automated detection coverage (marked in one color), which techniques have hunting-only coverage (marked in another), and which have no coverage (unmarked). This visualization identifies detection gaps and drives hypothesis prioritization. Export the matrix to share detection coverage status with leadership as a measurable program metric.

How do we measure the ROI of a threat hunting program?

The most direct ROI metrics are: true positives found by hunting before automated detection would have found them (the detection lead time); estimated breach cost avoided per confirmed threat found (using industry cost models); mean time to detect reduction attributable to hunting-driven detection improvements; and ATT&CK technique coverage increase from detections created based on hunt findings. Indirect metrics include analyst skill development and threat intelligence feedback loop quality. Present these metrics quarterly to justify program investment.

Sources & references

  1. MITRE ATT&CK Framework
  2. SANS: Threat Hunting and Incident Response Survey 2025
  3. Sqrrl: A Framework for Cyber Threat Hunting
  4. CISA: Threat Hunting Guidance
  5. Cyb3r Dude: The Threat Hunter's Handbook

Free resources

25
Free download

Critical CVE Reference Card 2025–2026

25 actively exploited vulnerabilities with CVSS scores, exploit status, and patch availability. Print it, pin it, share it with your SOC team.

No spam. Unsubscribe anytime.

Free download

Ransomware Incident Response Playbook

Step-by-step 24-hour IR checklist covering detection, containment, eradication, and recovery. Built for SOC teams, IR leads, and CISOs.

No spam. Unsubscribe anytime.

Free newsletter

Get threat intel before your inbox does.

50,000+ security professionals read Decryption Digest for early warnings on zero-days, ransomware, and nation-state campaigns. Free, weekly, no spam.

Unsubscribe anytime. We never sell your data.

Eric Bang
Author

Founder & Cybersecurity Evangelist, Decryption Digest

Cybersecurity professional with expertise in threat intelligence, vulnerability research, and enterprise security. Covers zero-days, ransomware, and nation-state operations for 50,000+ security professionals weekly.

Free Brief

The Mythos Brief is free.

AI that finds 27-year-old zero-days. What it means for your security program.

Joins Decryption Digest. Unsubscribe anytime.

Daily Briefing

Get briefings like this every morning

Actionable threat intelligence for working practitioners. Free. No spam. Trusted by 50,000+ SOC analysts, CISOs, and security engineers.

Unsubscribe anytime.

Mythos Brief

Anthropic's AI finds zero-days your scanners miss.