$1.76M
average cost of a security breach that a functioning SOC could have detected faster, per the IBM Cost of a Data Breach Report 2024
277 days
average time to identify and contain a breach without mature detection capabilities, per IBM 2024
Tier 1 analysts
handle 60 to 80 percent of all SOC alerts through triage and initial investigation
3:1
approximate ratio of Tier 1 to Tier 2 analysts in a well-structured SOC

Building a Security Operations Center from scratch is one of the most ambitious security program investments an organization can make. It requires sustained commitment across people, process, and technology, and the failure modes are well-documented: underfunded SOCs with too few analysts produce alert backlogs that defeat the purpose of monitoring; SOCs built on the wrong technology stack spend more time managing the platform than detecting threats; and SOCs without executive sponsorship get deprioritized when budget cycles tighten. The organizations that build effective SOCs treat it as a multi-year program, not a technology purchase.

This guide covers every dimension of a SOC build, from initial scope definition and sponsorship through technology stack selection, staffing models, detection engineering, runbook development, and the maturity progression that separates a reactive alert queue from an intelligence-driven detection and response capability. Whether you are building a first SOC, maturing an existing one, or evaluating whether an MDR provider is a better fit, this guide provides the practitioner-level framework to make those decisions with clarity.

Before You Build: Defining Scope, Mission, and Sponsorship

The most common reason SOC programs fail is not technology selection or staffing ratios. It is the absence of clear executive sponsorship and a documented scope that defines what the SOC is responsible for monitoring and what it is not. Before a single SIEM instance is provisioned, the SOC mission needs to be articulated in writing: what assets are in scope, what the coverage hours will be, how incidents will be escalated to leadership, and what constitutes success. Without this foundation, the SOC becomes a reactive alert queue without authority to act.

Executive sponsorship means more than nominal support. It means the CISO or CTO has committed budget for multiple fiscal years, has authority to require log source integration from business units that would prefer to opt out, and has aligned the SOC mission with the organization's risk appetite statement. SOCs that lack this authority spend months negotiating log access from application teams and end up with monitoring gaps that attackers exploit. Frame the business case in financial terms: a breach without mature detection costs an average of $1.76 million more than one detected quickly, per IBM research. That number provides a concrete ROI anchor for the investment ask.

Scope decisions are among the most consequential early choices. A SOC that attempts to monitor everything immediately will be overwhelmed with volume and produce poor quality detections. A better approach is to tier the asset inventory by criticality and bring the highest-risk environments into scope first: Active Directory, cloud identity providers, internet-facing applications, endpoint fleet, and the network perimeter. Lower-risk environments can follow once the core detection capability is operational and tuned. Document what is explicitly out of scope so that business units understand the coverage boundary and do not assume that uncovered assets are being monitored.

The coverage model decision, whether to operate 24x7, business hours only, or a hybrid with on-call after hours, is driven by your threat profile and risk tolerance. Organizations in financial services, healthcare, or critical infrastructure that are active ransomware targets cannot accept business-hours-only coverage. Attackers deliberately execute their most destructive actions overnight and on weekends because they know detection is weakest then. Organizations with lower risk profiles may start with business-hours monitoring and on-call escalation, then expand to 24x7 as the SOC matures. Whatever model you choose, document the coverage hours explicitly and communicate them to leadership so that the residual risk of off-hours gaps is an informed decision rather than an undocumented assumption.

The SOC Technology Stack: SIEM, SOAR, EDR, and Beyond

The SIEM (Security Information and Event Management) platform is the technical foundation of any SOC. It aggregates logs from across the environment, normalizes them into a common schema, applies correlation rules to detect patterns that indicate threats, and provides the search and investigation interface that analysts use daily. The leading SIEM platforms each have distinct strengths: Splunk offers unmatched search flexibility and a rich marketplace of integrations but carries high licensing costs; Microsoft Sentinel provides native Azure integration, consumption-based pricing, and strong coverage for Microsoft-centric environments; Elastic SIEM delivers an open-source-derived platform with lower licensing cost but higher operational overhead; IBM QRadar and Chronicle (Google) are well-established in large enterprise and government environments. The choice should be driven by your existing ecosystem, analyst expertise, and total cost of ownership rather than marketing positioning.

What a SIEM ingests is as important as which SIEM you choose. A SOC monitoring only perimeter firewall logs while endpoints and identity systems go dark is a half-built detection capability. The minimum viable log source list for a new SOC includes: endpoint detection and response (EDR) telemetry from the entire endpoint fleet; Windows Event Logs from domain controllers (authentication events, group changes, GPO changes); cloud identity provider logs (Entra ID, Okta, or Google Workspace sign-in and audit logs); network flow logs or firewall logs covering internet egress; email security gateway logs; and VPN or remote access logs. Each of these sources addresses a different phase of the attack lifecycle. Endpoint telemetry catches execution and lateral movement; identity logs catch credential-based initial access and privilege escalation; network logs catch command-and-control and exfiltration.

SOAR (Security Orchestration, Automation, and Response) platforms extend the SIEM by automating repetitive analyst tasks. When a phishing alert fires, a SOAR playbook can automatically pull the email headers, check the sender domain against threat intel feeds, extract URLs and check them against VirusTotal and URLScan, look up the targeted user's role and prior incident history, and present the analyst with a pre-populated case enriched with all relevant context. This reduces investigation time from 30 minutes to 5 minutes and lets Tier 1 analysts handle a much higher alert volume. Leading SOAR platforms include Palo Alto XSOAR, Splunk SOAR, and Microsoft Sentinel Playbooks. Start with automation for the highest-volume alert types (phishing, brute force, impossible travel) before building more complex playbooks.

EDR platforms (CrowdStrike Falcon, SentinelOne, Microsoft Defender for Endpoint) are both a log source and a response tool. Unlike traditional antivirus, EDR records process execution, file writes, network connections, and registry modifications at the endpoint level, providing the forensic detail that SIEM correlation rules need to detect sophisticated attacks. EDR also provides remote response capability: analysts can isolate an endpoint from the network, terminate processes, and collect forensic artifacts directly from the SIEM or SOAR workflow without requiring physical access. If your organization does not have EDR deployed, it should be the first technology investment before SIEM, because the SIEM without endpoint telemetry has a fundamental detection blind spot. Threat intelligence platform integration adds another layer: by ingesting curated threat feeds and contextualizing SIEM alerts against known adversary infrastructure, analysts can quickly distinguish targeted attacks from generic scanning noise.

Free daily briefing

Briefings like this, every morning before 9am.

Threat intel, active CVEs, and campaign alerts, distilled for practitioners. 50,000+ subscribers. No noise.

SOC Staffing: Roles, Tiers, and Team Structure

The SOC analyst tier model reflects the progression from alert triage to advanced investigation and threat hunting. Tier 1 analysts are the front line: they review queued alerts, apply initial triage to determine if an alert is a true positive or false positive, gather basic enrichment (is this IP malicious? has this user had prior incidents?), and escalate confirmed incidents to Tier 2. Tier 1 is a high-volume, high-repetition role that benefits from clear runbooks, automation support, and defined escalation thresholds. Tier 2 investigators take over escalated cases, conduct deeper investigation using SIEM queries and EDR telemetry, determine scope of compromise, and drive the incident response process. Tier 3 is the most senior technical level, responsible for threat hunting (proactively searching for attacker activity that evaded detection), detection engineering (building and tuning detection rules), and handling the most complex incidents.

The Detection Engineer role deserves special attention because it is often undersourced in new SOC builds. Detection engineers own the detection library: they build new SIEM rules, tune existing ones to reduce false positives, map coverage against MITRE ATT&CK, and work with the threat intelligence team to operationalize new intelligence into detection content. A SOC without a dedicated detection engineer gradually accumulates stale, noisy rules and falls behind the threat landscape. Even a part-time allocation from a Tier 3 analyst to detection engineering produces significantly better outcomes than none at all.

Analyst burnout is the primary retention challenge for SOC programs. Alert fatigue from high false positive rates, repetitive triage work without variety, overnight shift schedules, and the psychological weight of high-stakes incident response all contribute to turnover rates that can exceed 50% annually at poorly managed SOCs. Mitigation strategies include aggressive false positive tuning (the single highest-impact intervention), SOAR automation to eliminate the most repetitive tasks, rotation between shifts and between triage and investigation duties, career development pathways with defined progression criteria from Tier 1 to Tier 2, training budgets for certifications (Security+, CySA+, GCIA, GCIH), and explicit recognition of analyst contributions during incidents. The total cost of analyst turnover, including recruiting, onboarding, and productivity ramp, typically exceeds $50,000 per analyst, making retention investment a clear financial priority.

Shift coverage models each have trade-offs. The 8-hour three-shift model provides the most analyst-friendly schedule and cleanest handoffs but requires the most headcount. The 12-hour two-shift model (day and night) reduces staffing requirements but produces fatigue over extended shifts and can complicate handoffs for multi-day incidents. The follow-the-sun model, where analysts in different geographic regions cover their business hours and hand off to the next region, is common in global organizations and provides 24x7 coverage without overnight shifts, but requires strong handoff documentation and communication protocols. Regardless of model, shift overlap periods are critical for complex incident continuity and should be explicitly scheduled.

Detection Engineering: Building Your Detection Library

Detection engineering is the discipline of systematically building, testing, and maintaining the detection rules that allow a SOC to identify attacker activity in telemetry. The starting point for any detection engineering program is the MITRE ATT&CK framework, which documents the tactics, techniques, and procedures used by real-world threat actors across the full attack lifecycle from initial access through impact. By mapping your existing detection rules against ATT&CK, you can visualize which techniques you have coverage for and which represent blind spots. The MITRE ATT&CK Navigator tool makes this visualization straightforward and should be used as a regular artifact in SOC program reporting.

Use-case prioritization is the second foundational decision. You cannot build detection rules for every ATT&CK technique simultaneously. Instead, prioritize based on your threat profile: which threat actors are most likely to target your industry, what initial access techniques they favor, and which assets represent the most valuable targets. A financial institution faces different priority use cases than a healthcare organization or a critical infrastructure operator. Threat intelligence, past incident data, and industry-specific reporting from sources like FS-ISAC, H-ISAC, or sector-specific CISA advisories should drive prioritization decisions.

Detection rules fall into two categories: baseline detections that fire on clearly anomalous activity (impossible travel, new local admin account, PowerShell with Base64 encoding) and behavioral detections that look for patterns across multiple events over time (multiple failed logins followed by a success, lateral movement from a workstation to multiple servers within an hour). Baseline detections are faster to build and produce higher-fidelity alerts with less tuning. Behavioral detections require more development effort and more sophisticated SIEM queries but catch more sophisticated attack techniques that evade signature-based detection.

Sigma rules are a vendor-neutral detection rule format that allows detection content to be written once and compiled into the query language of any major SIEM platform. Building your detection library in Sigma format provides portability if you change SIEM platforms and access to the large community repository of public Sigma rules maintained on GitHub. Alert tuning is the ongoing maintenance task that separates a productive SOC from a noisy one: every new detection rule should be deployed to a test window, observed for false positive rate, and tuned with appropriate whitelists and filters before going into production. A target false positive rate of under 5% per rule is a reasonable starting benchmark; rules exceeding 20% false positive rate should be immediately reviewed and refined.

Runbooks and Playbooks: Operationalizing Response

Runbooks and playbooks are the procedural backbone of a SOC. A runbook is a step-by-step technical procedure for investigating a specific type of alert or incident: it tells the analyst exactly what queries to run, what questions to answer, what enrichment to gather, and what thresholds trigger escalation. A playbook is a higher-level response strategy that defines roles, communication paths, escalation criteria, and decision points for a broader incident type. Together, they ensure that incident response is consistent, documented, and executable by any analyst regardless of experience level, which is critical for shift-based operations where institutional knowledge cannot rely on a single expert being available.

Every SOC needs a baseline set of runbooks covering the most common alert types before going live. The minimum viable runbook set includes: phishing investigation (how to analyze a suspicious email, check links and attachments, assess user exposure, and determine if credentials were compromised); malware alert investigation (how to review EDR telemetry, assess scope, determine if network isolation is warranted); ransomware response (how to identify the initial access vector, determine scope of encryption, initiate containment, and engage the CSIRT); account compromise (how to assess unauthorized activity, reset credentials, review lateral movement, and determine data exposure); and data exfiltration indicators (how to review network flow data, DLP alerts, and cloud access logs to assess what was taken and where it went).

Runbook automation through SOAR transforms these step-by-step procedures from analyst tasks into automated workflows. When the phishing alert fires, the SOAR playbook executes the first fifteen steps of the runbook automatically, presenting the analyst with pre-enriched context rather than starting from raw alert data. This dramatically increases throughput and reduces the skill level required for Tier 1 triage. Start automation with the highest-volume runbooks first and build out the library over time.

Runbook documentation standards matter for maintainability. Each runbook should include a version number, last review date, the analyst tier it is written for, prerequisites (which tools require access), decision tree with clear escalation thresholds, and a post-incident update process so that lessons from real incidents are incorporated. Runbooks that are not regularly reviewed become stale as the environment changes, and stale runbooks produce inconsistent incident handling. Schedule a quarterly review cycle for the runbook library, prioritizing the runbooks that handled the highest volume of incidents in the prior quarter.

SOC Metrics: What to Measure and Report

Effective SOC metrics serve two audiences: the SOC team, which needs operational metrics to identify bottlenecks and quality problems; and leadership, which needs outcome metrics to assess whether the SOC investment is delivering risk reduction. Conflating these two audiences produces reports that satisfy neither. The SOC team needs granular, high-frequency data on alert backlog, false positive rates by rule, analyst throughput, and escalation patterns. Leadership needs trend lines on MTTD, MTTR, incident volume by severity, and qualitative assessments of threat landscape changes.

Mean Time to Detect (MTTD) is the primary detection quality metric: it measures the time from when attacker activity began in the environment to when the SOC identified it as a potential incident. MTTD improvement requires both better detection coverage (fewer blind spots) and faster analyst triage (lower alert queue depth). Mean Time to Respond (MTTR) measures from initial detection to containment, which is a function of both SOC efficiency and the organization's incident response process maturity. Both metrics should be trended over time and segmented by incident type, because phishing investigations and ransomware responses have fundamentally different time profiles.

False positive rate is the metric with the most direct impact on analyst capacity and burnout. A SOC where 80% of alerts are false positives is not operating at 80% efficiency; it is operating at 20% efficiency while generating 80% of the analyst workload that produces no outcome. False positive rate should be tracked per detection rule and reviewed weekly, with rules exceeding threshold immediately triaged for tuning. Alert backlog, the volume of alerts older than your SLA threshold that remain unreviewed, is the leading indicator of a SOC capacity problem and should trigger staffing or automation conversations before the backlog becomes unmanageable.

For quarterly business reviews with leadership, present metrics in the context of the risk they represent rather than as operational statistics. A reduction in MTTD from 48 hours to 6 hours over the quarter should be presented as reducing the window available for lateral movement and data exfiltration, not as an abstract improvement in an operations number. Connect the metrics to the incidents that were detected and their potential business impact if not caught. This framing demonstrates SOC value in terms that resonate with executives who are weighing the program investment against other business priorities.

SOC Maturity Model: Crawl, Walk, Run

SOC maturity models provide a framework for assessing current state, identifying gaps, and building a roadmap for improvement. The most widely used framework describes five maturity levels, each representing a distinct capability threshold. Level 1 is reactive and ad hoc: there is monitoring of some kind, but it is inconsistent, there are no documented processes, alert review is sporadic, and response depends on individual heroics rather than defined procedures. Most organizations at this level are not aware of threats until they cause visible damage. Level 2 is defined and repeatable: a SIEM is deployed, core log sources are ingested, there are documented runbooks for common alert types, and incidents are tracked in a ticketing system. This is the foundational operational state that every SOC should reach in its first 6 to 12 months.

Level 3 is proactive: detection rules are tuned to reduce false positives, coverage is mapped against MITRE ATT&CK with documented gaps, threat intelligence is integrated into alert enrichment, and SOAR automation handles high-volume triage tasks. SOCs at Level 3 are detecting most commodity threats and some targeted attack techniques. They have defined escalation paths, consistent analyst training, and metrics programs that track MTTD and MTTR. This is a realistic 12 to 24-month target for a new SOC with adequate investment. Level 4 introduces threat hunting: analysts proactively search for attacker activity that evaded automated detection, detection content is continuously improved based on hunt findings, and purple team exercises are used to validate detection coverage against specific threat actor techniques.

Level 5 is intelligence-driven and continuously improving: the SOC has a dedicated threat intelligence function that produces finished intelligence tailored to the organization's threat profile, detection engineering is a mature capability with a library of custom detection content tuned to the environment, and every major incident produces documented lessons learned that drive program improvement. Few organizations outside of financial services, large technology companies, and defense contractors operate at Level 5. The realistic progression for most organizations is 12 to 18 months to reach Level 2, another 12 to 18 months to reach Level 3, and then several years of sustained investment to reach Levels 4 and 5.

Building a maturity roadmap starts with an honest assessment of current state. Assess across five dimensions: people (staffing, training, retention), process (runbooks, escalation procedures, review cadence), technology (log coverage, SIEM quality, automation), detection quality (coverage against ATT&CK, false positive rates, MTTD), and program governance (metrics, reporting, budget planning). Score each dimension against the maturity level criteria, identify the lowest-scoring areas as priorities, and build a 12-month improvement plan with specific, measurable milestones. Revisit the assessment semi-annually to track progress and update priorities as the threat landscape and organizational environment evolve.

Level 1: Ad Hoc

Reactive monitoring with no documented processes. Alert review is sporadic and response depends on individual knowledge rather than defined procedures.

Level 2: Defined

SIEM deployed with core log sources ingested. Documented runbooks for common alert types. Incidents tracked in a ticketing system with consistent escalation paths.

Level 3: Proactive

Detection rules tuned for low false positive rates. ATT&CK coverage mapped with documented gaps. Threat intelligence integrated. SOAR automation deployed for high-volume alert types.

Level 4: Threat Hunting

Analysts proactively hunt for attacker activity that evaded automated detection. Purple team exercises validate detection coverage. Detection content continuously improved from hunt findings.

Level 5: Intelligence-Driven

Dedicated threat intelligence function produces finished intelligence tailored to the organization's threat profile. Every incident drives documented program improvement. Detection library continuously updated.

The bottom line

Building a SOC is a multi-year investment that must start with process and coverage before pursuing sophistication. A SOC with thorough log coverage, low false positive rates, and consistent runbook execution will outperform a more technically advanced SOC with poor data quality and ad hoc processes. For most organizations, the right starting point is either an MDR provider for 24x7 coverage or a hybrid model where MDR handles triage while an internal team builds detection content and handles escalations. This approach gets defensive capability operational quickly while internal expertise develops.

The goal is not a perfect SOC on day one. It is a SOC that improves measurably every quarter. Track MTTD, MTTR, and false positive rate as trend lines, run a purple team exercise annually to validate coverage, and review the maturity model assessment semi-annually to keep the roadmap aligned with where the program actually is. Organizations that approach SOC development with this disciplined, iterative mindset consistently outperform those that treat it as a one-time technology deployment.

Frequently asked questions

How much does it cost to build an internal SOC?

The cost of building an internal SOC varies significantly based on coverage model, staffing levels, and technology choices. A minimal business-hours SOC with two to three analysts, a mid-tier SIEM, and basic tooling typically runs between $800,000 and $1.5 million annually when fully loaded salaries, benefits, training, and licensing are accounted for. A 24x7 SOC with eight to twelve analysts across three shifts, a mature SIEM such as Splunk or Microsoft Sentinel with full data ingest, EDR, SOAR, and threat intel platform integration, and dedicated detection engineering resources can exceed $3 to $5 million annually. These figures do not include the one-time capital costs for infrastructure, integration work, and initial program buildout, which often add another $200,000 to $500,000. For most organizations under 2,000 employees, the economics of an internal SOC are difficult to justify compared to an MDR provider or hybrid model.

Should I build an internal SOC or use an MDR provider?

The decision depends on your organization's size, regulatory environment, risk tolerance, and internal security maturity. MDR providers offer 24x7 coverage, pre-built detection content, and access to a large analyst team at a fraction of the cost of building internally, typically $50,000 to $250,000 annually depending on endpoint count and scope. The trade-off is reduced control over detection logic, dependency on the provider's tooling, and lower organizational learning that comes from having analysts embedded in your environment. Internal SOCs provide greater customization, deeper institutional knowledge, and the ability to build detection content tailored to your specific environment. A hybrid model, where an MDR provider handles 24x7 alert triage while an internal detection engineering team builds and maintains detection content and handles escalations, often delivers the best outcome for mid-market organizations. Large enterprises with complex environments, regulated industries with data residency requirements, and organizations with unique threat profiles are the strongest candidates for fully internal SOC investment.

What SIEM should a new SOC start with?

The right SIEM depends on your existing technology ecosystem, data volume, in-house expertise, and budget. Organizations already in the Microsoft ecosystem are well-served by Microsoft Sentinel, which offers native integration with Entra ID, Defender, and Azure services, consumption-based pricing, and a large library of community detection rules. Splunk remains the market leader for organizations with complex, heterogeneous environments and teams that value its search language and marketplace of integrations, but licensing costs are high. Elastic SIEM provides a strong open-source option with lower licensing cost but higher operational complexity. IBM QRadar and Securonix are well-established in large enterprise and MSSP environments. For a new SOC with a limited budget and a primarily Microsoft or AWS environment, Microsoft Sentinel or the Elastic stack are typically the most cost-effective starting points. Regardless of which SIEM you choose, the most important factors are ensuring all critical log sources are ingested and normalized, and that analysts are trained to write and tune detection rules in the platform's query language.

How many analysts do I need for a 24x7 SOC?

A 24x7 SOC requires a minimum of eight analysts to cover three shifts (three analysts per shift with one for overlap and coverage of vacations and sick leave), though most practitioners recommend ten to twelve to provide sustainable coverage without analyst burnout. This assumes an 8-hour shift model with three shifts per day, seven days per week. With a 12-hour shift model, you can reduce headcount to six to eight analysts but at the cost of longer shifts that increase fatigue. These numbers are for Tier 1 analysts only. A complete SOC also requires Tier 2 investigators (two to three), a Tier 3 threat hunter or detection engineer (one to two), a SOC Manager, and a Threat Intelligence analyst, bringing a fully staffed 24x7 SOC to fifteen to twenty people. Organizations that cannot staff to this level should consider an MDR provider for coverage and staff internal analysts for daytime investigation, escalation, and detection engineering.

What are the first 10 detection rules every SOC should have?

The highest-priority detection rules for a new SOC are: impossible travel logins (authentication from geographically separated locations within a time window that is physically impossible); brute force and password spray attacks against Active Directory or cloud identity providers; new local administrator account creation outside of standard provisioning; outbound connections to known malicious IPs and domains via threat intelligence feed integration; PowerShell execution with encoded command-line arguments (Base64 encoded payloads); scheduled task and service creation by non-standard accounts; lateral movement via PsExec, WMI, or SMB; large volumes of files accessed or modified in a short period (ransomware indicator); email forwarding rules created to external addresses; and new accounts added to privileged groups such as Domain Admins or Global Administrators. These ten rules address the most common initial access, persistence, lateral movement, and impact techniques seen in real-world incidents and provide immediate detection value without requiring extensive tuning.

How do I measure SOC effectiveness?

SOC effectiveness is measured through a combination of detection quality metrics and operational efficiency metrics. The primary metrics are Mean Time to Detect (MTTD), which measures how long it takes from threat activity to SOC awareness; Mean Time to Respond (MTTR), which measures how long from detection to containment; and false positive rate, which measures the percentage of alerts that are not genuine threats and directly impacts analyst capacity. Secondary metrics include escalation rate (percentage of Tier 1 alerts escalated to Tier 2), alert backlog (unreviewed alerts older than SLA threshold), and dwell time (how long a threat was present before detection). For leadership reporting, frame these as trend lines rather than point-in-time figures. A SOC that reduces MTTD from 72 hours to 8 hours over a quarter demonstrates clear value even if the absolute numbers are still high. Purple team exercises and tabletop simulations also provide qualitative effectiveness data by testing whether specific attack scenarios would be detected by current detection content.

What is the difference between a SOC and a CSIRT/CIRT?

A SOC (Security Operations Center) is a continuous monitoring and detection function that operates 24x7, triaging alerts and investigating potential incidents as they are detected. A CSIRT (Computer Security Incident Response Team) or CIRT (Computer Incident Response Team) is a defined team and process activated in response to a declared security incident. In practice, many organizations treat these as overlapping functions: the SOC detects and triages, and when an incident is declared, the CSIRT process kicks in to coordinate response, communication, legal, and recovery activities. In smaller organizations, the same people play both roles. In larger organizations, the SOC operates continuously with defined escalation paths to a separate CSIRT that includes legal counsel, communications, executive leadership, and external forensics resources. The distinction matters most during major incidents, where the SOC continues monitoring for ongoing activity while the CSIRT focuses on containment and recovery. NIST SP 800-61 defines the incident response lifecycle (preparation, detection, containment, eradication, recovery, lessons learned) that governs CSIRT operations.

Sources & references

  1. NIST SP 800-61: Computer Security Incident Handling Guide
  2. SANS Institute: Building a World-Class Security Operations Center
  3. Microsoft Sentinel SOC Guide
  4. CIS Controls v8
  5. Gartner: SOC Model Guide

Free resources

25
Free download

Critical CVE Reference Card 2025–2026

25 actively exploited vulnerabilities with CVSS scores, exploit status, and patch availability. Print it, pin it, share it with your SOC team.

No spam. Unsubscribe anytime.

Free download

Ransomware Incident Response Playbook

Step-by-step 24-hour IR checklist covering detection, containment, eradication, and recovery. Built for SOC teams, IR leads, and CISOs.

No spam. Unsubscribe anytime.

Free newsletter

Get threat intel before your inbox does.

50,000+ security professionals read Decryption Digest for early warnings on zero-days, ransomware, and nation-state campaigns. Free, weekly, no spam.

Unsubscribe anytime. We never sell your data.

Eric Bang
Author

Founder & Cybersecurity Evangelist, Decryption Digest

Cybersecurity professional with expertise in threat intelligence, vulnerability research, and enterprise security. Covers zero-days, ransomware, and nation-state operations for 50,000+ security professionals weekly.

Free Brief

The Mythos Brief is free.

AI that finds 27-year-old zero-days. What it means for your security program.

Joins Decryption Digest. Unsubscribe anytime.

Daily Briefing

Get briefings like this every morning

Actionable threat intelligence for working practitioners. Free. No spam. Trusted by 50,000+ SOC analysts, CISOs, and security engineers.

Unsubscribe anytime.

Mythos Brief

Anthropic's AI finds zero-days your scanners miss.