November 2026disaster recovery plan

Disaster Recovery Plan for SaaS Companies: How to Build a DR Strategy That Passes Compliance Audits

Build a disaster recovery plan for SaaS companies that meets SOC 2, ISO 27001, and HIPAA requirements. Covers RPO/RTO, cloud DR strategies, failover architecture, and testing procedures.

By QuickTrust EditorialUpdated 2026-03-19

Disaster Recovery Plan for SaaS Companies: How to Build a DR Strategy That Passes Compliance Audits

In November 2024, a misconfigured database migration at a mid-market SaaS company brought down its production environment for 38 hours. The engineering team had backups -- daily snapshots stored in the same AWS region as the production database. But when the entire availability zone experienced degraded performance during the recovery attempt, those backups were unreachable for 14 hours. The company had no documented recovery procedures, no predefined failover target, and no tested process for restoring from cross-region replicas that, it turned out, had never been configured.

The financial damage was significant: $2.3 million in SLA credits, three enterprise customers initiated contract reviews, and the subsequent SOC 2 audit produced a qualified opinion citing the absence of a tested disaster recovery plan. But the most revealing detail was this -- the company had a disaster recovery plan. It was a 12-page document written during their Series A, reviewed once, and never tested. It referenced an AWS architecture that no longer existed and named a recovery team lead who had left 18 months earlier.

This is the pattern that repeats across the SaaS industry. The disaster recovery plan exists on paper but has never been validated against reality. When a real disruption occurs, the plan is useless. When auditors arrive, the plan is indefensible.

This guide walks through how to build a disaster recovery plan that does both things: it actually recovers your systems during a real disruption, and it withstands the scrutiny of SOC 2, ISO 27001, HIPAA, and PCI DSS auditors. The two objectives are not in conflict. A DR plan that works in practice produces the documentation and evidence that auditors need. A DR plan that satisfies auditors but has never been tested is a liability waiting to surface.


What Is a Disaster Recovery Plan?

A disaster recovery plan is a documented, structured set of procedures for restoring IT systems, applications, and data after a disruption that renders them unavailable. The scope of a DR plan covers the technical recovery of infrastructure and services -- it defines what gets restored, in what order, to what state, and within what timeframe.

The disruptions a DR plan addresses include:

  • Infrastructure failures. Cloud provider outages, availability zone failures, hardware failures, network partitions.
  • Data corruption or loss. Database corruption from failed migrations, accidental deletion, ransomware encryption, storage failures.
  • Application failures. Deployment errors that crash production, cascading service failures, dependency outages.
  • Natural disasters. Events that affect physical infrastructure -- relevant even for cloud-native companies when single-region architectures are in use.
  • Cyber attacks. Ransomware, destructive malware, attacks targeting data integrity or availability.

A disaster recovery plan is not a general emergency response document. It is a technical plan with specific procedures, named systems, defined recovery targets, and tested runbooks. It answers the question: when a critical system goes down, what exactly do we do to bring it back?

What a DR plan contains

At a high level, a complete disaster recovery plan includes:

  • An inventory of systems and data classified by criticality
  • Defined Recovery Point Objectives (RPO) and Recovery Time Objectives (RTO) for each system tier
  • Backup strategies and backup validation procedures
  • Recovery procedures documented as step-by-step runbooks
  • Failover architecture and infrastructure requirements
  • Roles and responsibilities for DR execution
  • Communication procedures during a DR event
  • Testing schedules, test types, and documentation requirements
  • Maintenance and review cadence

Each of these components will be addressed in detail in the sections that follow.


Disaster Recovery vs. Business Continuity: How They Work Together

Disaster recovery and business continuity are related but distinct disciplines. Conflating them -- or treating one as a substitute for the other -- is a common audit finding and a common operational mistake.

Business continuity planning (BCP) is the broader discipline. It addresses how the entire organization continues to operate during and after a disruption. BCP covers people, processes, facilities, communications, and vendor dependencies in addition to technology. A business continuity plan answers: how do we keep the business running?

Disaster recovery (DR) is a subset of business continuity. It focuses specifically on restoring IT systems, applications, and data. A disaster recovery plan answers: how do we get our technology back?

The relationship between the two:

DimensionBusiness Continuity PlanDisaster Recovery Plan
ScopeEntire organization -- people, processes, technology, facilitiesIT systems, applications, data
FocusMaintaining business operationsRestoring technology infrastructure
Key questionsHow do employees work? How do we serve customers? How do we communicate?How do we restore databases? How do we failover to a secondary region? How do we recover from backup?
OwnsExecutive leadership / COO / Head of OperationsCTO / VP Engineering / Infrastructure Lead
TimeframeFrom disruption through full recovery and normalizationFrom system failure through system restoration
Compliance mappingSOC 2 A1.2, ISO 27001 A.5.29-A.5.30, HIPAA 164.308(a)(7)SOC 2 A1.2/A1.3, ISO 27001 A.5.29/A.8.13-A.8.14, HIPAA 164.308(a)(7)(ii)(B-D), PCI DSS 12.10.1

For SaaS companies, the disaster recovery plan is the most technically detailed component of the broader business continuity program. BCP sets the business-level requirements -- the maximum tolerable downtime for each business function, the minimum service levels customers must receive during a disruption, the communication cadence with stakeholders. The DR plan translates those business requirements into technical recovery procedures.

Every compliance framework requires both. Auditors will check that your DR plan is connected to your business continuity plan -- that recovery time objectives in the DR plan align with maximum tolerable downtime in the BCP, and that the two documents reference each other.


Why Compliance Frameworks Require DR Plans

Disaster recovery is not a discretionary control in any major compliance framework. Every framework that SaaS companies commonly pursue includes explicit requirements for documented, tested DR capabilities. Here is exactly what each framework demands.

SOC 2 -- Availability Criteria (A1.2 and A1.3)

If the Availability trust service criterion is in scope for your SOC 2 report -- and for SaaS companies, it almost always is -- two criteria directly address disaster recovery:

  • A1.2 -- Recovery planning. The entity authorizes, designs, develops, implements, operates, approves, maintains, and monitors environmental protections, software, data backup processes, and recovery infrastructure to meet its objectives. This requires a documented DR plan with defined recovery objectives.
  • A1.3 -- Recovery testing. The entity tests recovery plan procedures supporting system recovery to meet its objectives. This means auditors will request evidence of DR testing -- not just the plan, but proof that you have validated the plan works.

SOC 2 auditors will specifically ask for: the documented DR plan, defined RPO and RTO values, evidence of backup procedures, evidence of DR testing (test results, test dates, findings), and evidence that testing gaps were remediated.

ISO 27001 -- Annex A Controls

ISO 27001:2022 addresses disaster recovery through several Annex A controls:

  • A.5.29 -- Information security during disruption. The organization shall plan how to maintain information security at an appropriate level during disruption. This requires that your DR plan addresses not just system recovery but maintaining security controls during the recovery process.
  • A.5.30 -- ICT readiness for business continuity. ICT readiness shall be planned, implemented, maintained, and tested based on business continuity objectives. This is the direct DR requirement -- your IT systems must be recoverable within defined targets, and that recoverability must be tested.
  • A.8.13 -- Information backup. Backup copies of information, software, and systems shall be maintained and regularly tested in accordance with the agreed topic-specific policy on backup.
  • A.8.14 -- Redundancy of information processing facilities. Information processing facilities shall be implemented with sufficient redundancy to meet availability requirements.

ISO 27001 certification auditors will examine whether your DR plan is integrated into your ISMS, whether it is based on a risk assessment, whether recovery objectives are tied to business impact analysis, and whether testing evidence exists.

HIPAA -- Security Rule Section 164.308(a)(7)

HIPAA's Contingency Plan standard is one of the most prescriptive DR requirements of any framework:

  • 164.308(a)(7)(i) -- Contingency Plan. Establish (and implement as needed) policies and procedures for responding to an emergency or other occurrence that damages systems containing ePHI.
  • 164.308(a)(7)(ii)(A) -- Data Backup Plan (Required). Establish and implement procedures to create and maintain retrievable exact copies of ePHI.
  • 164.308(a)(7)(ii)(B) -- Disaster Recovery Plan (Required). Establish (and implement as needed) procedures to restore any loss of data.
  • 164.308(a)(7)(ii)(C) -- Emergency Mode Operation Plan (Required). Establish (and implement as needed) procedures to enable continuation of critical business processes for protection of the security of ePHI while operating in emergency mode.
  • 164.308(a)(7)(ii)(D) -- Testing and Revision Procedures (Addressable). Implement procedures for periodic testing and revision of contingency plans.
  • 164.308(a)(7)(ii)(E) -- Applications and Data Criticality Analysis (Addressable). Assess the relative criticality of specific applications and data in support of other contingency plan components.

Note that "addressable" under HIPAA does not mean optional. It means you must implement the specification or document why an equivalent alternative measure is reasonable and appropriate. For SaaS companies handling ePHI, DR testing is effectively mandatory.

For more on HIPAA requirements, see our HIPAA Compliance Guide.

PCI DSS -- Requirement 12.10

PCI DSS v4.0 Requirement 12.10.1 explicitly requires that the incident response plan include "business recovery and continuity procedures" and "data backup processes." While PCI DSS addresses DR within the broader incident response requirement rather than as a standalone requirement, assessors will verify that:

  • Backup procedures for cardholder data environments are documented and tested
  • Recovery procedures exist for systems within the cardholder data environment
  • Recovery time targets are defined and achievable
  • Testing has been performed at least annually

The common thread

Across all four frameworks, the requirements converge on the same three elements: (1) a documented disaster recovery plan with defined recovery objectives, (2) implemented backup and recovery procedures that support those objectives, and (3) evidence that the plan has been tested. A DR plan that is documented but untested will fail every framework. A DR capability that works but is undocumented will fail every audit.


RPO and RTO: Defining Your Recovery Objectives

RPO and RTO are the two metrics that drive every technical decision in your disaster recovery plan. They define what you can afford to lose and how long you can afford to be down. Getting them wrong -- or failing to define them at all -- is one of the most common DR plan audit failures.

Recovery Point Objective (RPO)

RPO defines the maximum acceptable amount of data loss measured in time. An RPO of 1 hour means you can tolerate losing up to 1 hour of data. If your last backup was taken at 2:00 PM and a failure occurs at 2:45 PM, you lose 45 minutes of data -- within your 1-hour RPO. If the failure occurs at 3:30 PM, you lose 1.5 hours of data -- an RPO violation.

RPO determines your backup frequency and replication strategy. A 24-hour RPO can be met with daily backups. A 1-hour RPO requires at least hourly backups or continuous replication. A near-zero RPO requires synchronous replication.

Recovery Time Objective (RTO)

RTO defines the maximum acceptable duration of downtime from the moment a disruption occurs until the system is restored to operational status. An RTO of 4 hours means the system must be back online within 4 hours of the failure.

RTO determines your recovery architecture. A 24-hour RTO can be met with manual restoration from backups. A 4-hour RTO typically requires pre-provisioned infrastructure and automated recovery procedures. A sub-1-hour RTO requires hot standby or active-active architecture.

Defining RPO and RTO by system tier

Not every system requires the same recovery objectives. Applying the same RPO and RTO to every system is both unnecessarily expensive and a sign of an immature DR program. Instead, classify systems into tiers based on business impact and assign recovery objectives accordingly.

TierDescriptionExamplesTypical RPOTypical RTODR Strategy
Tier 1 -- Mission CriticalSystems whose failure immediately halts revenue generation or customer operationsProduction application, primary database, authentication service, payment processing< 15 minutes< 1 hourHot standby / active-active, synchronous replication
Tier 2 -- Business CriticalSystems whose extended outage significantly impacts operations but does not immediately halt revenueInternal APIs, background job processing, analytics pipelines, notification services1 hour4 hoursWarm standby, asynchronous replication
Tier 3 -- ImportantSystems that support business operations but whose outage is tolerable for a defined periodStaging environments, internal tools, reporting dashboards, dev environments24 hours24 hoursPilot light or backup/restore
Tier 4 -- Non-CriticalSystems whose outage has minimal immediate business impactDocumentation sites, internal wikis, sandbox environments72 hours72 hoursBackup/restore only

These values are starting points. Your specific RPO and RTO targets must be derived from a business impact analysis (BIA) that assesses the financial, operational, legal, and reputational consequences of system unavailability at each time interval. The BIA is also a requirement for ISO 27001 (A.5.30) and HIPAA (164.308(a)(7)(ii)(E)).

What auditors check

Auditors verify three things about your RPO/RTO:

  1. Defined and documented. RPO and RTO are explicitly stated in the DR plan for each system or system tier.
  2. Justified. The values are based on a business impact analysis, not arbitrary numbers.
  3. Achievable. DR testing results demonstrate that the organization can actually meet the stated RPO and RTO targets. If your DR plan states a 1-hour RTO but your last DR test took 6 hours to complete recovery, auditors will flag the gap.

DR Strategies for Cloud-Native and SaaS Companies

Cloud infrastructure offers DR capabilities that were cost-prohibitive in traditional data center environments. But those capabilities must be intentionally architected -- they do not exist by default. Deploying to AWS, GCP, or Azure does not make your application disaster-recoverable any more than it makes your application secure.

Here are the five primary DR strategies, ordered from least to most resilient and from lowest to highest cost.

Strategy 1: Backup and Restore

How it works: Data is backed up regularly (snapshots, database dumps, file-level backups) and stored in a separate location. When a disaster occurs, infrastructure is provisioned from scratch in the recovery location, and data is restored from backups.

RPO: Depends on backup frequency -- typically 1 to 24 hours. RTO: Hours to days, depending on the size of the environment and the complexity of the restoration process. Cost: Lowest. You pay only for backup storage until a disaster occurs. Best for: Tier 3 and Tier 4 systems.

Limitation: This is the slowest recovery strategy. Provisioning infrastructure, restoring data, configuring networking, and validating the environment takes time. For production SaaS applications, this strategy rarely meets customer SLA expectations.

Strategy 2: Pilot Light

How it works: The core infrastructure components -- databases, directory services, domain controllers -- are kept running in the recovery region at minimal capacity. When a disaster occurs, application servers, load balancers, and other compute resources are provisioned and scaled up around the already-running core.

RPO: Minutes to 1 hour, depending on replication configuration. RTO: 1 to 4 hours. Cost: Low to moderate. You pay for the always-on core infrastructure plus storage replication costs. Best for: Tier 2 and Tier 3 systems where some recovery time is acceptable.

Strategy 3: Warm Standby

How it works: A scaled-down but fully functional copy of the production environment runs continuously in the recovery region. Data is replicated asynchronously. When a disaster occurs, the standby environment is scaled up to handle production traffic and DNS or load balancer routing is switched.

RPO: Minutes, depending on replication lag. RTO: 30 minutes to 2 hours. Cost: Moderate. You pay for a continuously running environment, though at reduced scale. Best for: Tier 1 and Tier 2 systems where recovery time must be measured in minutes, not hours.

Strategy 4: Hot Standby (Active-Passive)

How it works: A fully provisioned, production-scale copy of the environment runs continuously in the recovery region. Data is replicated synchronously or with near-zero lag. When a disaster occurs, traffic is routed to the standby environment via DNS failover or global load balancing.

RPO: Near-zero (seconds of data loss, depending on replication mode). RTO: Minutes. Cost: High. You effectively pay for two production environments. Best for: Tier 1 mission-critical systems where any material downtime is unacceptable.

Strategy 5: Active-Active (Multi-Region)

How it works: The application runs simultaneously in two or more regions, each serving production traffic. Data is replicated bidirectionally. If one region fails, the other regions absorb the traffic with no manual intervention required.

RPO: Zero (no data loss if using synchronous replication or conflict-free replicated data types). RTO: Seconds (limited to DNS TTL or health check intervals). Cost: Highest. Requires application-level support for multi-region operation, including data consistency management, conflict resolution, and cross-region latency handling. Best for: Tier 1 systems at companies where the cost of downtime exceeds the cost of multi-region infrastructure.

Multi-AZ vs. Multi-Region

A critical distinction for cloud-native DR planning:

  • Multi-AZ (Availability Zone) deploys across multiple data centers within the same geographic region. It protects against single data center failures but does not protect against region-wide outages. AWS RDS Multi-AZ, GCP regional persistent disks, and Azure zone-redundant storage all operate at this level.
  • Multi-region deploys across geographically separate regions. It protects against region-wide outages but introduces complexity around data replication, latency, and consistency.

For most SaaS companies pursuing compliance certifications, multi-AZ is the minimum for Tier 1 and Tier 2 systems. Multi-region is required when customers demand it contractually, when regulatory requirements mandate geographic redundancy, or when the cost of a region-wide outage exceeds the cost of multi-region architecture.

For a detailed mapping of cloud security controls to compliance frameworks, see our Cloud Security Compliance Guide.


Building Your DR Plan: Step-by-Step

This section provides an eight-step process for building a disaster recovery plan from scratch or rebuilding one that would not survive an audit.

Step 1: Conduct a Business Impact Analysis (BIA)

The BIA is the foundation of your DR plan. It identifies which systems and data are critical to business operations, what the impact of their unavailability is at each time interval, and what the maximum tolerable downtime and data loss are for each.

What the BIA should document:

  • Every business-critical application and system
  • The business functions each system supports
  • The financial impact of downtime per hour (lost revenue, SLA penalties, productivity loss)
  • The regulatory impact (notification obligations, compliance violations)
  • The reputational impact (customer trust, market perception)
  • Maximum Tolerable Period of Disruption (MTPD) for each business function
  • Dependencies between systems (a system may be Tier 3 on its own but Tier 1 as a dependency of a mission-critical system)

Step 2: Classify Systems and Define RPO/RTO

Based on the BIA, classify every in-scope system into tiers and assign RPO and RTO values. Document the classification rationale. Ensure that RPO/RTO targets are signed off by business stakeholders, not just the engineering team -- recovery objectives are business decisions, not technical ones.

Step 3: Select DR Strategies by Tier

Map each system tier to an appropriate DR strategy (backup/restore, pilot light, warm standby, hot standby, or active-active) based on the RPO/RTO requirements and cost constraints. Document the strategy selection rationale, including why more resilient (and more expensive) strategies were not selected for lower-tier systems.

Step 4: Design and Implement the Recovery Architecture

Translate the selected strategies into actual infrastructure. This includes:

  • Configuring cross-region or cross-AZ replication for databases and storage
  • Provisioning recovery infrastructure (or infrastructure-as-code templates for on-demand provisioning)
  • Configuring DNS failover or global load balancing
  • Setting up monitoring and alerting for replication lag, backup completion, and recovery environment health
  • Documenting network configurations, security group rules, and IAM roles in the recovery environment

Step 5: Implement and Validate Backup Procedures

Backups are the foundation of every DR strategy. Implement backup procedures for each system and validate them:

  • Configure automated backups at the frequency required to meet RPO targets
  • Store backups in a separate region or account from production
  • Encrypt all backups at rest and in transit
  • Implement backup monitoring with alerts for failed or missed backups
  • Validate backup integrity through regular test restores (not just checking that the backup job completed -- actually restore from the backup and verify data integrity)

Step 6: Write Recovery Runbooks

For each system or system tier, write a detailed, step-by-step recovery runbook. This is covered in detail in the DR Runbooks section below. The runbook must be specific enough that someone who did not design the system can execute the recovery.

Step 7: Test the DR Plan

Testing is not optional. Every compliance framework requires it, and untested DR plans fail when they are needed. DR testing types, frequency, and documentation are covered in the DR Testing section below.

Step 8: Establish Maintenance and Review Cadence

A DR plan is a living document. Establish a review cadence:

  • Quarterly: Review contact lists, verify backup completion rates, check replication health
  • After major infrastructure changes: Update the DR plan whenever significant architecture changes occur (new databases, new services, region migrations, provider changes)
  • After DR tests: Update runbooks based on test findings
  • Annually: Full DR plan review including BIA reassessment, RPO/RTO validation, strategy review, and management sign-off

Document every review, even if no changes are made. The evidence of review is what auditors need.


DR Architecture Patterns for AWS, GCP, and Azure

Each major cloud provider offers native services that support DR architecture. The specific service names differ, but the patterns are consistent.

AWS

DR ComponentAWS Services
Database replicationRDS Multi-AZ (automatic failover within a region), RDS Cross-Region Read Replicas, Aurora Global Database (sub-second replication across regions)
Object storage replicationS3 Cross-Region Replication (CRR), S3 Same-Region Replication (SRR)
Block storage snapshotsEBS Snapshots (cross-region copy), AWS Backup (centralized backup management across services)
Infrastructure as CodeCloudFormation StackSets (deploy recovery infrastructure across regions), Terraform with multi-region providers
DNS failoverRoute 53 health checks with failover routing policy
Global load balancingAWS Global Accelerator, CloudFront with origin failover
DR orchestrationAWS Elastic Disaster Recovery (automated failover and failback for EC2 workloads)

GCP

DR ComponentGCP Services
Database replicationCloud SQL cross-region replicas, Cloud Spanner (globally distributed, synchronous replication by design), AlloyDB cross-region replication
Object storage replicationCloud Storage dual-region and multi-region buckets (automatic replication), Turbo Replication (15-minute RPO for dual-region)
Block storage snapshotsPersistent Disk snapshots (multi-regional), Backup and DR Service
Infrastructure as CodeDeployment Manager, Terraform with multi-region configuration
DNS failoverCloud DNS routing policies with health checks
Global load balancingCloud Load Balancing (global, single anycast IP, automatic failover across regions)

Azure

DR ComponentAzure Services
Database replicationAzure SQL Database geo-replication (active geo-replication, failover groups), Cosmos DB multi-region writes
Object storage replicationAzure Blob Storage geo-redundant storage (GRS), read-access geo-redundant storage (RA-GRS)
Block storage snapshotsAzure Managed Disk snapshots, Azure Backup
Infrastructure as CodeARM Templates, Bicep, Terraform with multi-region providers
DNS failoverAzure Traffic Manager (DNS-based global load balancing with health probes)
Global load balancingAzure Front Door (layer 7 global load balancing with instant failover)
DR orchestrationAzure Site Recovery (automated replication and failover for VMs and physical servers)

Regardless of cloud provider, the principle is the same: DR architecture must be intentionally designed, implemented, and tested. Managed services reduce operational burden but do not eliminate the need for a documented, validated recovery plan.


Data Backup Strategies: The 3-2-1-1-0 Rule

The traditional 3-2-1 backup rule (3 copies, 2 different media types, 1 offsite) has evolved for cloud-native environments. The modern standard is the 3-2-1-1-0 rule:

  • 3 copies of your data (production plus two backups)
  • 2 different storage types (e.g., block storage snapshots plus object storage exports, or managed database backups plus logical dumps to a separate storage service)
  • 1 copy offsite (in a different region or a different cloud provider account)
  • 1 copy offline or immutable (air-gapped or stored in a write-once, read-many format that cannot be modified or deleted -- critical for ransomware protection)
  • 0 errors (backup integrity is verified through regular test restores, not just monitoring backup job completion status)

Why immutable backups matter

Ransomware attacks increasingly target backup systems. If an attacker gains access to your environment, they will look for backups and attempt to encrypt or delete them before deploying ransomware to production. If your backups are stored in the same account with the same credentials, the attacker can destroy your recovery capability.

Immutable backups mitigate this risk:

  • AWS: S3 Object Lock (Governance or Compliance mode), AWS Backup Vault Lock
  • GCP: Cloud Storage Bucket Lock, retention policies with the locked flag
  • Azure: Blob Storage immutability policies (time-based retention or legal hold)

Backup validation

A backup that has never been restored is a hypothesis, not a backup. Implement regular backup validation:

  • Automated restore tests: Schedule automated restore operations on a weekly or monthly cadence. Spin up a test environment, restore from the most recent backup, run validation queries against the restored data, and tear down the test environment.
  • Integrity checks: Verify checksums of backup files. Compare row counts, record hashes, or data fingerprints between production and the restored backup.
  • Document results: Every backup validation test should produce a dated record with the backup source, restore target, validation method, results, and any discrepancies found. This documentation is audit evidence.

DR Testing: Types, Frequency, and Documentation

A disaster recovery plan that has never been tested is indistinguishable from a plan that does not work. Every compliance framework requires DR testing, and auditors will request evidence not just that testing occurred but that it was meaningful, that findings were documented, and that gaps were remediated.

Test types

There are three primary types of DR tests, each with increasing scope and rigor.

1. Tabletop Exercise

A discussion-based walkthrough where the DR team reviews the recovery procedures for a specific failure scenario without executing any technical actions. The team works through the runbook step by step, identifies gaps in documentation, discusses decision points, and validates that roles and responsibilities are clear.

When to use: Quarterly, or when significant changes to infrastructure or personnel have occurred. Duration: 1 to 3 hours. Risk: None. No production systems are affected. Value: Identifies documentation gaps, validates team readiness, satisfies baseline audit requirements.

2. Simulation Test

A controlled test where recovery procedures are executed against a non-production environment. The team restores from backups, brings up recovery infrastructure, validates data integrity, and measures actual recovery time -- but production traffic is never affected.

When to use: Semi-annually or annually. Duration: 4 to 8 hours. Risk: Low. Production is not affected, but the test consumes team time and cloud resources. Value: Validates that recovery procedures actually work. Measures actual RTO against target RTO. Identifies technical gaps that tabletop exercises cannot surface.

3. Full Failover Test

A test where production traffic is actually routed to the recovery environment. This is the most rigorous form of DR testing and the only one that validates the complete end-to-end recovery process, including DNS propagation, load balancer reconfiguration, session handling, and user experience during failover.

When to use: Annually for Tier 1 systems (if architecture supports it safely). Duration: 2 to 8 hours, including failover, validation, and failback. Risk: Moderate to high. Production traffic is affected. Careful planning, rollback procedures, and a maintenance window are required. Value: Highest. This is the only test type that proves the DR plan works end-to-end. Auditors view failover test evidence as the strongest possible DR validation.

Testing frequency

Test TypeMinimum FrequencyBest PracticeAudit Evidence Value
Tabletop exerciseAnnuallyQuarterlySatisfies baseline requirements
Simulation testAnnuallySemi-annuallyStrong evidence of plan validity
Full failover testAs architecture permitsAnnually for Tier 1Strongest possible evidence

What to document for auditors

Every DR test must produce a formal report containing:

  • Test date and duration
  • Test type (tabletop, simulation, or failover)
  • Scenario description (what failure was simulated)
  • Participants (names and roles)
  • Systems tested
  • RPO/RTO targets for the tested systems
  • Actual recovery time achieved during the test
  • Actual data loss (if any) measured during the test
  • Pass/fail determination based on whether RPO/RTO targets were met
  • Findings (what worked, what did not, what gaps were identified)
  • Remediation items with owners and deadlines
  • Sign-off from the DR plan owner or CISO

This report is your primary audit evidence for DR testing. Without it, the test did not happen from an auditor's perspective.

For a parallel discussion of testing requirements for incident response plans, see our Incident Response Plan Guide.


DR Runbooks: What to Include

A DR runbook is a step-by-step recovery procedure for a specific system or failure scenario. It is the operational document that an engineer follows during an actual DR event. If the DR plan is the strategy, the runbook is the tactical execution guide.

Why runbooks matter for compliance

Auditors assess whether your recovery procedures are documented at a level of detail that enables execution by qualified personnel who may not have designed the system. If your DR plan says "restore the database from backup" without specifying which database, which backup, what tool to use, what validation steps to perform, and how to reconnect dependent services, the plan is insufficient.

Runbook structure

Each DR runbook should contain:

1. Scope and trigger conditions

  • What system or service this runbook covers
  • What failure scenarios trigger this runbook
  • Prerequisites (access credentials, tools, network connectivity)

2. Pre-recovery checklist

  • Confirm the nature and scope of the failure
  • Notify the DR team and initiate communication procedures
  • Preserve evidence or logs if the failure may be security-related
  • Verify backup availability and integrity before beginning restoration

3. Step-by-step recovery procedures

  • Numbered, sequential steps with specific commands, console actions, or API calls
  • Expected output or result for each step (so the operator can verify each step completed correctly)
  • Decision points (if step X produces result Y, proceed to step 7; if result Z, escalate to the database lead)
  • Time estimates for each step

4. Post-recovery validation

  • How to verify the recovered system is functioning correctly
  • Data integrity checks (row counts, checksum comparisons, functional tests)
  • Dependency verification (confirm all dependent services can connect to the recovered system)
  • Performance baseline verification (is the recovered system performing within normal parameters?)

5. Failback procedures

  • How to return to the primary environment once the disruption is resolved
  • Data reconciliation procedures if writes occurred in the recovery environment
  • DNS or routing changes required for failback

6. Communication checkpoints

  • When to send status updates and to whom
  • When to notify customers of recovery completion
  • When to stand down the DR team

Runbook maintenance

Runbooks become stale quickly. Any infrastructure change, deployment process change, or tooling change can invalidate runbook steps. Establish a maintenance procedure:

  • Review and test runbooks after every significant infrastructure change
  • Review all runbooks during annual DR plan review
  • Version-control runbooks alongside infrastructure-as-code (or in the same repository)
  • Assign an owner for each runbook who is responsible for keeping it current

DR Metrics and KPIs

Measuring disaster recovery performance is essential for both operational improvement and audit compliance. Without metrics, you cannot demonstrate to auditors that your DR program is effective, and you cannot identify areas that need investment.

Core DR metrics

1. Actual RTO vs. Target RTO

The most important DR metric. Measured during every DR test and every real recovery event. If your target RTO for the production database is 1 hour and your last DR test achieved recovery in 47 minutes, your RTO compliance ratio is 100% (target met). If the test achieved recovery in 1 hour 23 minutes, you have a 23-minute gap that must be addressed.

2. Actual RPO vs. Target RPO

Measured by comparing the timestamp of the most recent backup or replication point to the time of the simulated (or actual) failure. If your target RPO is 15 minutes and the most recent backup was 12 minutes old at the time of failure, RPO compliance is 100%. If the most recent backup was 45 minutes old, you have a 30-minute gap.

3. DR Test Pass Rate

The percentage of DR tests that successfully met both RPO and RTO targets. Track this over time. A declining pass rate indicates infrastructure drift, architectural changes that have not been reflected in the DR plan, or growing complexity that exceeds current recovery capabilities.

4. Backup Success Rate

The percentage of scheduled backups that completed successfully. Target: 99.9% or higher. Monitor for failed, missed, or incomplete backups and investigate every failure.

5. Backup Validation Success Rate

The percentage of backup restoration tests that produced valid, complete, and usable data. A backup that completes but produces corrupted or incomplete data is worse than a missing backup -- it creates false confidence.

6. Mean Time to Recovery (MTTR)

The average time to restore a system from the moment of failure detection to the moment the system is operational. Track MTTR by system tier and compare against RTO targets.

7. Replication Lag

For systems using asynchronous replication, track the replication lag between primary and replica. Sustained replication lag that exceeds your RPO target means your actual RPO is worse than your documented RPO -- a finding auditors will identify.

8. Time Since Last DR Test

Track the elapsed time since the last DR test for each system tier. If your policy requires annual testing and 14 months have elapsed since the last test, you have a compliance gap.

Reporting

Compile DR metrics into a quarterly report reviewed by the DR plan owner and shared with executive leadership. This report serves as both an operational management tool and audit evidence demonstrating ongoing DR program governance.


Common DR Plan Audit Failures

After reviewing disaster recovery plans across hundreds of SaaS companies and multiple compliance frameworks, these are the failures that most consistently produce audit findings.

1. RPO/RTO targets are not defined

The DR plan describes backup procedures and recovery processes but never explicitly states what recovery time or data loss is acceptable for each system. Auditors cannot evaluate whether your DR capability is adequate without defined targets.

Fix: Define RPO and RTO for every in-scope system, justified by a business impact analysis.

2. The plan has never been tested

The single most common finding. The plan was written, approved, and never validated. When the auditor asks for test results, there are none.

Fix: Schedule your first DR test within 30 days. Start with a tabletop exercise if you have never tested before, then progress to simulation tests.

3. Test results show RTO/RPO targets were not met -- and no remediation occurred

A DR test was performed, and the results showed that recovery took 6 hours against a 2-hour RTO target. But no remediation was documented, no architecture changes were made, and the same failing configuration is still in place.

Fix: Every DR test that fails to meet targets must produce documented remediation items with owners and deadlines. The next test must demonstrate improvement.

4. Backups are in the same region/account as production

The DR plan describes backup procedures, but all backups are stored in the same AWS region, GCP project, or Azure subscription as production. A region-wide outage or an account compromise takes out both production and backups.

Fix: Store at least one backup copy in a separate region. For ransomware protection, store at least one copy in a separate account with immutable storage.

5. The plan references infrastructure that no longer exists

The DR plan was written two years ago and references an AWS architecture that has since been replaced. Database names, service configurations, and region designations are outdated. The runbooks contain commands for systems that no longer exist.

Fix: Review and update the DR plan after every significant infrastructure change. Include DR plan review as a step in your change management process. For more on change management, see our Change Management Guide.

6. No defined roles and responsibilities

The plan describes what must happen but not who does it. When a real disaster occurs, nobody knows who initiates failover, who communicates with customers, or who has the authority to make recovery decisions.

Fix: Define specific roles (DR Coordinator, Recovery Lead, Communications Lead) with named individuals and alternates.

7. Backup integrity has never been validated

Backups run daily and complete successfully. But no one has ever restored from a backup and verified the data. When a real restoration is needed, the backup turns out to be corrupted, incomplete, or incompatible with the current schema.

Fix: Implement monthly automated restore tests. Verify data integrity after every test restore.

8. The DR plan is disconnected from the BCP

The disaster recovery plan exists as a standalone technical document with no connection to the broader business continuity plan. Recovery objectives in the DR plan do not align with maximum tolerable downtime in the BCP. Auditors see two plans that do not reference each other.

Fix: Ensure the DR plan explicitly references the BCP, that RPO/RTO values are derived from BCP-level business impact analysis, and that both documents are reviewed together.

9. No maintenance or review cadence

The DR plan has no defined review cycle, or the defined review cycle has not been followed. The version history shows no updates in the last 18 months.

Fix: Establish a formal review cadence (at minimum, annually and after major changes). Document every review, including reviews where no changes were made.

10. Emergency mode operations are not addressed (HIPAA)

For companies subject to HIPAA, the DR plan does not address how critical business processes that protect ePHI security continue during emergency mode operations -- a specific required implementation specification under 164.308(a)(7)(ii)(C).

Fix: Add an emergency mode operations section that documents how ePHI is protected during and after a disaster, including access controls, audit logging, and encryption that remain in effect during recovery.


FAQ

What is the difference between a disaster recovery plan and a business continuity plan?

A disaster recovery plan focuses specifically on restoring IT systems, applications, and data after a disruption. A business continuity plan is broader -- it addresses how the entire organization continues to operate during a disruption, including people, processes, communications, facilities, and vendor dependencies. The DR plan is a component of the broader BCP. Both are required by SOC 2, ISO 27001, HIPAA, and PCI DSS, and auditors expect the two plans to be consistent and cross-referenced.

How often should a disaster recovery plan be tested?

At minimum, annually. PCI DSS and HIPAA require periodic testing. SOC 2 auditors expect at least annual testing. ISO 27001 auditors require evidence that ICT readiness has been tested. Best practice is quarterly tabletop exercises, semi-annual simulation tests, and annual full failover tests for Tier 1 systems. Any test is better than no test -- if you have never tested your DR plan, start with a tabletop exercise this month.

What is the difference between RPO and RTO?

RPO (Recovery Point Objective) defines the maximum acceptable amount of data loss, measured in time. RTO (Recovery Time Objective) defines the maximum acceptable duration of system downtime. RPO determines your backup frequency and replication strategy. RTO determines your recovery architecture and level of standby infrastructure. Both must be defined for every critical system, justified by a business impact analysis, and validated through DR testing.

How much does disaster recovery cost for a SaaS company?

DR costs vary widely based on recovery objectives and architecture. A backup-and-restore strategy for Tier 3 systems costs little beyond storage (typically $50 to $500/month depending on data volume). A warm standby for Tier 2 systems adds 30-50% of the primary environment's compute cost. A hot standby or active-active architecture for Tier 1 systems can double your infrastructure spend for those systems. The right question is not "how much does DR cost?" but "how much does downtime cost?" -- and designing DR architecture that balances recovery speed against infrastructure investment.

Do we need a separate DR plan for each compliance framework?

No, and maintaining separate plans is counterproductive. Build one comprehensive DR plan that addresses the most prescriptive requirements across all applicable frameworks, then include a compliance mapping appendix that shows how each section of your DR plan satisfies each framework's specific requirements. SOC 2 A1.2/A1.3, ISO 27001 A.5.29/A.5.30/A.8.13/A.8.14, HIPAA 164.308(a)(7), and PCI DSS 12.10 all require fundamentally similar capabilities.

What is the minimum DR plan needed to pass a SOC 2 audit?

If Availability is in scope (A1 criteria), the minimum is: a documented DR plan with defined RPO/RTO, implemented backup procedures, evidence of DR testing within the audit period, documentation of test results and findings, and evidence that gaps identified during testing were addressed. SOC 2 does not prescribe specific RTO/RTO values -- but the values you define must be achievable and demonstrated through testing. Unrealistic targets that testing shows you cannot meet will produce a finding.

Should our DR plan cover SaaS vendor outages?

Yes. Your application's availability depends on every service in the dependency chain. If your application relies on a third-party authentication provider, payment processor, or database-as-a-service, your DR plan should include procedures for handling outages of those services. This does not mean you need to build redundancy for every vendor, but you need documented procedures: what is the impact, how does the team respond, what is the customer communication plan, and are there alternative services that can be activated?

How do we handle DR for a microservices architecture?

DR for microservices requires service-level classification and recovery planning. Not every microservice needs the same DR strategy. Classify each service by tier, define RPO/RTO per service, and design recovery procedures that account for service dependencies. Pay particular attention to state -- stateless services are inherently easier to recover than stateful services. Document the dependency graph and define the recovery sequence (which services must be restored first to enable recovery of dependent services). Test recovery of the full service mesh, not just individual services in isolation.


Automate Your Disaster Recovery Compliance with QuickTrust

Building a disaster recovery plan that satisfies four compliance frameworks simultaneously is a substantial undertaking. Maintaining it -- keeping recovery procedures current as infrastructure evolves, tracking testing schedules, documenting test results, validating that backup procedures meet RPO targets, and producing audit-ready evidence on demand -- is an ongoing operational discipline.

QuickTrust reduces that burden. Our platform provides:

  • Pre-built DR policy and plan templates mapped to SOC 2, ISO 27001, HIPAA, and PCI DSS requirements, customizable to your specific architecture and recovery objectives
  • Automated control mapping that cross-references your DR plan to every applicable framework requirement, ensuring coverage across SOC 2 A1.2/A1.3, ISO 27001 A.5.29/A.5.30/A.8.13/A.8.14, HIPAA 164.308(a)(7), and PCI DSS 12.10
  • DR testing management with scheduling, scenario templates, test report generation, and remediation tracking -- so you always have current test evidence when auditors ask
  • Continuous compliance monitoring that tracks whether your DR plan is current, tested within the required cadence, and aligned with your infrastructure's actual state
  • Backup monitoring integration that validates backup completion rates and flags gaps before auditors find them

Stop treating disaster recovery compliance as a document-filing exercise. Start treating it as a continuously monitored, always audit-ready capability.

Get started with QuickTrust and see how SaaS companies like yours build disaster recovery programs that pass every audit -- the first time.


This guide is maintained by the QuickTrust Editorial team and updated as compliance frameworks, regulatory requirements, and industry best practices evolve. Last updated: March 19, 2026.

Ready to get audit-ready?

Our engineers implement controls, prepare evidence, and coordinate your audit.

Get a Free Assessment

Related Articles