Business Continuity & Disaster Recovery (BC/DR) Plan

ECS Technology Solutions

Version 1.1 – Updated: 17 April 2025


1. Purpose

This BC/DR Plan ensures ECS Technology Solutions (“ECS”) can continue to deliver critical managed services and meet contractual and regulatory obligations during and after a disruptive incident, while safeguarding customer data and minimizing downtime.

2. Scope

  • People: All employees, contractors, and key third-party partners.
  • Facilities: Headquarters in Elkhorn, NE and any remote offices.
  • Technology: On-premises infrastructure, cloud resources in AWS & Azure, SaaS platforms (M365, ticketing, SIEM), and customer-hosted systems managed by ECS.
  • Services: 24×7 Service Desk, Remote Monitoring & Management (RMM), Security Operations Center (SOC), Backup & Recovery, Professional Services, and Hosting.

3. Objectives

ObjectiveTarget
Protect life & safetyImmediate
Stabilize critical services≤ 4 hours
Restore all customer-facing services≤ 24 hours
Return to normal operations≤ 7 days
Communicate status updatesFirst notice ≤ 30 min; subsequent every 2 h

4. Definitions

TermMeaning
RTORecovery Time Objective – max acceptable downtime.
RPORecovery Point Objective – max tolerable data loss.
Critical SystemOutage halts customer operations or security.
Incident Commander (IC)Individual who leads BC/DR response (default: Information Security Manager).

5. Business Impact Analysis (BIA)

TierExample SystemsRTORPO
Tier 1 – CriticalTicketing (HaloPSA), SIEM, backup vaults, RMM, MFA IdP4 h1 h
Tier 2 – ImportantDocumentation (Hudu), ERP/finance, internal Git repos24 h8 h
Tier 3 – RoutineMarketing website, historical archives72 h24 h

6. Roles & Responsibilities

RoleResponsibility
PresidentActivate BC/DR plan, allocate resources, approve external communications.
Incident Commander (IC)Lead response, coordinate teams, track progress.
Chief Technology Officer (CTO)Execute infrastructure failover, validate restorations.
Information Security ManagerManage security incidents, coordinate forensics, maintain BC/DR documentation, schedule tests, and notify regulators if required.
Service Delivery ManagersLiaise with clients, prioritise tickets, provide status reports.
All StaffFollow instructions, execute playbooks, escalate issues.

7. Risk Assessment (Top Threats)

  • Cyber-Attack / Ransomware – Data encryption & service disruption.
  • Cloud Region Outage – Loss of primary AWS or Azure region.
  • Critical SaaS Failure – Ticketing or SIEM vendor outage.
  • Physical Disaster – Fire, flood, or tornado impacts HQ.
  • Pandemic / Workforce Unavailability – Staff unable to work onsite.

8. Preventive & Mitigation Controls

  • Encryption: AES-256 (or strongest available) for data at rest; TLS 1.3 for transit.
  • Backups: Daily backups to isolated Backup account/vault; cross-region replication; immutable retention.
  • Real-Time Patch & Vulnerability Management: Immediate critical patching; continuous scanning.
  • Least-Privilege IAM & MFA: Mandatory on Tier 1 systems.
  • Redundant Cloud Architecture: Multi-AZ, cross-region failover for critical workloads.
  • Endpoint Protection & EDR: Detect and contain threats rapidly.
  • UPS & Generator: Protect on-prem equipment from power loss.

9. Recovery Strategies

9.1 Data Backup & Restore

  • Snapshots stored in isolated AWS/Azure backup vaults (encrypted).
  • Restore precedence: Tier 1 → Tier 2 → Tier 3.

9.2 Cloud Infrastructure Failover

PlatformPrimarySecondaryMethod
AWSus-east-1us-west-2Infrastructure-as-Code redeploy via CloudFormation + data restore
AzureCentral USEast US 2Azure Site Recovery + ARM templates

9.3 SaaS Service Outage

  • Switch to out-of-band comms (Teams → SMS) if IdP down.
  • Exported data restored into secondary provider or database.

9.4 Office Inaccessibility

  • Staff operate fully remote using secured endpoints and MFA.
  • VPN not required (Zero-Trust SaaS); fallback cellular hotspots issued.

10. Disaster Recovery Procedures

  1. Detection & Triage: SOC escalates event to IC.
  2. Declare Incident: IC records start time, severity, notifies President.
  3. Activate Plan: Mobilise recovery teams, reference relevant playbooks.
  4. Communication: Send initial client advisory ≤ 30 min via status page & email.
  5. Recovery Actions: Execute appropriate strategy (e.g., failover, restore).
  6. Validation: QA confirms service functionality; monitoring re-enabled.
  7. Return to Normal: Switch back to primary sites; debrief scheduled.

11. Communication Plan

AudienceMethodFrequency
Internal StaffTeams / phone bridge30-min stand-ups
ClientsEmail from support@ecs.rocks & Status PageInitial + 2-hourly
Vendors / ISPsPhone & ticket portalsAs needed
Regulators / AuthoritiesLetter or secure emailWithin legal timeframes

12. Testing & Training

  • Annual Tabletop Exercises – Simulate scenarios, update gaps.
  • Monthly KnowBe4 Training – Reinforce BC/DR awareness & phishing resilience.

13. Plan Maintenance

  • Reviewed annually by the Information Security Manager and CTO, or sooner after major changes.
  • Lessons learned post-incident feed into next revision.
  • Latest approved plan stored and distributed via SharePoint.