FAQ

How do we test security changes without risking unplanned downtime?

There is no way to introduce security changes with zero risk of downtime, but you can reduce and contain that risk with an OT-aware change process. The key is to test in conditions that are as close as practical to production, prove safe behavior, and have a controlled path to roll back.

1. Start with an OT-focused change control process

Before testing, make sure security changes follow the same rigor as process or equipment changes:

  • Formal risk assessment: Identify which cells, lines, and batch records could be impacted. Include safety, product quality, and availability.
  • Asset and dependency mapping: Understand what depends on the system in scope (MES interfaces, historians, HMIs, recipe servers, QMS, calibration systems, etc.).
  • Vendor and OEM input: Confirm supported OS versions, patches, and configuration changes for PLCs, DCS, MES, and lab systems. Unsupported changes often create hidden failure modes.
  • Documented test & rollback plans: Define how you will test, what success looks like, what evidence you will retain, and exactly how you will revert if something misbehaves.

2. Use representative test and staging environments

In brownfield plants, it is rare to have a perfect replica of production, but some form of staging is still critical.

  • Tiered environments: Aim for at least three tiers: development / lab, staging / pre-production, and production.
  • Representative configurations: Ensure staging runs the same OS versions, antivirus, group policies, firewall rules, and application versions as production where feasible.
  • Cloned images: Where you cannot build full staging, create VM or disk-image clones of critical servers and engineering workstations to test patches and configuration changes offline.
  • Network realism: Include key network elements in test: VLANs, firewalls, jump hosts, domain controllers, time servers, and typical latency. Many issues show up only when security controls interact (e.g., firewalls + AV + logging agents).

For safety or validation reasons, you may not be able to duplicate all field devices or control networks. In that case, explicitly document what is not represented in test and adjust the production deployment plan accordingly.

3. Define realistic test scenarios and acceptance criteria

Patches and security configuration changes rarely fail during simple smoke tests; they fail during typical operations and edge cases.

  • Critical workflows: Derive test cases from your highest-risk workflows: batch start/stop, recipe download, equipment changeover, data collection to MES/QMS, label printing, electronic records, and electronic signatures where used.
  • Protocol and integration checks: Explicitly test OT-relevant protocols and integrations (OPC, Modbus/TCP, PROFINET, control network routing, historian feeds, ERP/MES/QMS interfaces).
  • Performance and timing: Observe for slowdowns, timeouts, or buffering issues under realistic load, not just with a single operator.
  • Security + operations: Confirm that new security rules (firewall, allowlisting, AV, logging) do not block expected behavior, background services, or license checks.
  • Clear acceptance criteria: Define in advance what must pass (and for how long) for the change to be considered safe enough for production.

4. Use phased deployment and maintenance windows

Even with strong testing, production behavior can differ. Reduce unplanned downtime risk by controlling rollout:

  • Maintenance windows: Schedule changes during planned downtime or low-risk periods when product impact, safety implications, and capacity constraints are manageable.
  • Pilot lines or systems: Deploy first to a less critical line, cell, or non-GxP area that still reflects your architecture. Monitor closely before expanding.
  • Staged rollout: Sequence systems so dependencies are respected (e.g., domain controllers and jump servers before HMIs; data brokers before historians; historians before MES).
  • Change freezes: Avoid overlapping major changes (e.g., patching OS, updating AV, and replacing a firewall at the same time). Combined changes increase diagnostic complexity and downtime risk.

5. Plan explicit rollback and contingency procedures

Rollback planning is as important as the test plan. Without it, you are accepting higher downtime risk.

  • System images and backups: Take validated backups and, where possible, full system images of servers, engineering workstations, and controller configurations before change.
  • Documented rollback steps: Write step-by-step, tested instructions to revert patches or config changes, including data restoration and verification steps.
  • Time-boxed go/no-go: Set a maximum acceptable outage window for troubleshooting. If exceeded, trigger rollback rather than continuing to “tune in production.”
  • Degraded but safe modes: For some plants, a defined degraded mode (manual recording, reduced automation, bypass of non-critical integrations) can be an alternative while issues are resolved. This must be pre-approved and documented.

6. Coordinate with validation, quality, and regulatory expectations

In regulated environments, security changes may have validation impact:

  • Impact assessment: Determine whether the change is configuration-only, infrastructure-only, or affects validated functions (e.g., data integrity, audit trails, e-signatures, recipe management).
  • Targeted re-validation: Where impact is non-trivial, run targeted regression tests of validated functions in test or staging before production rollout.
  • Traceable evidence: Capture test results, approvals, and configuration baselines in systems where auditors will expect them (QMS, change control tools, or validated document management).
  • Change control alignment: Ensure your cybersecurity change record links to any deviations, CAPAs, or risk assessments that justified the change window or rollout approach.

7. Accept that full replacement strategies are rarely feasible

In mixed-vendor, long-lifecycle OT environments, it is usually not realistic to replace major systems just to simplify security testing. Full replacements often fail due to:

  • Qualification and validation burden: Revalidating an MES, DCS, or core PLC platform can take months and draw heavily on operations and quality resources.
  • Downtime risk: Cutover windows for complete replacements are long and brittle, which conflicts with production and regulatory commitments.
  • Integration complexity: Older ERP, PLM, and QMS integrations may rely on undocumented behavior that is difficult to reproduce in a new stack.

Instead, most plants adopt a coexistence strategy: harden legacy systems as much as practical, introduce new security controls at the network and infrastructure layers, and test changes incrementally with strong containment and rollback options.

8. Monitor aggressively after deployment

Testing does not stop when the patch is applied in production.

  • Heightened monitoring window: Define a period (for example, one to two weeks) of closer monitoring of alarms, performance indicators, and error logs on affected systems.
  • Operator feedback loop: Give operators and maintenance a simple, fast path to report anomalies that might be linked to the change.
  • Security & operations review: After the monitoring window, review incidents, near-misses, and performance changes, and refine your test cases for the next cycle.

9. How this plays out in typical brownfield plants

In many plants, you will not have perfect labs, full vendor support, or enough downtime. The practical pattern is usually:

  • Use a limited OT lab or virtual clones for first-pass testing of patches and configuration changes.
  • Perform targeted testing in staging where available, focused on your most critical workflows and integrations.
  • Roll out in small, well-documented steps during planned windows, with validated backups and explicit rollback plans.
  • Capture test and deployment evidence in your change control and validation systems so you can explain the rationale and residual risks to auditors and internal stakeholders.

This approach does not eliminate risk, but it makes security changes predictable, auditable, and significantly less likely to trigger unplanned downtime.

Get Started

Built for Speed, Trusted by Experts

Whether you're managing 1 site or 100, Connect 981 adapts to your environment and scales with your needs—without the complexity of traditional systems.

Get Started

Built for Speed, Trusted by Experts

Whether you're managing 1 site or 100, C-981 adapts to your environment and scales with your needs—without the complexity of traditional systems.