How to Manage Planned Downtime the Right Way, with Synthetics

Your synthetic browser tests should give you confidence, not confusion. If you have been following this series, you now have tests that actually matter. They reflect real user journeys, validate outcomes, and alert with purpose.

But even the best synthetic tests can create noise during planned change.

Every time you roll out a release or perform scheduled maintenance, synthetic tests start capturing every side effect that comes with change. Maybe a page loads slower while caches warm. Maybe a dependency restarts. Maybe a flow becomes inconsistent for a few minutes as infrastructure shifts under it.

Nothing is wrong. You are just making changes. The problem is the noise.

This is where planned downtime management comes in. By watermarking releases, suppressing expected failures, and marking maintenance periods with clear business context, you keep your synthetic tests clean and your signals meaningful during the moments when your environment changes the most.

This article covers the next best practice in the Getting Synthetics Right Series: Manage Planned Downtime the Right Way. If you’re new to the series, check out the introduction article to learn how these best practices come together to make your synthetic browser tests reliable and actionable.

What it is: Planned downtime management for synthetic tests

Planned downtime management is the practice of preparing your synthetic browser tests for expected change so they do not produce misleading failures or skew your performance data. It tells your observability platform, “We know things will look different right now. Interpret the results in context.”

With Splunk Observability Cloud, downtime configurations let you:

This helps preserve the clarity of synthetic signals before, during, and after deployments, upgrades, pathing changes, and other forms of controlled operational work.

Why it matters

Planned downtime matters because change often creates noise, and noise hides real issues (or minimally diminishes confidence).

During releases and maintenance, synthetic tests behave correctly by detecting slow pages, inconsistent flows, and temporary failures. The challenge is that these signals are expected during maintenance windows. If you treat them like real incidents, you end up with false alerts, polluted dashboards, and misleading performance trends.

Here are the core risks:

  1. Alert storms during planned work. Deployments often cause brief, expected failures. Without downtime, every fault is an opportunity to fire an alert.
  2. Distorted baselines. Warm-up behavior and temporary latency shifts skew performance and uptime metrics.
  3. No record of when change occurred. Without marking maintenance windows, you cannot tie regressions or improvements to the actual event. Remember context is critical.

Planned downtime solves these problems by marking expected change, suppressing noisy signals, and keeping your synthetic data clean. The goal? When a synthetic alert fires outside a maintenance window, you know it means something.

Putting it into practice: How to manage planned downtime

Here is how to manage planned downtime effectively in Splunk Observability Cloud so your synthetic test data stays clean, accurate, and aligned with real operational change.

1. Choose the right downtime rule

Splunk Synthetic Monitoring supports two downtime rules. Each changes how synthetic data is collected and interpreted during maintenance. Review the table below to better understand the downtime options and some example use cases.

Downtime options

Rule
Description
Result
Best Use Cases
Pause Tests
Stops selected synthetic tests from running during the downtime window.
No synthetic data is collected. Charts show gaps.
Login maintenance, database cutovers, known breakage windows, infrastructure work.
Augment Data
Tests continue running but each run is tagged with under_maintenance=true. These runs are excluded from uptime, SLAs, SLOs, and averages.
Continuous visibility without polluting availability or baseline reports.
Releases where you want to observe warm-up behavior, dependency restarts, or drift.

Pro Tip: Use under maintenance data for insight

Even though under_maintenance=true runs are excluded from SLAs and SLOs, they can still reveal valuable behavior during change.

What to Look For
Why It Matters

• Warm-up impact after deployments

• Short-lived latency spikes

Shows where the application may need caching, readiness checks, or stabilization steps.
• Dependency sensitivity or third-party instability
Helps identify fragile integration points that behave differently during change.
• Regression indicators compared with RUM or APM
Confirms whether synthetic slowdowns match real user impact or backend service issues.

Learn more about Downtime Configurations.

Pro-tip: Track maintenance Visually in Dashboards

Splunk displays downtime directly on synthetic test charts so you can instantly see when maintenance occurred and how performance behaved around it.

Visual indicators include:

Check out these sample screencaps:

Downtime records remain available for thirteen months, preserving clean historical visibility.

2. Configure synthetics downtime

There are two ways to configure downtime for your synthetic tests in Splunk Observability Cloud. You can either:

Both approaches help you keep your test data clean, the choice really depends on how your teams operate.

2.1 Configure downtime in the UI

The Splunk UI is ideal for operationally driven maintenance windows, scheduled releases, and situations where teams want visual clarity without automation. In a past life, part of our change management process was the association of a change task to enable downtime windows post change approval.

When the UI is the best fit:

Downtime actions in the UI

Action
Description
Learn More
Create a downtime window
Configure a one-time or recurring window for selected tests.
Schedule a downtime configuration
Modify an existing downtime
Edit, extend, end early, or delete downtime windows depending on their lifecycle.
Modify an existing downtime

UI-based management is simple, predictable, and easy to operationalize across teams.

2.2 Configure Downtime as Code Using the API

Downtime can also be created and managed programmatically through the Observability cloud API . This is the right choice for teams that automate deployment workflows or want downtime to be controlled directly by their release pipelines or change management tooling

When the API is the best fit:

Learn more: Leveraging the API to manage synthetics downtime configurations.

Example: Create a downtime window with the API

Snippet
curl -X POST "https://api.us1.signalfx.com/v2/synthetics/downtime_configurations" 

  -H "Content-Type: application/json" 

  -H "X-SF-TOKEN: $TOKEN" 

  -d '{

        "downtimeConfiguration": {

          "name": "release-maintenance",

          "rule": "augment_data",

          "testIds": [12345],

          "startTime": "2025-05-01T02:00:00Z",

          "endTime": "2025-05-01T03:00:00Z",

          "timezone": "America/New_York"

        }

      }'
Title
JSON
Label
Create a downtime window with the API
Type
json
Show Copy Button
true

API-managed downtime ensures observability stays aligned with how you deploy and operate your applications.

3. Add buffers before and after maintenance

Maintenance windows need time on both sides. Systems warm up, dependencies initialize, caches rebuild, and traffic stabilizes after deployment.

Here’s why buffers help:

Python example with API call

Snippet
import requests
from datetime import datetime, timedelta

TOKEN = ""
REALM = "us1"
TEST_IDS = [12345]

start = datetime(2025, 5, 1, 2, 0)
end   = datetime(2025, 5, 1, 3, 0)
buffer = timedelta(minutes=15)

downtime_start = start - buffer
downtime_end   = end + buffer

payload = {
    "downtimeConfiguration": {
        "name": "release-maintenance",
        "rule": "augment_data",
        "testIds": TEST_IDS,
        "startTime": downtime_start.isoformat() + "Z",
        "endTime": downtime_end.isoformat() + "Z",
        "timezone": "America/New_York"
    }
}

resp = requests.post(
    url=f"https://api.{REALM}.signalfx.com/v2/synthetics/downtime_configurations",
    json=payload,
    headers={"Content-Type": "application/json", "X-SF-TOKEN": TOKEN},
)

print(resp.status_code, resp.text))
Title
Python
Label
Python example with API call
Type
python
Show Copy Button
true

4. Trigger a validation run after maintenance

Once maintenance ends, run your most important tests immediately to confirm key workflows are functioning correctly. You can do this either:

This gives you immediate confirmation that your release completed successfully.

Learn more:

The bottom line

Planned downtime is not just about stopping alerts. It is about keeping your synthetic data clean and your signals trustworthy during moments of controlled change. With the right downtime rules, clean buffers, API-based automation, post-maintenance validation, and thoughtful use of augmented test runs, your synthetic monitoring becomes a stable, reliable part of your observability practice.

Handled well, downtime becomes a strength. When a synthetic test fires outside a maintenance window, you know it is real. When it stays quiet during a release, you know your configuration is working. And when augmented runs reveal drift or fragility, you see early warning signs before users are affected.

Call to action

Review your current downtime configuration and ensure your most important tests have the right rules in place. Add buffers around your next release, try the downtime API to automate the process, and validate key workflows once maintenance completes.

To continue building synthetic tests your team can trust, follow the rest of this series. You can also explore Splunk Observability Cloud with a free trial and see how synthetics, RUM, and APM come together to provide complete end to end visibility.

Related Articles

What the North Pole Can Teach Us About Digital Resilience
Observability
3 Minute Read

What the North Pole Can Teach Us About Digital Resilience

Discover North Pole lessons for digital resilience. Prioritise operations, just like the reliable Santa Tracker, for guaranteed outcomes. Explore our dashboards for deeper insights!
The Next Step in your Metric Data Optimization Starts Now
Observability
6 Minute Read

The Next Step in your Metric Data Optimization Starts Now

We're excited to introduce Dimension Utilization, designed to tackle the often-hidden culprit of escalating costs and data bloat – high-cardinality dimensions.
How to Manage Planned Downtime the Right Way, with Synthetics
Observability
6 Minute Read

How to Manage Planned Downtime the Right Way, with Synthetics

Planned downtime management ensures clean synthetic tests and meaningful signals during environment changes. Manage downtime the right way, with synthetics.
Smart Alerting for Reliable Synthetics: Tune for Signal, Not Noise
Observability
7 Minute Read

Smart Alerting for Reliable Synthetics: Tune for Signal, Not Noise

Smart alerting is the way to get reliable signals from your synthetic tests. Learn how to set up and use smart alerts for better synthetic signaling.
How To Choose the Best Synthetic Test Locations
Observability
6 Minute Read

How To Choose the Best Synthetic Test Locations

Running all your synthetic tests from one region? Discover why location matters and how the right test regions reveal true customer experience.
Advanced Network Traffic Analysis with Splunk and Isovalent
Observability
6 Minute Read

Advanced Network Traffic Analysis with Splunk and Isovalent

Splunk and Isovalent are redefining network visibility with eBPF-powered insights.
Conquer Complexity, Accelerate Resolution with the AI Troubleshooting Agent in Splunk Observability Cloud
Observability
4 Minute Read

Conquer Complexity, Accelerate Resolution with the AI Troubleshooting Agent in Splunk Observability Cloud

Learn more about how AI Agents in Observability Cloud can help you and your teams troubleshoot, identify root cause, and remediate issues faster.
Instrument OpenTelemetry for Non-Kubernetes Environments in One Simple Step
Observability
2 Minute Read

Instrument OpenTelemetry for Non-Kubernetes Environments in One Simple Step

The OpenTelemetry Injector makes implementation incredibly easy and expands OpenTelemetry's reach and ease of use for organizations with diverse infrastructure.
Resolve Database Performance Issues Faster With Splunk Database Monitoring
Observability
3 Minute Read

Resolve Database Performance Issues Faster With Splunk Database Monitoring

Introducing Splunk Database Monitoring, which helps you identify and resolve slow, inefficient queries; correlate application issues to specific queries for faster root cause analysis; and accelerate fixes with AI-powered recommendations.