Skip to main content

datadog-probe

Last updated on

Datadog APM probe allows you to query Datadog metrics or run Synthetic tests and compare the results against specified criteria.

When to use

  • Monitor Datadog metrics (e.g., system.cpu.user, trace.servlet.request.duration) as steady-state indicators during chaos
  • Use Datadog Synthetic tests to validate end-to-end user flows under failure conditions
  • Validate that Datadog-monitored SLOs remain within acceptable thresholds during fault injection

Prerequisites

  • An active Datadog account
  • Access to the Datadog API from the Kubernetes execution plane
  • A Datadog API key and Application key for authentication. See Datadog API Keys and Application Keys

Steps to configure

  1. Navigate to Project Settings > Chaos Probes and click + New Probe

  2. Select APM Probe, provide a name, and select Datadog under APM Type

  3. Under Variables, define any reusable values you want to reference in probe properties or run properties. For each variable, specify the type (String or Number), name, value (fixed or runtime input), and whether it's required at runtime.

  4. Under Datadog Connector, select an existing connector or click + New Connector to create one. Provide the Datadog instance URL, Application key, and API key, configure the delegate, verify the connection, and click Finish.

  5. Under Probe Properties, select the query mode and configure:

    Metrics mode:

    FieldDescription
    Datadog QueryDatadog metrics query.
    Example: avg:system.cpu.user{host:my-host}. See Datadog Metrics documentation
    Lookback Window (in minutes)Time range from the specified number of minutes ago to now

    Synthetic Test mode:

    FieldDescription
    Synthetic TestProvide the Synthetic test details (API test or Browser test) to evaluate the probe outcome. See Datadog Synthetics documentation

    Under Datadog Data Comparison, provide:

    FieldDescription
    TypeData type for comparison: Float or Int
    Comparison CriteriaComparison operator: >=, <=, ==, !=, >, <, oneOf, between
    ValueThe expected value to compare against the metric result
  6. Provide the Run Properties:

    FieldDescription
    TimeoutMaximum time for probe execution (e.g., 10s)
    IntervalTime between successive executions (e.g., 2s)
    AttemptNumber of retry attempts (e.g., 1)
    Polling IntervalTime between retries (e.g., 30s)
    Initial DelayDelay before first execution (e.g., 5s)
    VerbosityLog detail level
    Stop On Failure (optional)Stop the experiment if the probe fails
  7. Click Create Probe