Drift Alerts
Drift detection identifies gradual performance degradation that individual anomaly scores might miss. By comparing recent execution metrics against established baselines, drift alerts surface trends before they become critical incidents.
Drift Types
NodeLoom monitors four categories of drift for each workflow:
| Drift Type | What It Measures |
|---|---|
| Duration drift | Average execution time is increasing compared to the baseline. Indicates performance degradation, slow API responses, or resource contention. |
| Token drift | Average token consumption per execution is increasing. May indicate prompt bloat, unnecessary tool calls, or model changes. |
| Output size drift | Average output payload size is growing. Could indicate unbounded data fetches or downstream API changes returning more data. |
| Error rate drift | The percentage of failed executions is increasing relative to the baseline error rate. |
How Detection Works
Drift detection compares a recent window of executions (the last 50 executions or last 7 days, whichever is smaller) against the baseline (calculated from the previous 30 days of successful executions). If the recent average exceeds the baseline by more than the configured threshold percentage, a drift alert is triggered.
Trigger Timing
Drift checks are triggered in two ways:
- After execution: Each completed execution triggers a drift check for the workflow. This provides near-real-time detection for high-frequency workflows.
- Scheduled scan: A periodic background scan checks all active workflows for drift, catching slow-moving trends in low-frequency workflows that do not execute often enough for per-execution detection.
Minimum data requirement
Configurable Thresholds
Each team can customise drift thresholds from the workspace monitoring settings. Thresholds are expressed as a percentage above the baseline:
| Drift Type | Description |
|---|---|
| Duration threshold | How much slower the recent average duration can be before triggering an alert. |
| Token threshold | How much more tokens the recent average can consume before triggering an alert. |
| Output size threshold | How much larger the recent average output size can be before triggering an alert. |
| Error rate threshold | The absolute increase in error rate (percentage points) that triggers an alert. |
Per-workflow overrides
Alert Lifecycle
Drift alerts follow a simple lifecycle:
| State | Description |
|---|---|
| Active | The drift condition has been detected and the alert is visible in the monitoring dashboard. Notifications are sent. |
| Acknowledged | A team member has reviewed the alert. It remains visible but is marked as acknowledged. |
| Resolved | The recent metrics have returned within the threshold. The alert is automatically resolved. |
Alerts are automatically resolved when the recent window returns below the threshold on subsequent checks. You do not need to manually close resolved alerts.
Notifications
Drift alerts use the same notification channels as anomaly detection: email and webhook. Configure notification preferences per team from the workspace monitoring settings.
To avoid alert fatigue, NodeLoom deduplicates drift notifications. A new notification is only sent when:
- A new drift type is detected for a workflow (e.g., duration drift appears for the first time).
- A previously resolved drift type re-triggers.
- The drift severity increases significantly (e.g., from 30% over baseline to 60% over baseline).
Drift vs Anomaly Detection
While both features monitor execution health, they serve different purposes:
| Anomaly Detection | Drift Alerts | |
|---|---|---|
| Scope | Individual execution | Trend across many executions |
| Question answered | Was this specific execution unusual? | Is this workflow getting worse over time? |
| Detection speed | Immediate (per execution) | Gradual (requires a window of data) |
| Best for | Catching one-off spikes, security events | Identifying regressions, performance degradation |
Next Steps
- Anomaly Detection -- per-execution anomaly scoring and alerting.
- Token Usage -- monitor token consumption trends in detail.
- Sentiment Tracking -- monitor conversational quality in AI agent workflows.