How do I configure alerts for different severity levels (Warning vs. Error)?
Eliminate alert fatigue with Nodinite graduated severity levels. Configure intelligent dual-threshold monitoring that provides early warnings without crying wolf, and critical alerts that demand immediate action.
✅ Set Warning threshold - Early signal (75% of normal) for proactive investigation
✅ Set Error threshold - Critical level (50% of normal) for immediate escalation
✅ Graduate response actions - Different recipients, frequencies, and escalation paths
✅ Prevent false escalations - Gap between thresholds allows appropriate responses
Set dual thresholds per Non-Events resource. E-commerce order monitoring example:
Non-Events resource: "Order Confirmations - Business Hours"
Warning threshold: Min 600 orders/hour (75% of normal 800, early heads-up)
- Alert recipient: Operations team email + Slack #operations-alerts channel
- Alert frequency: Every 30 minutes while condition persists
- Escalation: None (operational awareness, investigate when convenient)
- Action: Check order processing service health, review error logs, monitor trend (is volume dropping further?)
Error threshold: Min 400 orders/hour (50% of normal 800, critical data outage)
- Alert recipient: Operations team pager + SMS to on-call engineer + Slack #critical-alerts
- Alert frequency: Immediately, repeat every 5 minutes until resolved
- Escalation: If not acknowledged within 15 minutes, escalate to engineering manager + CTO
- Action: Immediate investigation, assume critical outage (payment gateway down, integration service crashed, API authentication failure), engage incident response procedure
Threshold tuning rationale:
- Warning at 75%: Provides early signal (volume trending down but not catastrophic, proactive investigation prevents Error)
- Error at 50%: Clear critical threshold (half normal volume = obvious data outage, not normal variance)
- Gap between Warning/Error: Allows graduated response (Warning = watch, Error = act immediately), reduces false escalations
AckNak severity example:
- Warning: Payment acknowledgment >25 seconds (trending toward 30-second SLA violation, proactive capacity planning)
- Error: Payment acknowledgment >30 seconds (SLA violated, regulatory compliance issue, customer escalation risk)
Next Step
Need more help? Check the Troubleshooting Overview for all FAQs, or refer to the Configuration Guide to fine-tune your Non-Events monitoring setup.