######################### Monitoring and operations ######################### Atmosphere includes a Grafana deployment with dashboards created by default and a Prometheus deployment that collects metrics from the cluster and sends alerts to AlertManager. Loki also collects logs from the cluster using Vector. ****************************** Philosophy and alerting levels ****************************** Atmosphere's monitoring philosophy is strongly aligned with the principles outlined in the `Google Site Reliability Engineering (SRE) book `_. The approach focuses on alerting that's meaningful, practical, and directly tied to user-visible impact. Alerting philosophy =================== Symptom-based alerting ---------------------- Symptom-based alerts take priority over cause-based alerts. They focus on conditions that directly affect users or service health, such as elevated error rates or increased latency, rather than internal system state. Cause-based alerts are acceptable only when no reasonable symptom-based proxy exists, or when they serve as leading indicators for capacity planning (typically P4 or P5). Error budget driven alerting ---------------------------- When service-level objectives (SLOs) exist for a service, burn-rate alerts work better than static threshold alerts. A burn-rate alert fires when the error budget depletes faster than expected, which accounts for baseline error rates and avoids alerting on acceptable levels of degradation. Escalation model ----------------- The alerting system scales severity with impact. Most alerts affecting a single system trigger a lower priority level (P4 or P5). If an issue affects the entire control plane of a specific service, it escalates to a P3 or P2. If a service is entirely unavailable, it becomes a P1. Inhibition and grouping ----------------------- Alerts have dependency awareness. When a parent component fails (for example, a node goes down), inhibition rules in AlertManager suppress child component alerts (for example, pods on that node) to avoid cascading alert storms. The system groups related alerts so that a single notification represents a coherent incident rather than dozens of individual symptoms. Minimizing noise ----------------- Minimizing alert noise keeps every notification meaningful and worthy of action. Every alert should provide enough information to initiate an immediate and effective response, regardless of business hours for high priority alerts. Monitoring and alerting strategies improve over time to better identify and respond to incidents. The ultimate goal is to provide a reliable and high-quality service to all users. Severity levels =============== The alerting system classifies incidents into different severity levels based on their impact on the system and users. **P1**: Critical This level covers incidents causing a complete service disruption or significant loss of functionality across the entire Atmosphere platform. Immediate response, attention, and action are necessary regardless of business hours. Notification: Page to on-call engineer with escalation. **P2**: High This level is for incidents that affect a large group of users or critical system components. These incidents require swift attention and action, regardless of business hours, but don't cause a total disruption. Notification: Page to on-call engineer. **P3**: Moderate This level is for incidents that affect a smaller group of users or a single system. These incidents require attention and may necessitate action during business hours. Notification: Alert channel with mention. **P4**: Low This level covers minor issues that have a limited impact on a small subset of users or system functionality. These incidents require attention and action, if necessary, during standard business hours. Notification: Alert channel, no mention. **P5**: Informational This is the lowest level of severity, used for providing information about normal system activities or minor issues that don't significantly impact users or system functionality. These incidents typically don't require immediate attention or action and operators address them during standard business hours. Notification: Informational channel or email digest. ********************** Operational procedures ********************** Creating silences ================= To create a silence, log in to your Grafana instance that Atmosphere deploys as an admin user. 1. Click on the hamburger menu in the top left corner and select "Alerting" and then "Silences" from the menu. .. image:: images/monitoring-silences-menu.png :alt: Silences menu :width: 200 2. Make sure that you select "AlertManager" on the top right corner of the page, this ensures that you create a silence inside of the AlertManager that's managed by the Prometheus operator instead of the built-in Grafana AlertManager which isn't used. .. image:: images/monitoring-alertmanger-list.png :alt: AlertManager list :width: 200 .. admonition:: AlertManager selection :class: warning It's important that you select the AlertManager that's managed by the Prometheus operator, otherwise your silence won't apply to the Prometheus instance that Atmosphere deploys. 3. Click the "Add Silence" button and use the AlertManager format to create your silence, which you can test by seeing if it matches any alerts in the list labeled "Affected alert instances". .. admonition:: Limit the number of labels :class: info It's important to limit the number of labels that you use in your silence to make sure it continues to work even if the alerts change. For example, if you have an alert that carries the following labels: - ``alertname`` - ``instance`` - ``job`` - ``severity`` You should only use the ``alertname`` and ``severity`` labels in your silence to make sure it continues to work even if the ``instance`` or ``job`` labels change. ************** Configurations ************** Dashboard management ==================== For Grafana, rather than enabling persistence through the application's user interface or manual Helm chart modifications, manage dashboards directly via the Helm chart values. .. admonition:: Avoid Manual Persistence Configurations! :class: warning It's important to avoid manual persistence configurations, especially for services like Grafana, where dashboards and data sources can persist. Such practices aren't captured in version control and pose a risk of data loss, configuration drift, and upgrade complications. To manage Grafana dashboards through Helm, include the dashboard definitions within your configuration file. This approach enables version-controlled dashboard configurations that you can replicate across different deployments without manual intervention. For example, you can define a dashboard in the Helm values like this: .. code-block:: yaml kube_prometheus_stack_helm_values: grafana: dashboards: default: my-dashboard: gnetId: 10000 revision: 1 datasource: Prometheus This instructs Helm to fetch and configure the specified dashboard from `Grafana.com dashboards `_, using Prometheus as the data source. You can find more examples of how to do this in the Grafana Helm chart `Import Dashboards `_ documentation. ************ Viewing data ************ The monitoring stack offers a few different ways to view collected data. The most common ways are through AlertManager, Grafana, and Prometheus. Grafana dashboard ================= By default, Atmosphere creates an ``Ingress`` for Grafana using the ``kube_prometheus_stack_grafana_host`` variable. Keycloak handles authentication, and Atmosphere deploys it by default. Inside Keycloak, Atmosphere creates two client roles for Grafana: ``grafana:admin`` Has access to all organization resources, including dashboards, users, and teams. ``grafana:editor`` Can view and edit dashboards, folders, and playlists. ``grafana:viewer`` Can view dashboards, playlists, and query data sources. You can view the existing dashboards by going to *Manage* > *Dashboards*. You can also check any alerts that are currently firing by going to *Alerting* > *Alerts*. Prometheus ========== By default, Prometheus sits behind an ``Ingress`` using the ``kube_prometheus_stack_prometheus_host`` variable. It also runs behind the `oauth2-proxy` service, which handles authentication so that only authenticated users can access the Prometheus UI. Alternative authentication -------------------------- You can bypass the `oauth2-proxy` service and use an alternative authentication method to access the Prometheus UI. In both cases, you override the ``servicePort`` on the ``Ingress`` to point to the port where Prometheus runs instead of the `oauth2-proxy` service. .. admonition:: Advanced Usage Only :class: warning It's strongly recommended that you stick to keeping the `oauth2-proxy` service in front of the Prometheus UI. The `oauth2-proxy` service is responsible for authenticating users and ensuring that only authenticated users can access the Prometheus UI. Basic authentication ~~~~~~~~~~~~~~~~~~~~ To use basic authentication for the Prometheus UI instead of the `oauth2-proxy` service with single sign-on, make the following changes to your inventory: .. code-block:: yaml kube_prometheus_stack_helm_values: prometheus: ingress: servicePort: 8080 annotations: nginx.ingress.kubernetes.io/auth-type: basic nginx.ingress.kubernetes.io/auth-secret: basic-auth-secret-name In this example, the ``basic-auth-secret-name`` secret handles user authentication. Create the secret in the same namespace as the Prometheus deployment based on the `Ingress NGINX annotations `_. Restricting by address ~~~~~~~~~~~~~~~~~~~~~ To restrict Prometheus UI access to specific IP addresses, make the following changes to your inventory: .. code-block:: yaml kube_prometheus_stack_helm_values: prometheus: ingress: servicePort: 8080 annotations: nginx.ingress.kubernetes.io/whitelist-source-range: "10.0.0.0/24,172.10.0.1" In this example, the configuration restricts access to the IP range ``10.0.0.0/24`` and the IP address ``172.10.0.1``. AlertManager ============ By default, the AlertManager dashboard points to the Ansible variable ``kube_prometheus_stack_alertmanager_host`` and sits behind an ``Ingress`` with the `oauth2-proxy` service, protected by Keycloak similar to Prometheus. ************ Integrations ************ Since Atmosphere relies on AlertManager to send alerts, you can integrate it with services like OpsGenie, PagerDuty, email, and more. To receive monitoring alerts using your preferred notification tools, integrate them with AlertManager. OpsGenie ======== To get started, complete the following steps inside OpsGenie: 1. Create an integration inside OpsGenie, you can do this by going to *Settings* > *Integrations* > *Add Integration* and selecting *Prometheus*. 2. Copy the API key that OpsGenie generates for you and set up correct assignment rules inside OpsGenie. 3. Create a new heartbeat inside OpsGenie, you can do this by going to *Settings* > *Heartbeats* > *Create Heartbeat*. Set the interval to 1 minute. Afterward, configure the following options for the Atmosphere ``config``, making sure that you replace the placeholders with the correct values: ``API_KEY`` The API key that you copied from the OpsGenie integration. ``HEARTBEAT_NAME`` The name of the heartbeat that you created inside OpsGenie .. code-block:: yaml kube_prometheus_stack_helm_values: alertmanager: config: receivers: - name: "null" - name: notifier opsgenie_configs: - api_key: API_KEY message: >- {% raw -%} {{ .GroupLabels.alertname }} {%- endraw %} priority: >- {% raw -%} {{- if eq .GroupLabels.severity "critical" -}} P1 {{- else if eq .GroupLabels.severity "warning" -}} P3 {{- else if eq .GroupLabels.severity "info" -}} P5 {{- else -}} {{ .GroupLabels.severity }} {{- end -}} {%- endraw %} description: |- {% raw -%} {{ if gt (len .Alerts.Firing) 0 -}} Alerts Firing: {{ range .Alerts.Firing }} - Message: {{ .Annotations.message }} Labels: {{ range .Labels.SortedPairs }} - {{ .Name }} = {{ .Value }} {{ end }} Annotations: {{ range .Annotations.SortedPairs }} - {{ .Name }} = {{ .Value }} {{ end }} Source: {{ .GeneratorURL }} {{ end }} {{- end }} {{ if gt (len .Alerts.Resolved) 0 -}} Alerts Resolved: {{ range .Alerts.Resolved }} - Message: {{ .Annotations.message }} Labels: {{ range .Labels.SortedPairs }} - {{ .Name }} = {{ .Value }} {{ end }} Annotations: {{ range .Annotations.SortedPairs }} - {{ .Name }} = {{ .Value }} {{ end }} Source: {{ .GeneratorURL }} {{ end }} {{- end }} {%- endraw %} - name: heartbeat webhook_configs: - url: https://api.opsgenie.com/v2/heartbeats/HEARTBEAT_NAME/ping send_resolved: false http_config: basic_auth: password: API_KEY Once you deploy the changes, you'll start to see alerts inside OpsGenie and you can also verify that the heartbeat shows as *ACTIVE*. PagerDuty ========= To integrate with PagerDuty, first prepare an *Integration key*. To do that, decide how you want to integrate with PagerDuty since there are two ways: **Event Orchestration** This method is beneficial if you want to build different routing rules based on the events coming from the integrated tool. **PagerDuty Service Integration** This method is beneficial if you don't need to route alerts from the integrated tool to different responders based on the event payload. For both of these methods, you need to create an *Integration key* in PagerDuty using the `PagerDuty Integration Guide `_. Once you're done, you'll need to configure the inventory with the following options: .. code-block:: yaml kube_prometheus_stack_helm_values: alertmanager: config: receivers: - name: notifier pagerduty_configs: - service_key: '' You can find more details about ``pagerduty_configs`` in the `Prometheus documentation `_. Email ===== To integrate with email, you need to configure the following options in the inventory: .. code-block:: yaml kube_prometheus_stack_helm_values: alertmanager: config: receivers: - name: notifier email_configs: - smarthost: 'smtp.gmail.com:587' auth_username: '' auth_password: '' from: '' to: '' headers: subject: 'Prometheus Mail Alerts' You can find more details about ``email_configs`` in the `Prometheus documentation `_. **************** Alerts reference **************** ``CoreDNSCriticalErrorBudgetBurn`` ================================== This alert fires when the CoreDNS SERVFAIL rate exceeds 14.4x the burn rate for a 99.9% SLO. At this rate, the 30-day error budget exhausts in under 2.1 days. It uses multi-window burn-rate detection with 1-hour and 5-minute windows. **Likely root causes** - Upstream DNS servers are unreachable or returning errors - CoreDNS configuration errors after a recent change - Network connectivity issues between CoreDNS and upstream resolvers - Resource exhaustion (CPU or memory) on CoreDNS pods **Diagnostic and remediation steps** 1. Check CoreDNS pod health and logs: .. code-block:: console kubectl -n kube-system get pods -l k8s-app=kube-dns kubectl -n kube-system logs -l k8s-app=kube-dns --tail=100 2. Verify upstream DNS server connectivity: .. code-block:: console kubectl -n kube-system exec -it deploy/coredns -- nslookup example.com 3. Check the current SERVFAIL rate: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum(rate(coredns_dns_responses_total{rcode="SERVFAIL"}[5m])) / sum(rate(coredns_dns_responses_total[5m]))' 4. Review CoreDNS ConfigMap for configuration issues: .. code-block:: console kubectl -n kube-system get configmap coredns -o yaml 5. Restart CoreDNS pods if configuration looks correct: .. code-block:: console kubectl -n kube-system rollout restart deployment coredns ``CoreDNSDown`` =============== This alert fires when CoreDNS disappears from Prometheus target discovery for more than 15 minutes. This could indicate crashed pods or an incorrectly configured scrape target. **Likely root causes** - CoreDNS pods crashed or aren't scheduling - Prometheus scrape configuration changed - Node-level issues preventing pod scheduling - Resource limits causing out-of-memory termination **Diagnostic and remediation steps** 1. Check CoreDNS pod status: .. code-block:: console kubectl -n kube-system get pods -l k8s-app=kube-dns kubectl -n kube-system describe pods -l k8s-app=kube-dns 2. Check for out-of-memory events: .. code-block:: console kubectl -n kube-system get events --field-selector reason=OOMKilling 3. Verify Prometheus can reach the CoreDNS metrics endpoint: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 'up{job="coredns"}' 4. Check node availability if pods aren't scheduling: .. code-block:: console kubectl get nodes -o wide ``CoreDNSHighErrorBudgetBurn`` ============================== This alert fires when the CoreDNS SERVFAIL rate exceeds 6x the burn rate for a 99.9% SLO. At this rate, the 30-day error budget exhausts in under 5 days. It uses multi-window burn-rate detection with 6-hour and 30-minute windows. **Likely root causes** - Intermittent upstream DNS server issues - Partial network connectivity problems - DNS zone transfer failures - Upstream rate limiting **Diagnostic and remediation steps** 1. Check CoreDNS logs for recurring errors: .. code-block:: console kubectl -n kube-system logs -l k8s-app=kube-dns --tail=200 | grep -i error 2. Check the SERVFAIL rate broken down over time: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum(rate(coredns_dns_responses_total{rcode="SERVFAIL"}[30m])) / sum(rate(coredns_dns_responses_total[30m]))' 3. Verify upstream DNS server health by testing resolution: .. code-block:: console kubectl -n kube-system exec -it deploy/coredns -- nslookup example.com 4. Review CoreDNS forward plugin configuration: .. code-block:: console kubectl -n kube-system get configmap coredns -o yaml ``CoreDNSLowErrorBudgetBurn`` ============================= This alert fires when the CoreDNS SERVFAIL rate exceeds 1x the burn rate for a 99.9% SLO. At this rate, the error budget exhausts before the 30-day window resets. It uses multi-window burn-rate detection with 3-day and 6-hour windows. **Likely root causes** - Chronic low-level DNS resolution failures - Specific zones or domains consistently failing - Degraded upstream DNS server performance - Incorrectly configured DNS records causing intermittent failures **Diagnostic and remediation steps** 1. Identify which DNS queries are failing: .. code-block:: console kubectl -n kube-system logs -l k8s-app=kube-dns --tail=500 | grep SERVFAIL 2. Check the long-term SERVFAIL trend: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum(rate(coredns_dns_responses_total{rcode="SERVFAIL"}[6h])) / sum(rate(coredns_dns_responses_total[6h]))' 3. Review if specific upstream servers are consistently problematic: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (to) (rate(coredns_forward_responses_total{rcode="SERVFAIL"}[6h]))' ``CoreDNSModerateErrorBudgetBurn`` ================================== This alert fires when the CoreDNS SERVFAIL rate exceeds 3x the burn rate for a 99.9% SLO. At this rate, the 30-day error budget exhausts in under 10 days. It uses multi-window burn-rate detection with 1-day and 2-hour windows. **Likely root causes** - Degraded upstream DNS server responding with errors - Network path issues causing intermittent resolution failures - DNS zone configuration errors causing partial failures - Resource pressure on CoreDNS pods **Diagnostic and remediation steps** 1. Check CoreDNS resource usage: .. code-block:: console kubectl -n kube-system top pods -l k8s-app=kube-dns 2. Review the SERVFAIL rate trend over the last day: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum(rate(coredns_dns_responses_total{rcode="SERVFAIL"}[2h])) / sum(rate(coredns_dns_responses_total[2h]))' 3. Check forward plugin health: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum(rate(coredns_forward_healthcheck_failures_total[2h]))' 4. Review CoreDNS logs for patterns: .. code-block:: console kubectl -n kube-system logs -l k8s-app=kube-dns --tail=300 | grep -i "error\|fail" ``etcdDatabaseHighFragmentationRatio`` ====================================== This alert fires when the etcd database has a high fragmentation ratio that can cause performance issues on the cluster. To resolve this issue, use the following command: .. code-block:: console kubectl -n kube-system exec svc/kube-prometheus-stack-kube-etcd -- \ etcdctl defrag \ --cluster \ --cacert /etc/kubernetes/pki/etcd/ca.crt \ --key /etc/kubernetes/pki/etcd/server.key \ --cert /etc/kubernetes/pki/etcd/server.crt ``EtcdMembersDown`` =================== If any alerts fire from Prometheus for ``etcd`` issues such as ``TargetDown``, ``etcdMembersDown``, or ``etcdInsufficientMembers``, it could be due to expired certificates. You can update the certificates that ``kube-prometheus-stack`` uses for talking with ``etcd`` with the following commands: .. code-block:: console kubectl -n monitoring delete secret/kube-prometheus-stack-etcd-client-cert kubectl -n monitoring create secret generic kube-prometheus-stack-etcd-client-cert \ --from-file=/etc/kubernetes/pki/etcd/ca.crt \ --from-file=/etc/kubernetes/pki/etcd/healthcheck-client.crt \ --from-file=/etc/kubernetes/pki/etcd/healthcheck-client.key ``GoldpingerHighErrorRate`` =========================== This alert fires when more than 5% of Goldpinger ping attempts are failing for at least 15 minutes, indicating network connectivity issues across the cluster. **Likely Root Causes** - Intermittent network connectivity issues - DNS resolution failures - Firewall rules blocking ICMP or health check traffic - Goldpinger pods experiencing resource constraints - Node-level network stack issues **Diagnostic and Remediation Steps** 1. Check error rates by node to identify patterns: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (instance) (rate(goldpinger_errors_total{type="ping"}[5m]))' 2. Verify Goldpinger pods are healthy: .. code-block:: console kubectl -n monitoring get pods -l app.kubernetes.io/name=goldpinger kubectl -n monitoring describe pods -l app.kubernetes.io/name=goldpinger 3. Check for any network policies that might be blocking traffic: .. code-block:: console kubectl get networkpolicies --all-namespaces 4. Review Goldpinger logs for specific error messages: .. code-block:: console kubectl -n monitoring logs -l app.kubernetes.io/name=goldpinger --tail=100 5. Check if the errors correlate with specific target nodes by examining which targets have high latency: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'topk(10, histogram_quantile(0.95, sum by (instance, host_ip, le) (rate(goldpinger_peers_response_time_s_bucket{call_type="ping"}[5m]))))' ``GoldpingerHighPeerLatency`` ============================= This alert fires when the 95th percentile of Goldpinger peer-to-peer latency exceeds 500ms for at least 15 minutes, indicating network congestion or performance issues. **Likely Root Causes** - Network congestion on the cluster network - Overloaded network switches or routers - High CPU or I/O load on nodes causing delayed responses - Network interface saturation - Incorrect network Quality of Service (QoS) configuration **Diagnostic and Remediation Steps** 1. Check which node pairs have the highest latency: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'topk(10, histogram_quantile(0.95, sum by (instance, host_ip, le) (rate(goldpinger_peers_response_time_s_bucket{call_type="ping"}[5m]))))' 2. Check node-exporter metrics for network saturation: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'rate(node_network_transmit_bytes_total[5m])' 3. Review CPU and I/O wait metrics on affected nodes: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (instance) (rate(node_cpu_seconds_total{mode="iowait"}[5m]))' 4. Check for network drops which may indicate congestion: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'rate(node_network_receive_drop_total[5m]) > 0' 5. If latency is consistently high between specific node pairs, investigate the network path between them for hardware issues or configuration errors. ``GoldpingerHighUnhealthyRatio`` ================================ This alert fires when more than 10% of nodes in the cluster are reporting as unhealthy according to Goldpinger health checks for at least 5 minutes. **Likely Root Causes** - Widespread network connectivity issues affecting multiple nodes - A network partition isolating a segment of the cluster - Multiple nodes experiencing high load or resource exhaustion - Infrastructure-level network problems (switch failures, routing issues) **Diagnostic and Remediation Steps** 1. Check the Goldpinger dashboard in Grafana to visualize which nodes are affected and the connectivity patterns. 2. Query Prometheus to identify which specific nodes are reporting unhealthy: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'goldpinger_nodes_health_total{status="unhealthy"} > 0' 3. Check for network issues on the affected nodes: .. code-block:: console kubectl get nodes -o wide kubectl describe node 4. Review node-exporter metrics for network errors or drops: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'rate(node_network_receive_errs_total[5m]) > 0' 5. If the issue affects only specific nodes, check their network configuration and physical connectivity. ``GoldpingerNodeUnreachable`` ============================= This alert fires when more than 50% of Goldpinger instances can't reach a specific node for at least 5 minutes. The target node may be down or experiencing network issues. **Likely Root Causes** - The target node is down or unresponsive - Network interface failure on the target node - Firewall or security group configuration error blocking traffic - The Goldpinger pod on the target node has crashed **Diagnostic and Remediation Steps** 1. Identify the affected node by its IP address from the alert labels: .. code-block:: console kubectl get nodes -o wide | grep 2. Check if the node is reachable and healthy: .. code-block:: console kubectl get node kubectl describe node 3. Verify the Goldpinger pod is running on the affected node: .. code-block:: console kubectl -n monitoring get pods -l app.kubernetes.io/name=goldpinger \ -o wide | grep 4. Check network connectivity from another node: .. code-block:: console kubectl debug node/ -it --image=busybox -- \ ping -c 3 5. Review system logs on the affected node for network or kernel issues: .. code-block:: console kubectl debug node/ -it --image=busybox -- \ cat /host/var/log/syslog | tail -100 ``MySQLGaleraOutOfSync`` ======================== This alert fires when a Percona XtraDB Cluster (PXC) Galera node has a ``wsrep_local_state`` that's not 4 (Synced) and not 2 (Donor), while ``wsrep_desync`` isn't enabled. The alert excludes state 2 (Donor) because it's a normal transient state during donor operations for State Snapshot Transfer (SST) or PXC operator backups. The separate ``MySQLGaleraDonorFallingBehind`` alert covers problematic donor scenarios. This is a cause-based alert because no reasonable symptom-based proxy exists for a Galera node losing cluster sync. The cluster may continue to serve requests with the remaining nodes, but the reduced quorum margin increases the risk of a full outage if another node fails. **Likely Root Causes** - A node remains in the Joining state (1) and can't complete SST. - A node completed SST but hasn't finished catching up (state 3, Joined) for an extended period. - Network partitioning prevents the node from rejoining the cluster. - Corrupted Galera cache or write-set replication failure. **Diagnostic and Remediation Steps** 1. Check the current ``wsrep_local_state`` on the affected node: .. code-block:: console kubectl -n openstack exec -it percona-xtradb-pxc-0 -- \ mysql -e "SHOW STATUS LIKE 'wsrep_local_state%';" 2. Check overall cluster status: .. code-block:: console kubectl -n openstack exec -it percona-xtradb-pxc-0 -- \ mysql -e "SHOW STATUS LIKE 'wsrep_cluster%';" 3. Review the PXC pod logs for the affected replica: .. code-block:: console kubectl -n openstack logs percona-xtradb-pxc- --tail=200 4. If the node remains in Joining state, it may need an SST restart. Delete the affected pod to trigger a fresh SST: .. code-block:: console kubectl -n openstack delete pod percona-xtradb-pxc- 5. Verify that no PXC backup runs at the moment (a backup puts a node into Donor state, which the ``MySQLGaleraDonorFallingBehind`` alert covers): .. code-block:: console kubectl -n openstack get pxc-backup ``NodeDiskHighLatency`` ======================= This alert fires when the average IO latency on a disk device exceeds 20ms for at least 1 hour. It measures the time the kernel spends servicing reads and writes divided by the number of completed operations, which represents the true per-operation latency at the block device layer. **Likely root causes:** - Failing or degraded SSD or HDD (wear-out, bad sectors, firmware issues) - RAID array running in degraded mode after a disk failure - Storage controller problems - Severely overloaded storage subsystem (too many concurrent IO operations) - Incorrect IO scheduler for the workload type **Diagnostic and remediation steps:** 1. Identify the affected host and device from the alert labels (``instance`` and ``device``). 2. Check the current IO latency and throughput on the affected device: .. code-block:: console iostat -xz 1 5 3. Check for disk errors in the kernel log: .. code-block:: console dmesg | grep -i -E "error|fault|reset|i/o" | tail -30 4. If the device is part of a RAID array, check the array status: .. code-block:: console cat /proc/mdstat 5. Check the SMART health status of the underlying drives: .. code-block:: console smartctl -a /dev/ 6. Review whether the host is under unusual IO load: .. code-block:: console iotop -aoP 7. If the disk shows degradation or failure, plan a replacement. For RAID arrays, replace the failed member. For standalone disks, migrate workloads before the disk fails completely. ``NginxIngressCriticalErrorBudgetBurn`` ======================================= This alert fires when a service behind NGINX Ingress consumes its 30-day error budget at more than 14.4x the sustainable rate, based on a 99.9% availability SLO. It uses multi-window burn-rate detection with 1-hour and 5-minute windows to confirm the issue is both sustained and ongoing. At this burn rate, the entire 30-day error budget exhausts in under 2.1 days. A minimum traffic guard of 1 request per second prevents false positives on low-traffic services. **Likely Root Causes** - Service pods are crashing or in a crash loop - Database connection failures affecting all service replicas - Configuration errors in the service deployment - Resource exhaustion (CPU, memory, or file descriptors) on service pods - Network connectivity failures between NGINX and service pods - Service code bugs causing widespread failures **Diagnostic and Remediation Steps** 1. Check the alert labels to identify the affected service and query the current error rate: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (service) (rate(nginx_ingress_controller_requests{status=~"5[0-9]{2}"}[1h])) / sum by (service) (rate(nginx_ingress_controller_requests[1h]))' 2. Check the health and status of the backend service pods: .. code-block:: console kubectl get pods -A | grep kubectl describe pod -n 3. Review recent logs from the affected service pods: .. code-block:: console kubectl logs -n --tail=100 4. Check NGINX Ingress controller logs for upstream connection errors: .. code-block:: console kubectl logs -n ingress-nginx -l app.kubernetes.io/name=ingress-nginx --tail=200 5. Verify resource usage on the backend service pods: .. code-block:: console kubectl top pods -n | grep 6. If the issue persists, consider scaling the service or restarting affected pods to attempt recovery while investigating the root cause. ``NginxIngressHighErrorBudgetBurn`` ==================================== This alert fires when a service behind NGINX Ingress consumes its 30-day error budget at more than 6x the sustainable rate, based on a 99.9% availability SLO. It uses multi-window burn-rate detection with 6-hour and 30-minute windows to confirm the issue is both sustained and ongoing. At this burn rate, the entire 30-day error budget exhausts in under 5 days. A minimum traffic guard of 1 request per second prevents false positives on low-traffic services. **Likely Root Causes** - Intermittent issues with a subset of service replicas - Occasional database query timeouts or connection pool exhaustion - Resource pressure on some service pods - Networking issues affecting specific nodes or pods - Recent deployment causing partial service degradation - Cache or session storage issues **Diagnostic and Remediation Steps** 1. Check the current error rate by service: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (service) (rate(nginx_ingress_controller_requests{status=~"5[0-9]{2}"}[6h])) / sum by (service) (rate(nginx_ingress_controller_requests[6h]))' 2. Identify which specific HTTP error codes the service returns: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (service, status) (rate(nginx_ingress_controller_requests{status=~"5[0-9]{2}"}[30m]))' 3. Check for unhealthy or recently restarted pods: .. code-block:: console kubectl get pods -A | grep kubectl get events -n --sort-by='.lastTimestamp' | grep 4. Review application logs for errors: .. code-block:: console kubectl logs -n -l app= --tail=100 | grep -i error 5. Check if the error rate correlates with increased load or specific API endpoints by examining the NGINX Ingress metrics in Grafana. 6. Verify database connectivity and performance if the service depends on a database: .. code-block:: console kubectl exec -it -n -- -e "SHOW PROCESSLIST;" ``NginxIngressLowErrorBudgetBurn`` =================================== This alert fires when a service behind NGINX Ingress consumes its 30-day error budget at the sustainable rate or faster, based on a 99.9% availability SLO. It uses multi-window burn-rate detection with 3-day and 6-hour windows to confirm the issue is both sustained and ongoing. At this burn rate, the 30-day error budget exhausts before the window resets. A minimum traffic guard of 1 request per second prevents false positives on low-traffic services. **Likely Root Causes** - Degrading dependency (database, cache, or external service) - Gradual resource leak causing occasional failures - Configuration drift across replicas - Intermittent infrastructure issues during off-peak hours - Elevated baseline error rate after a deployment **Diagnostic and Remediation Steps** 1. Check the error rate trend over the last 3 days: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (service) (rate(nginx_ingress_controller_requests{status=~"5[0-9]{2}"}[3d])) / sum by (service) (rate(nginx_ingress_controller_requests[3d]))' 2. Compare the error rate with the previous period to identify when it started increasing by examining trends in Grafana dashboards. 3. Review recent deployments or configuration changes that correlate with the error rate increase: .. code-block:: console kubectl get events -n --sort-by='.lastTimestamp' | head -50 4. Check if specific error codes dominate: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (service, status) (rate(nginx_ingress_controller_requests{status=~"5[0-9]{2}"}[6h]))' 5. Investigate resource usage and dependency health for the affected service. ``NginxIngressModerateErrorBudgetBurn`` ======================================== This alert fires when a service behind NGINX Ingress consumes its 30-day error budget at more than 3x the sustainable rate, based on a 99.9% availability SLO. It uses multi-window burn-rate detection with 1-day and 2-hour windows to confirm the issue is both sustained and ongoing. At this burn rate, the entire 30-day error budget exhausts in under 10 days. A minimum traffic guard of 1 request per second prevents false positives on low-traffic services. **Likely Root Causes** - A subset of service replicas returning errors intermittently - Slow database queries causing periodic timeouts - Resource contention during peak traffic periods - Partial deployment with a faulty version still in rotation - Upstream dependency experiencing intermittent issues **Diagnostic and Remediation Steps** 1. Check the error rate over the last day: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (service) (rate(nginx_ingress_controller_requests{status=~"5[0-9]{2}"}[1d])) / sum by (service) (rate(nginx_ingress_controller_requests[1d]))' 2. Identify which error codes the service returns: .. code-block:: console kubectl -n monitoring exec svc/kube-prometheus-stack-prometheus -- \ promtool query instant http://localhost:9090 \ 'sum by (service, status) (rate(nginx_ingress_controller_requests{status=~"5[0-9]{2}"}[2h]))' 3. Check for unhealthy pods or recent restarts: .. code-block:: console kubectl get pods -A | grep kubectl get events -n --sort-by='.lastTimestamp' | grep 4. Review application logs for recurring errors: .. code-block:: console kubectl logs -n -l app= --tail=200 | grep -i error 5. Check if errors correlate with traffic patterns or specific time windows in Grafana. 6. Review recent deployments or configuration changes that might have introduced the issue. ``NodeNetworkMulticast`` ======================== This alert fires when a node receives large volumes of multicast traffic, which can indicate an incorrectly configured network or a malicious actor. This can result in high CPU usage on the node and can cause the node to become unresponsive. It can also cause a high amount of software interrupts on the node. To find the root cause of this issue, use the following commands: .. code-block:: console iftop -ni $DEV -f 'multicast and not broadcast' With this command, you can see which IP addresses send the multicast traffic. Once you have the IP address, use the following command to find the server behind it: .. code-block:: console openstack server list --all-projects --long -n --ip $IP