Alerts reworked

Change alerts names, severities and descriptions.

Change-Id: Ib06f08a6f336d28592d5f70e97aedfeb12eb603c
Closes-bug: PROD-19698
diff --git a/prometheus/map.jinja b/prometheus/map.jinja
index abc2a73..4df2a85 100644
--- a/prometheus/map.jinja
+++ b/prometheus/map.jinja
@@ -29,11 +29,8 @@
 {%- set monitoring = salt['grains.filter_by']({
   'default': {
     'remote_storage_adapter': {
-      'sent_vs_received_ratio': 10.0,
-      'ignored_vs_sent_ratio': 5.0,
-    },
-    'alertmanager': {
-      'notification_failed_rate': 0.3
+      'sent_vs_received_ratio': 0.9,
+      'ignored_vs_sent_ratio': 0.05,
     },
     'prometheus': {
       'remote_storage_queue_full_percent': 75.0,
@@ -51,3 +48,4 @@
     'config_dir': '/srv/volumes/local/alerta',
   },
 }, merge=salt['pillar.get']('prometheus:alerta')) %}}
+
diff --git a/prometheus/meta/prometheus.yml b/prometheus/meta/prometheus.yml
index 5655064..5c78c64 100644
--- a/prometheus/meta/prometheus.yml
+++ b/prometheus/meta/prometheus.yml
@@ -5,45 +5,55 @@
 {%- if server.get('enabled', False) %}
 {% raw %}
     PrometheusTargetDown:
-      if: 'up != 1'
+      if: up != 1
       for: 2m
       labels:
         severity: critical
         service: prometheus
       annotations:
-        summary: 'Prometheus endpoint {{ $labels.instance }} down'
-        description: 'The Prometheus target {{ $labels.instance }} is down for the job {{ $labels.job }}.'
+        summary: "Prometheus target is down"
+        description: "The Prometheus target for the {{ $labels.job }} job on the {{ $labels.host or $labels.instance }} node is down for at least 2 minutes."
+    PrometheusTargetSamplesOrderWarning:
+      if: increase(prometheus_target_scrapes_sample_out_of_order_total[1m]) > 0
+      labels:
+        severity: warning
+        service: prometheus
+      annotations:
+        summary: "Prometheus samples are out of order"
+        description: "{{ $value }} last-minute samples of Prometheus on the {{ $labels.instance }} instance are out of order."
+    PrometheusTargetSamplesBoundsWarning:
+      if: increase(prometheus_target_scrapes_sample_out_of_bounds_total[1m]) > 0
+      labels:
+        severity: warning
+        service: prometheus
+      annotations:
+        summary: "Prometheus samples timestamps are out of bounds"
+        description: "{{ $value }} last-minute samples of Prometheus on the {{ $labels.instance }} instance have timestamps out of bounds."
+    PrometheusTargetSamplesDuplicateWarning:
+      if: increase(prometheus_target_scrapes_sample_duplicate_timestamp_total[1m]) > 0
+      labels:
+        severity: warning
+        service: prometheus
+      annotations:
+        summary: "Prometheus samples have duplicate timestamps"
+        description: "{{ $value }} last-minute samples of Prometheus on the {{ $labels.instance }} instance have duplicate timestamps."
 {% endraw %}
 {%- if server.version == 1.7 %}
 {% raw %}
-    PrometheusRushMode:
-      if: 'prometheus_local_storage_rushed_mode != 0'
+    PrometheusDataIngestionWarning:
+      if: prometheus_local_storage_rushed_mode != 0
       for: 10m
       labels:
         severity: warning
         service: prometheus
       annotations:
-        summary: 'Prometheus {{ $labels.instance }} in rush mode'
-        description: 'The Prometheus {{ $labels.instance }} is in rush mode for 10m.'
+        summary: "Prometheus is in the rushed mode"
+        description: "The Prometheus service writes on the {{ $labels.instance }} instance do not keep up with data ingestion speed for at least 10 minutes."
 {% endraw %}
 {%- endif %}
 {%- endif %}
-{%- if alertmanager.get('enabled', False) %}
-    AlertmanagerNotificationFailed:
-      {%- set threshold = monitoring.alertmanager.notification_failed_rate|float %}
-      if: >-
-        rate(alertmanager_notifications_failed_total[5m]) > {{ threshold }}
-      for: 2m
-{%- raw %}
-      labels:
-        severity: warning
-        service: alertmanager
-      annotations:
-        summary: 'Alertmanager {{ $labels.instance }} failed notifications'
-        description: 'Alertmanager {{ $labels.instance }} failed notifications for {{ $labels.integration }} (current value={{ $value }}, threshold={%- endraw %}{{ threshold }})'
-{%- endif %}
 {%- if server.get('config', {}).get('remote_write') %}
-    PrometheusRemoteStorageQueue:
+    PrometheusRemoteStorageQueueFullWarning:
       {%- set threshold = monitoring.prometheus.remote_storage_queue_full_percent|float %}
       if: >-
         prometheus_remote_storage_queue_length / prometheus_remote_storage_queue_capacity * 100 > {{ threshold }}
@@ -53,31 +63,59 @@
         severity: warning
         service: prometheus
       annotations:
-        summary: 'Prometheus {{ $labels.instance }} remote storage queue is filling'
-        description: 'The Prometheus {{ $labels.instance }} remote storage queue almost full (current value={{ $value }}%, threshold={%- endraw %}{{ threshold }}%)'
+        summary: "Prometheus remote storage queue is full in {%- endraw %} {{ threshold }}{%- raw %}%"
+        description: "The Prometheus remote storage queue on the {{ $labels.instance }} instance is {{ $value }}% full for at least 2 minutes."
+{%- endraw %}
 {%- endif %}
 {%- if remote_storage_adapter.get('enabled', False) %}
-    RemoteStorageAdapterSendingTooSlow:
+    RemoteStorageAdapterMetricsSendingWarning:
       {%- set threshold = monitoring.remote_storage_adapter.sent_vs_received_ratio|float %}
       if: >-
-        100.0 - (100.0 * sent_samples_total{job="remote_storage_adapter"} / on (job, instance) received_samples_total) > {{ threshold }}
+        increase(sent_samples_total{job="remote_storage_adapter"}[1m]) / on (job, instance) increase(received_samples_total[1m]) < {{ threshold }}
 {% raw %}
       labels:
         severity: warning
         service: remote_storage_adapter
       annotations:
-        summary: 'Remote storage adapter too slow on {{ $labels.instance }}'
-        description: 'Remote storage adapter can not ingest samples fast enough on {{ $labels.instance }} (current value={{ $value }}%, threshold={%- endraw %}{{ threshold }}%).'
-    RemoteStorageAdapterIgnoredTooHigh:
+        summary: "Ratio of sent to received remote storage adapter metrics is {%- endraw %} {{ threshold }}{%- raw %}"
+        description: "The ratio of the sent to received metrics of the remote storage adapter on the {{ $labels.instance }} instance is {{ $value }}."
+{% endraw %}
+    RemoteStorageAdapterMetricsIgnoredWarning:
       {%- set threshold = monitoring.remote_storage_adapter.ignored_vs_sent_ratio|float %}
       if: >-
-        100.0 * prometheus_influxdb_ignored_samples_total{job="remote_storage_adapter"} / on (job, instance) sent_samples_total > {{ threshold }}
+        increase(prometheus_influxdb_ignored_samples_total{job="remote_storage_adapter"}[1m]) / on (job, instance) increase(sent_samples_total[1m]) >= {{ threshold }}
 {% raw %}
       labels:
         severity: warning
         service: remote_storage_adapter
       annotations:
-        summary: 'Remote storage adapter receiving too many invalid metrics on {{ $labels.instance }}'
-        description: 'Remote storage adapter is receiving too many invalid metrics on {{ $labels.instance }} (current value={{ $value }}%, threshold={%- endraw %}{{ threshold }}%).'
+        summary: "{%- endraw %}{{ threshold }}{%- raw %}% of remote storage adapter metrics are invalid"
+        description: "{{ $value }}% of remote storage adapter metrics on the {{ $labels.instance }} instance are invalid."
+{%- endraw %}
+{%- endif %}
+{%- if alertmanager.get('enabled', False) %}
+{%- raw %}
+    AlertmanagerNotificationFailureWarning:
+      if: >-
+        increase(alertmanager_notifications_failed_total[2m]) > 0
+      for: 2m
+      labels:
+        severity: warning
+        service: alertmanager
+      annotations:
+        summary: "Alertmanager notifications fail"
+        description: "An average of {{ $value }} Alertmanager {{ $labels.integration }} notifications on the {{ $labels.instance }} instance fail for at least 2 minutes."
+    AlertmanagerAlertsInvalidWarning:
+      if: >-
+        increase(alertmanager_alerts_invalid_total[2m]) > 0
+      for: 2m
+      labels:
+        severity: warning
+        service: alertmanager
+      annotations:
+        summary: "Alertmanager alerts are invalid"
+        description: "An average of {{ $value }} Alertmanager {{ $labels.integration }} alerts on the {{ $labels.instance }} instance are invalid for at least 2 minutes."
+{%- endraw %}
 {%- endif %}
 {%- endif %}
+