> routes:
> - matchers:
>
alertname:['HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady']
That's invalid: alertmanager should not even start. I tested your config,
and I get the following error:
ts=2023-02-27T10:17:54.702Z caller=coordinator.go:118 level=error
component=configuration msg="Loading configuration file failed"
file=tmp.yaml err="yaml: unmarshal errors:\n line 22: cannot unmarshal
!!str `alertna...` into []string"
'matchers' is a list of strings, not a map. This should work:
route:
routes:
- matchers:
-
alertname=~"HostOutOfDiskSpace|HostHighCpuLoad|HostHighCpuLoad|KubeNodeNotReady"
receiver: elevate_alerts
See:
https://prometheus.io/docs/alerting/latest/configuration/#matcher
https://prometheus.io/docs/alerting/latest/configuration/#example
On Sunday, 26 February 2023 at 14:53:26 UTC Sampada Thorat wrote:
> Hello Everyone,
>
> I want to receive Alerts for
> 'HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady'
> alertnames in "elevate_alerts" channel and rest all other alerts in
> " default_receiver_test" channel. But for the below configuration, I'm
> getting all the alerts in "elevate_alerts" only.
>
> This is my ConfigMap:
>
> apiVersion: v1
> data:
> connectors.yaml: |
> connectors:
> - test:
> https://sasoffice365.webhook.office.com/webhookb2/d2415be1-2360-49c3-af48-7baf41aa1371@b1c14d5c-3625-45b3-a430-9552373a0c2f/IncomingWebhook/c7c62c1315d24c1fb5d1c731d2467dc6/5c8c1e6c-e827-4114-a893-9a1788ad41b5
> - alertmanager:
> https://sasoffice365.webhook.office.com/webhookb2/a7cb86de-1543-4e6d-b927-387c1f1e35ad@b1c14d5c-3625-45b3-a430-9552373a0c2f/IncomingWebhook/687a7973ffe248d081f58d94a090fb4c/05be66ae-90eb-42f5-8e0c-9c10975012ca
> kind: ConfigMap
> metadata:
> annotations:
> meta.helm.sh/release-name: prometheus-msteams
> meta.helm.sh/release-namespace: monitoring
> creationTimestamp: "2023-02-26T12:33:36Z"
> labels:
> app.kubernetes.io/managed-by: Helm
> name: prometheus-msteams-config
> namespace: monitoring
> resourceVersion: "18040490"
> uid: 795c96d5-8318-4885-804f-71bba707c885
>
>
> This is my alertmanager.yaml:
>
> global:
> resolve_timeout: 5m
> receivers:
> - name: elevate_alerts
> webhook_configs:
> - url: "
> http://prometheus-msteams.default.svc.cluster.local:2000/alertmanager"
> send_resolved: true
> - name: default_receiver_test
> webhook_configs:
> - url: "http://prometheus-msteams.default.svc.cluster.local:2000/test"
> send_resolved: true
> route:
> group_by:
> - alertname
> - severity
> group_interval: 5m
> group_wait: 30s
> repeat_interval: 3h
> receiver: default_receiver_test
> routes:
> - matchers:
>
> alertname:['HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady']
> receiver: elevate_alerts
>
> Please help
>
>
>
>
--
You received this message because you are subscribed to the Google Groups
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/prometheus-users/bccc531d-5c21-4de3-8371-67b7256eb148n%40googlegroups.com.