>   routes:
>   - matchers:
>      
 
alertname:['HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady']

That's invalid: alertmanager should not even start.  I tested your config, 
and I get the following error:

ts=2023-02-27T10:17:54.702Z caller=coordinator.go:118 level=error 
component=configuration msg="Loading configuration file failed" 
file=tmp.yaml err="yaml: unmarshal errors:\n  line 22: cannot unmarshal 
!!str `alertna...` into []string"

'matchers' is a list of strings, not a map.  This should work:

route:
  routes:
  - matchers:
    - 
alertname=~"HostOutOfDiskSpace|HostHighCpuLoad|HostHighCpuLoad|KubeNodeNotReady"
  receiver: elevate_alerts

See:
https://prometheus.io/docs/alerting/latest/configuration/#matcher
https://prometheus.io/docs/alerting/latest/configuration/#example

On Sunday, 26 February 2023 at 14:53:26 UTC Sampada Thorat wrote:

> Hello Everyone,
>
> I want to receive Alerts for  
> 'HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady' 
> alertnames in "elevate_alerts" channel and rest all other alerts in 
> " default_receiver_test" channel. But for the below configuration, I'm 
> getting all the alerts in   "elevate_alerts" only.
>
> This is my ConfigMap:
>
> apiVersion: v1
> data:
>   connectors.yaml: |
>     connectors:
>       - test: 
> https://sasoffice365.webhook.office.com/webhookb2/d2415be1-2360-49c3-af48-7baf41aa1371@b1c14d5c-3625-45b3-a430-9552373a0c2f/IncomingWebhook/c7c62c1315d24c1fb5d1c731d2467dc6/5c8c1e6c-e827-4114-a893-9a1788ad41b5
>       - alertmanager: 
> https://sasoffice365.webhook.office.com/webhookb2/a7cb86de-1543-4e6d-b927-387c1f1e35ad@b1c14d5c-3625-45b3-a430-9552373a0c2f/IncomingWebhook/687a7973ffe248d081f58d94a090fb4c/05be66ae-90eb-42f5-8e0c-9c10975012ca
> kind: ConfigMap
> metadata:
>   annotations:
>     meta.helm.sh/release-name: prometheus-msteams
>     meta.helm.sh/release-namespace: monitoring
>   creationTimestamp: "2023-02-26T12:33:36Z"
>   labels:
>     app.kubernetes.io/managed-by: Helm
>   name: prometheus-msteams-config
>   namespace: monitoring
>   resourceVersion: "18040490"
>   uid: 795c96d5-8318-4885-804f-71bba707c885
>
>
> This is my alertmanager.yaml:
>
> global:
>   resolve_timeout: 5m
> receivers:
> - name: elevate_alerts
>   webhook_configs:
>   - url: "
> http://prometheus-msteams.default.svc.cluster.local:2000/alertmanager";
>     send_resolved: true
> - name: default_receiver_test
>   webhook_configs:
>   - url: "http://prometheus-msteams.default.svc.cluster.local:2000/test";
>     send_resolved: true
> route:
>   group_by:
>   - alertname
>   - severity
>   group_interval: 5m
>   group_wait: 30s
>   repeat_interval: 3h
>   receiver: default_receiver_test
>   routes:
>   - matchers:
>       
> alertname:['HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady']
>     receiver: elevate_alerts
>
> Please help
>
>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/bccc531d-5c21-4de3-8371-67b7256eb148n%40googlegroups.com.

Reply via email to