Hello Brian I tried your change yet my alertmanager isn't taking config
changes and shows older config. Can u have a look ?
global:
resolve_timeout: 5m
receivers:
- name: pdmso_alerts
webhook_configs:
- url:
"http://prometheus-msteams.monitoring.svc.cluster.local:2000/pdmsoalert"
send_resolved: true
- name: default_receiver_test
webhook_configs:
- url:
"http://prometheus-msteams.monitoring.svc.cluster.local:2000/test"
send_resolved: true
route:
group_by:
- alertname
- severity
group_interval: 5m
group_wait: 30s
repeat_interval: 3h
receiver: default_receiver_test
routes:
- matchers:
alertname=~"HostOutOfDiskSpace|HostHighCpuLoad|HostHighCpuLoad|KubeNodeNotReady"
receiver: pdmso_alerts
On Monday, February 27, 2023 at 3:51:33 PM UTC+5:30 Brian Candler wrote:
> > routes:
> > - matchers:
> >
>
> alertname:['HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady']
>
> That's invalid: alertmanager should not even start. I tested your config,
> and I get the following error:
>
> ts=2023-02-27T10:17:54.702Z caller=coordinator.go:118 level=error
> component=configuration msg="Loading configuration file failed"
> file=tmp.yaml err="yaml: unmarshal errors:\n line 22: cannot unmarshal
> !!str `alertna...` into []string"
>
> 'matchers' is a list of strings, not a map. This should work:
>
> route:
> routes:
> - matchers:
> -
> alertname=~"HostOutOfDiskSpace|HostHighCpuLoad|HostHighCpuLoad|KubeNodeNotReady"
> receiver: elevate_alerts
>
> See:
> https://prometheus.io/docs/alerting/latest/configuration/#matcher
> https://prometheus.io/docs/alerting/latest/configuration/#example
>
> On Sunday, 26 February 2023 at 14:53:26 UTC Sampada Thorat wrote:
>
>> Hello Everyone,
>>
>> I want to receive Alerts for
>> 'HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady'
>> alertnames in "elevate_alerts" channel and rest all other alerts in
>> " default_receiver_test" channel. But for the below configuration, I'm
>> getting all the alerts in "elevate_alerts" only.
>>
>> This is my ConfigMap:
>>
>> apiVersion: v1
>> data:
>> connectors.yaml: |
>> connectors:
>> - test:
>> https://sasoffice365.webhook.office.com/webhookb2/d2415be1-2360-49c3-af48-7baf41aa1371@b1c14d5c-3625-45b3-a430-9552373a0c2f/IncomingWebhook/c7c62c1315d24c1fb5d1c731d2467dc6/5c8c1e6c-e827-4114-a893-9a1788ad41b5
>> - alertmanager:
>> https://sasoffice365.webhook.office.com/webhookb2/a7cb86de-1543-4e6d-b927-387c1f1e35ad@b1c14d5c-3625-45b3-a430-9552373a0c2f/IncomingWebhook/687a7973ffe248d081f58d94a090fb4c/05be66ae-90eb-42f5-8e0c-9c10975012ca
>> kind: ConfigMap
>> metadata:
>> annotations:
>> meta.helm.sh/release-name: prometheus-msteams
>> meta.helm.sh/release-namespace: monitoring
>> creationTimestamp: "2023-02-26T12:33:36Z"
>> labels:
>> app.kubernetes.io/managed-by: Helm
>> name: prometheus-msteams-config
>> namespace: monitoring
>> resourceVersion: "18040490"
>> uid: 795c96d5-8318-4885-804f-71bba707c885
>>
>>
>> This is my alertmanager.yaml:
>>
>> global:
>> resolve_timeout: 5m
>> receivers:
>> - name: elevate_alerts
>> webhook_configs:
>> - url: "
>> http://prometheus-msteams.default.svc.cluster.local:2000/alertmanager"
>> send_resolved: true
>> - name: default_receiver_test
>> webhook_configs:
>> - url: "http://prometheus-msteams.default.svc.cluster.local:2000/test"
>> send_resolved: true
>> route:
>> group_by:
>> - alertname
>> - severity
>> group_interval: 5m
>> group_wait: 30s
>> repeat_interval: 3h
>> receiver: default_receiver_test
>> routes:
>> - matchers:
>>
>> alertname:['HostOutOfDiskSpace','HostHighCpuLoad','HostOutOfMemory','KubeNodeNotReady']
>> receiver: elevate_alerts
>>
>> Please help
>>
>>
>>
>>
--
You received this message because you are subscribed to the Google Groups
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To view this discussion on the web visit
https://groups.google.com/d/msgid/prometheus-users/d93c30d6-e4b6-4b8b-8913-0c1c84662fd3n%40googlegroups.com.