This is documentation for the next version of Mimir. For the latest stable release, go to the latest version.

Open source

(Optional) Grafana Mimir Alertmanager

The Mimir Alertmanager adds multi-tenancy support and horizontal scalability to the Prometheus Alertmanager. The Mimir Alertmanager is an optional component that accepts alert notifications from the Mimir ruler. The Alertmanager deduplicates and groups alert notifications, and routes them to a notification channel, such as email, PagerDuty, or OpsGenie.


Like other Mimir components, multi-tenancy in the Mimir Alertmanager uses the tenant ID header. Each tenant has an isolated alert routing configuration and Alertmanager UI.

Tenant configurations

Each tenant has an Alertmanager configuration that defines notifications receivers and alerting routes. The Mimir Alertmanager uses the same configuration file that the Prometheus Alertmanager uses.

Note: The Mimir Alertmanager exposes the configuration API according to the path set by the -server.path-prefix flag. It does not use the path set by the -http.alertmanager-http-prefix flag. If you run Mimir with the default configuration (-server.path-prefix, where the default value is /), then only set the hostname for the --address flag of the mimirtool command; do not set a path-specific address. For example, / is correct, and /alertmanager is incorrect.

You can validate a configuration file using the mimirtool command:

mimirtool alertmanager verify <ALERTMANAGER CONFIGURATION FILE>

The following sample command shows how to upload a tenant’s Alertmanager configuration using mimirtool:

mimirtool alertmanager load <ALERTMANAGER CONFIGURATION FILE>  \
  --address=<ALERTMANAGER URL>
  --id=<TENANT ID>

The following sample command shows how to retrieve a tenant’s Alertmanager configuration using mimirtool:

mimirtool alertmanager get \
  --address=<ALERTMANAGER URL>
  --id=<TENANT ID>

The following sample commands shows how to delete a tenant’s Alertmanager configuration using mimirtool:

mimirtool alertmanager delete \
  --address=<ALERTMANAGER URL>
  --id=<TENANT ID>

After the tenant uploads an Alertmanager configuration, the tenant can access the Alertmanager UI at the /alertmanager endpoint.

Fallback configuration

When a tenant doesn’t have a Alertmanager configuration, the Grafana Mimir Alertmanager uses a fallback configuration. By default, there is always a fallback configuration set. You can overwrite the default fallback configuration via the -alertmanager.configs.fallback command-line flag.

Warning: Without a fallback configuration or a tenant specific configuration, the Alertmanager UI is inaccessible and ruler notifications for that tenant fail.

Tenant limits

The Grafana Mimir Alertmanager has a number of per-tenant limits documented in limits. Each Mimir Alertmanager limit configuration parameter has an alertmanager prefix.

Alertmanager UI

The Mimir Alertmanager exposes the same web UI as the Prometheus Alertmanager at the /alertmanager endpoint.

When running Grafana Mimir with multi-tenancy enabled, the Alertmanager requires that any HTTP request include the tenant ID header. Tenants only see alerts sent to their Alertmanager.

For a complete reference of the tenant ID header and Alertmanager endpoints, refer to HTTP API.

You can configure the HTTP path prefix for the UI and the HTTP API:

  • -http.alertmanager-http-prefix configures the path prefix for Alertmanager endpoints.
  • -alertmanager.web.external-url configures the source URLs generated in Alertmanager alerts and from where to fetch web assets.

Note: Unless you are using a reverse proxy in front of the Alertmanager API that rewrites routes, the path prefix set in -alertmanager.web.external-url must match the path prefix set in -http.alertmanager-http-prefix (/alertmanager by default). If the path prefixes do not match, HTTP requests routing might not work as expected.

Using a reverse proxy

When using a reverse proxy, use the following settings when you configure the HTTP path:

  • Set -http.alertmanager-http-prefix to match the proxy path in your reverse proxy configuration.
  • Set -alertmanager.web.external-url to the URL served by your reverse proxy.


The Mimir Alertmanager adds some custom template functions to the default ones of the Prometheus Alertmanager.

tenantID-Returns ID of the tenant the alert belongs to.
queryFromGeneratorURLgenerator_urlReturns the URL decoded query from GeneratorURL of an alert set by a Prometheus. Example: {{ queryFromGeneratorURL (index .Alerts 0).GeneratorURL }}
grafanaExploreURLgrafana_URL,datasource,from,to,exprReturns link to Grafana explore with range query based on the input parameters. Example: {{ grafanaExploreURL "" "xyz" "now-12h" "now" (queryFromGeneratorURL (index .Alerts 0).GeneratorURL) }}

Sharding and replication

The Alertmanager shards and replicates alerts by tenant. Sharding requires that the number of Alertmanager replicas is greater-than or equal-to the replication factor configured by the -alertmanager.sharding-ring.replication-factor flag.

Grafana Mimir Alertmanager replicas use a hash ring that is stored in the KV store to discover their peers. This means that any Mimir Alertmanager replica can respond to any API or UI request for any tenant. If the Mimir Alertmanager replica receiving the HTTP request doesn’t own the tenant to which the request belongs, the request is internally routed to the appropriate replica.

To configure the Alertmanagers’ hash ring, refer to configuring hash rings.

Note: When running with a single tenant, scaling the number of replicas to be greater than the replication factor offers no benefits as the Mimir Alertmanager shards by tenant and not individual alerts.


The Mimir Alertmanager stores the alerts state on local disk at the location configured using

Warning: When running the Mimir Alertmanager without replication, ensure persistence of the directory to avoid losing alert state.

The Mimir Alertmanager also periodically stores the alert state in the storage backend configured with -alertmanager-storage.backend. When an Alertmanager starts, it attempts to load the alerts state for a given tenant from other Alertmanager replicas. If the load from other Alertmanager replicas fails, the Alertmanager falls back to the state that is periodically stored in the storage backend.

In the event of a cluster outage, this fallback mechanism recovers the backup of the previous state. Because backups are taken periodically, this fallback mechanism does not guarantee that the lastest state is restored.

Ruler configuration

You must configure the ruler with the addresses of Alertmanagers via the -ruler.alertmanager-url flag.

Point the address to Alertmanager’s API. You can configure Alertmanager’s API prefix via the -http.alertmanager-http-prefix flag, which defaults to /alertmanager. For example, if Alertmanager is listening at http://mimir-alertmanager.namespace.svc.cluster.local and it is using the default API prefix, set -ruler.alertmanager-url to http://mimir-alertmanager.namespace.svc.cluster.local/alertmanager.