This is documentation for the next version of Loki. For the latest stable release, go to the latest version.
Note: Table manager is only needed if you are using a multi-store backend. If you are using either TSDB,(recommended) or BoltDB you do not need the Table Manager.
Grafana Loki supports storing indexes and chunks in table-based data storages. When such a storage type is used, multiple tables are created over the time: each table - also called periodic table - contains the data for a specific time range.
This design brings two main benefits:
- Schema config changes: each table is bounded to a schema config and version, so that changes can be introduced over the time and multiple schema configs can coexist
- Retention: the retention is implemented deleting an entire table, which allows to have fast delete operations
The Table Manager is a Loki component which takes care of creating a periodic table before its time period begins, and deleting it once its data time range exceeds the retention period.
The Table Manager supports the following backends:
- Index store
- Chunk store
- Filesystem (primarily used for local environments)
Loki does support the following backends for both index and chunk storage, but they are deprecated and will be removed in a future release:
The object storages - like Amazon S3 and Google Cloud Storage - supported by Loki to store chunks, are not managed by the Table Manager, and a custom bucket policy should be set to delete old data.
For detailed information on configuring the Table Manager, refer to the
section in the Loki configuration document.
Tables and schema config
A periodic table stores the index or chunk data relative to a specific period
of time. The duration of the time range of the data stored in a single table and
its storage type is configured in the
schema_config can contain
one or more
configs. Each config, defines the storage used between the day
from (in the format
yyyy-mm-dd) and the next config, or “now”
in the case of the last schema config entry.
This allows to have multiple non-overlapping schema configs over the time, in order to perform schema version upgrades or change storage settings (including changing the storage type).
The write path hits the table where the log entry timestamp falls into (usually the last table, except short periods close to the end of a table and the beginning of the next one), while the read path hits the tables containing data for the query time range.
Schema config example
For example, the following
schema_config defines two configurations: the first
one using the schema
v10 and the current one using the
The first config stores data between
then a new config has been added - to upgrade the schema version to
storing data using the
v11 schema from
For each config, multiple tables are created, each one storing data for
period time (168 hours = 7 days).
schema_config: configs: - from: 2019-01-01 store: dynamo schema: v10 index: prefix: loki_ period: 168h - from: 2019-04-15 store: dynamo schema: v11 index: prefix: loki_ period: 168h
The Table Manager creates new tables slightly ahead of their start period, in order to make sure that the new table is ready once the current table end period is reached.
creation_grace_period property - in the
configuration block - defines how long before a table should be created.
The retention - managed by the Table Manager - is disabled by default, due to
its destructive nature. You can enable the data retention explicitly enabling
it in the configuration and setting a
retention_period greater than zero:
table_manager: retention_deletes_enabled: true retention_period: 336h
The Table Manager implements the retention deleting the entire tables whose
data exceeded the
retention_period. This design allows to have fast delete
operations, at the cost of having a retention granularity controlled by the
Given each table contains data for
period of time and that the entire table
is deleted, the Table Manager keeps the last tables alive using this formula:
number_of_tables_to_keep = floor(retention_period / table_period) + 1
It’s important to note that - due to the internal implementation - the table
retention_period must be multiples of
24h in order to get
the expected behavior.
For detailed information on configuring the retention, refer to the Loki Storage Retention documentation.
Active / inactive tables
A table can be active or inactive.
A table is considered active if the current time is within the range:
- Table start period -
- Table end period + max chunk age (hardcoded to
Currently, the difference between an active and inactive table only applies to the DynamoDB storage settings: capacity mode (on-demand or provisioned), read/write capacity units and autoscaling.
|DynamoDB||Active table||Inactive table|
|Read capacity unit|
|Write capacity unit|
|Autoscaling||Enabled (if configured)||Always disabled|
When configuring DynamoDB with the Table Manager, the default on-demand provisioning capacity units for reads are set to 300 and writes are set to 3000. The defaults can be overwritten:
table_manager: index_tables_provisioning: provisioned_write_throughput: 10 provisioned_read_throughput: 10 chunk_tables_provisioning: provisioned_write_throughput: 10 provisioned_read_throughput: 10
If Table Manager is not automatically managing DynamoDB, old data cannot easily
be erased and the index will grow indefinitely. Manual configurations should
ensure that the primary index key is set to
h (string) and the sort key is set
r (binary). The “period” attribute in the configuration YAML should be set
Table Manager deployment mode
The Table Manager can be executed in two ways:
- Implicitly executed when Loki runs in monolithic mode (single process)
- Explicitly executed when Loki runs in microservices mode
When Loki runs in monolithic mode, the Table Manager is also started as component of the entire stack.
When Loki runs in microservices mode,
the Table Manager should be started as separate service named
You can check out a production grade deployment example at