Caution
Grafana Alloy is the new name for our distribution of the OTel collector. Grafana Agent has been deprecated and is in Long-Term Support (LTS) through October 31, 2025. Grafana Agent will reach an End-of-Life (EOL) on November 1, 2025. Read more about why we recommend migrating to Grafana Alloy.
Important: This documentation is about an older version. It's relevant only to the release noted, many of the features and functions have been updated or replaced. Please view the current version.
Modules
A Module is a unit of Grafana Agent Flow configuration, which combines all the other concepts, containing a mix of configuration blocks, instantiated components, and custom component definitions.
The module passed as an argument to the run
command is called the main configuration.
Modules can be imported to enable the reuse of custom components defined by that module.
Importing modules
A module can be imported, allowing the custom components defined by that module to be used by other modules, called the importing module.
Modules can be imported from multiple locations using one of the import
configuration blocks:
- import.file: Imports a module from a file or a directory on disk.
- import.git: Imports a module from a file located in a Git repository.
- import.http: Imports a module from the response of an HTTP request.
- import.string: Imports a module from a string.
Warning
You can’t import a module that contains top-level blocks other thandeclare
orimport
.
Modules are imported into a namespace where the top-level custom components of the imported module are exposed to the importing module.
The label of the import block specifies the namespace of an import.
For example, if a configuration contains a block called import.file "my_module"
, then custom components defined by that module are exposed as my_module.CUSTOM_COMPONENT_NAME
. Imported namespaces must be unique across a given importing module.
If an import namespace matches the name of a built-in component namespace, such as prometheus
, the built-in namespace is hidden from the importing module, and only components defined in the imported module may be used.
Example
This example module defines a component to filter out debug-level and info-level log lines:
declare "log_filter" {
// argument.write_to is a required argument that specifies where filtered
// log lines are sent.
//
// The value of the argument is retrieved in this file with
// argument.write_to.value.
argument "write_to" {
optional = false
}
// loki.process.filter is our component which executes the filtering,
// passing filtered logs to argument.write_to.value.
loki.process "filter" {
// Drop all debug- and info-level logs.
stage.match {
selector = `{job!=""} |~ "level=(debug|info)"`
action = "drop"
}
// Send processed logs to our argument.
forward_to = argument.write_to.value
}
// export.filter_input exports a value to the module consumer.
export "filter_input" {
// Expose the receiver of loki.process so the module importer can send
// logs to our loki.process component.
value = loki.process.filter.receiver
}
}
You can save this module to a file called helpers.river
and import it:
// Import our helpers.river module, exposing its custom components as
// helpers.COMPONENT_NAME.
import.file "helpers" {
filename = "helpers.river"
}
loki.source.file "self" {
targets = LOG_TARGETS
// Forward collected logs to the input of our filter.
forward_to = [helpers.log_filter.default.filter_input]
}
helpers.log_filter "default" {
// Configure the filter to forward filtered logs to loki.write below.
write_to = [loki.write.default.receiver]
}
loki.write "default" {
endpoint {
url = LOKI_URL
}
}
Classic modules (deprecated)
Caution
Modules were redesigned in v0.40 to simplify concepts. This section outlines the design of the original modules prior to v0.40. Classic modules are scheduled to be removed in the release after v0.40.
You use Modules to create Grafana Agent Flow configurations that you can load as a component. Modules are a great way to parameterize a configuration to create reusable pipelines.
Modules are Grafana Agent Flow configurations which have:
- Arguments: Settings that configure a module.
- Exports: Named values that a module exposes to the consumer of the module.
- Components: Grafana Agent Flow components to run when the module is running.
You use a Module loader to load Modules into Grafana Agent Flow.
Refer to argument block and export block to learn how to define arguments and exports for a module.
Module loaders
A Module loader is a Grafana Agent Flow component that retrieves a module and runs the defined components.
Module loader components are responsible for the following functions:
- Retrieving the module source.
- Creating a Component controller for the module.
- Passing arguments to the loaded module.
- Exposing exports from the loaded module.
Module loaders are typically called module.LOADER_NAME
.
Note
Some module loaders may not support running modules with arguments or exports.
Refer to Components for more information about the module loader components.
Module sources
Modules are flexible, and you can retrieve their configuration anywhere, such as:
- The local filesystem.
- An S3 bucket.
- An HTTP endpoint.
Each module loader component supports different ways of retrieving module.sources
.
The most generic module loader component, module.string
, can load modules from the export of another Grafana Agent Flow component.
local.file "my_module" {
filename = "PATH_TO_MODULE"
}
module.string "my_module" {
content = local.file.my_module.content
arguments {
MODULE_ARGUMENT_NAME_1 = MODULE_ARGUMENT_VALUE_1
MODULE_ARGUMENT_NAME_2 = MODULE_ARGUMENT_VALUE_2
// ...
}
}
Example module
This example module manages a pipeline that filters out debug-level and info-level log lines.
// argument.write_to is a required argument that specifies where filtered
// log lines are sent.
//
// The value of the argument is retrieved in this file with
// argument.write_to.value.
argument "write_to" {
optional = false
}
// loki.process.filter is our component which executes the filtering, passing
// filtered logs to argument.write_to.value.
loki.process "filter" {
// Drop all debug- and info-level logs.
stage.match {
selector = "{job!=\"\"} |~ \"level=(debug|info)\""
action = "drop"
}
// Send processed logs to our argument.
forward_to = argument.write_to.value
}
// export.filter_input exports a value to the module consumer.
export "filter_input" {
// Expose the receiver of loki.process so the module consumer can send
// logs to our loki.process component.
value = loki.process.filter.receiver
}
You can save the module to a file and then use it as a processing step before writing logs to Loki.
loki.source.file "self" {
targets = LOG_TARGETS
// Forward collected logs to the input of our filter.
forward_to = [module.file.log_filter.exports.filter_input]
}
module.file "log_filter" {
filename = "/path/to/modules/log_filter.river"
arguments {
// Configure the filter to forward filtered logs to loki.write below.
write_to = [loki.write.default.receiver],
}
}
loki.write "default" {
endpoint {
url = "LOKI_URL"
}
}