Important: This documentation is about an older version. It's relevant only to the release noted, many of the features and functions have been updated or replaced. Please view the current version.
Running distributed tests
It has already been established that k6 can run large load tests from a single instance, but what about multiple instances running a single test?
Several reasons why you may wish to run a distributed test include:
- Your system under test (SUT) should be accessed from multiple IP addresses.
- A fully optimized node cannot produce the load required by your extremely large test.
- Kubernetes is already your preferred operations environment.
For scenarios such as these, we’ve created the k6-operator.
Introducing k6-operator
k6-operator is an implementation of the operator pattern in Kubernetes, defining custom resources in Kubernetes. The intent is to automate tasks that a human operator would normally do; tasks like provisioning new application components, changing configurations, or resolving run-time issues.
The k6-operator defines the custom TestRun
resource type and listens for changes to, or creation of, TestRun
objects.
Each TestRun
object references a k6 test script, configures the environment, and specifies the number of instances, as parallelism
, for a test run.
Once a change is detected, the operator will react by modifying the cluster state, spinning up k6 test jobs as needed.
Get started with k6-operator
Let’s walk through the process for getting started with the k6-operator. The only requirement being access to a Kubernetes cluster and having the appropriate access and tooling.
Note
This process can be performed on a local Kubernetes cluster running within Docker! Using kind or k3d are awesome options to experiment with the process.
- Introducing k6-operator
- Get started with k6-operator
- 1. Install the operator
- 2. Create a test script
- 3. Add test scripts
- 4. Create a custom resource
- 5. Run your test
- 6. When things go wrong
- See also
1. Install the operator
The first step to running distributed tests in Kubernetes is to install the operator if not already installed in the cluster. At this time, installation does require having the project source code downloaded onto your system. Installation commands must be run from the source directory.
Note
Besides privileged access to a Kubernetes cluster, installation will require that the system performing the installation has the following tools installed:
From your command-line, execute the following:
git clone https://github.com/grafana/k6-operator && cd k6-operator
Ensure that your kubectl
tool is set for the appropriate Kubernetes cluster.
Then, from the k6-operator
directory, you may now perform the installation:
make deploy
By default, the operator will be installed into a new namespace, k6-operator-system
.
You can verify the successful installation by listing available resources within the namespace:
kubectl get pod -n k6-operator-system
After a few moments, your resulting status should become Running
as shown below:
NAME READY STATUS RESTARTS AGE
k6-operator-controller-manager-7664957cf7-llw54 2/2 Running 0 160m
You are now ready to start create and execute test scripts!
2. Create a test script
Creating k6 test scripts for Kubernetes is no different from creating the script for the command-line. If you haven’t already created test cases for your system, then we suggest having a read through one of our guides for creating tests for websites and APIs/microservices:
In general, it is advised to start small and expand on your scripts over iterations.
So let’s start simple and create a test.js
with the following content:
import http from 'k6/http';
import { sleep } from 'k6';
export const options = {
vus: 10,
duration: '10s',
};
export default function () {
http.get('https://test.k6.io/');
sleep(1);
}
Note
While creating scripts, run them locally before publishing to your cluster. This can give you immediate feedback if you have errors in your script.
Let’s go ahead and verify our script is valid by performing a brief test:
k6 run test.js
We should see a successful execution and resulting output summary.
3. Add test scripts
In order for your test scripts to be run, they must be published into your cluster. Two main methods for this are using ConfigMap or PersistentVolume resources.
Add as a ConfigMap
Using a ConfigMap
is a quick and straightforward mechanism for adding your test scripts to Kubernetes.
The kubectl
tool provides a convenient method to create a new ConfigMap
from a local script.
Let’s create our ConfigMap as my-test
with the content of our test.js
script we created in the previous step:
kubectl create configmap my-test --from-file test.js
Caution
Limitations exist on how large your test script can be when deployed within a
ConfigMap
. Kubernetes imposes a size limit of 1,048,576 bytes (1 MiB) for the data, therefore if your test scripts exceed this limit, you’ll need to mount aPersistentVolume
.Check the motivations for when you should use a
ConfigMap
versus aPersistentVolume
.
You should see confirmation with configmap/my-test created
.
Add inside a PersistentVolume
Setting up a PersistentVolume
is beyond the scope of this guide, but enables access to a shared filesystem from your Kubernetes cluster via PersistentVolumeClaim
.
When using this option, organize your test scripts in the applicable filesystem just as you would locally. This mechanism is ideal when breaking up monolithic scripts into reusable modules.
Note
Organizing your test scripts was part of the discussion during episode #76 of k6 Office Hours.
Caution
When using aPersistentVolume
, the operator will expect all test scripts to be contained within a directory named/test/
.
To learn more about creating PersistentVolume
and PersistentVolumeClaim
resources, review the Kubernetes documentation.
4. Create a custom resource
During installation, the TestRun
Custom Resource definition was added to the Kubernetes API.
The data we provide in the custom resource TestRun
object should contain all the information necessary for the k6-operator to start a distributed load test.
Specifically, the main elements defined within the TestRun
object relate to the name and location of the test script to run, and the amount of parallelism to utilize.
Note
TheTestRun
custom resource provides many configuration options to control the initialization and execution of tests. For the full listing of possible options, please refer to the project source and README.
The following examples will show some common variations for the custom resource:
Script in a ConfigMap
When the test script to be executed is contained within a ConfigMap
resource, we specify the script details within the configMap
block of YAML.
The name
is the name of the ConfigMap and the file
is the key-value for the entry.
Let’s create the file run-k6-from-configmap.yaml
with the following content:
apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-from-configmap
spec:
parallelism: 4
script:
configMap:
name: my-test
file: test.js
Recall when the script was added as a ConfigMap for our configuration values.
We created the ConfigMap named my-test
.
The test script content was added to the map using the filename as the key-value, therefore the file
value is test.js
.
The amount of parallelism
is up to you; how many pods do you want to split the test amongst?
The operator will split the workload between the pods using execution segments.
Caution
It is important that theConfigMap
andCustomResource
are created in the same Namespace.
Script in a PersistentVolume
If the test script to be executed is contained within a PersistentVolume
, creation of a PersistentVolumeClaim will be required.
We won’t go into the details of PersistentVolumes and PersistentVolumeClaims, but to learn more, you should review the Kubernetes documentation.
Assume we’ve created a PersistentVolumeClaim
named my-volume-claim
against a PersistentVolume
containing the test script /test/test.js
, we can create the file run-k6-from-volume.yaml
with the following content:
apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-from-volume
spec:
parallelism: 4
script:
volumeClaim:
name: my-volume-claim
# File is relative to /test/ directory within volume
file: test.js
Caution
It is important that thePersistentVolumeClaim
andCustomResource
are created in the same Namespace.
Configure the environment
Not everything should be included directly in your scripts. Well written scripts will allow for variability to support multiple scenarios and to avoid hard-coding values that tend to change. These could be anything from passwords to target urls, in addition to system options.
We can pass this data as environment variables for use with each pod executing your script.
This can be defined explicitly within the TestRun
resource, or by referencing a ConfigMap
or Secret
.
apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-with-vars
spec:
parallelism: 4
script:
configMap:
name: my-test
file: test.js
runner:
env:
- name: MY_CUSTOM_VARIABLE
value: 'this is my variable value'
envFrom:
- configMapRef:
name: my-config-vars
- secretRef:
name: my-secrets-vars
Note
The above YAML introduces therunner
section. This section applies to each pod that will be running a portion of your test, based upon the desiredparallelism
.
Now, with the referenced resources, our test scripts can use environment variables as in the following:
export function setup() {
console.log(`Variable is set as: ${__ENV.MY_CUSTOM_VARIABLE}`);
}
Change command-line arguments
k6 options can be specified in many ways, one being the command-line. Specifying options via command-line can still be accomplished when using the operator as shown with the following example:
apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-with-args
spec:
parallelism: 4
script:
configMap:
name: my-test
file: test.js
arguments: --tag testid=run-k6-with-args --log-format json
With the above arguments, we’re adding a test-wide custom tag to metrics and changing the output format of logs to JSON.
Note
Be sure to visit the options reference for a listing of available options.
5. Run your test
Tests are executed by applying the custom resource TestRun
to a cluster where the operator is running.
The test configuration is applied as in the following:
kubectl apply -f /path/to/your/k6-resource.yaml
After completing a test run, you need to clean up the test jobs created. This is done by running the following command:
kubectl delete -f /path/to/your/k6-resource.yaml
Note
If you make use of real-time results output, e.g. Prometheus Remote Write, use the
cleanup
option to automatically remove resources upon test completion as with the following example:apiVersion: k6.io/v1alpha1 kind: TestRun metadata: name: run-k6-with-realtime spec: parallelism: 4 # Removes all resources upon completion cleanup: post script: configMap: name: my-test file: test.js arguments: -o experimental-prometheus-rw
6. When things go wrong
Sadly nothing works perfectly all the time, so knowing where you can go for help is important.
Be sure to search the k6-operator category in the community forum. k6 has a growing and helpful community of engineers working with k6-operator, so there’s a good chance your issue has already been discussed and overcome. It’s also in these forums where you’ll be able to get help from members of the k6 development team.
See also
Here are some additional resources to help on your learning journey: