Getting started
Introduction
Dkron nodes can work in two modes, agents or servers.
A Dkron agent is a cluster member that can handle job executions, run your scripts and return the resulting output to the server.
A Dkron server is also a cluster member that send job execution queries to agents or other servers, so servers can execute jobs too.
The main distinction is that servers order job executions, can be used to schedule jobs, handles data storage and participate on leader election.
Dkron clusters have a leader, the leader is responsible of starting job execution queries in the cluster.
Any Dkron agent or server acts as a cluster member and it's available to run scheduled jobs.
You can choose whether a job is run on a node or nodes by specifying tags and a count of target nodes having this tag do you want a job to run. This gives an unprecedented level of flexibility in runnig jobs across a cluster of any size and with any combination of machines you need.
All the execution responses will be gathered by the scheduler and stored in the database.
State storage
Dkron deployment is just a single binary, it stores the state in an internal BadgerDB instance and replicate all changes between all server nodes using the Raft protocol, it doesn't need any other storage system outside itself.
Installation
See the installation.
Configuration
See the configuration.
Usage
By default Dkron uses the following ports:
8946
for serf layer between agents8080
for HTTP for the API and Dashboard6868
for gRPC and raft layer comunication between agents.
Be sure you have opened this ports (or the ones that you configured) in your firewall or AWS security groups.
Starting a single node
Works out of the box, good for non HA installations.
- System service: If no changes are done to the default config files, dkron will start as a service in single mode.
- Command line: Running a single node with default config can be done by running:
dkron agent --server --bootstrap-expect=1
Check your server is working: curl localhost:8080/v1
Simple as that, now it is time to add a job:
This job will only run in just one server
node due to the node count in the tag. Refer to the target node spec for details.
curl localhost:8080/v1/jobs -XPOST -d '{
"name": "job1",
"schedule": "@every 10s",
"timezone": "Europe/Berlin",
"owner": "Platform Team",
"owner_email": "[email protected]",
"disabled": false,
"tags": {
"server": "true"
},
"metadata": {
"user": "12345"
}
"concurrency": "allow",
"executor": "shell",
"executor_config": {
"command": "date"
}
}'
For full Job params description refer to the Job model in the API guide
That's it!