LOG ERROR :
lvl=error msg=“failed to connect to InfluxDB, retrying…” service=influxdb cluster=localhost err=“Get http://localhost:8086/ping: dial tcp [::1]:8086: connect: connection refused”
Config :
The hostname of this node.
Must be resolvable by any configured InfluxDB hosts.
hostname = “localhost”
Directory for storing a small amount of metadata about the server.
data_dir = “/var/lib/kapacitor”
Do not apply configuration overrides during startup.
Useful if the configuration overrides cause Kapacitor to fail startup.
This option is intended as a safe guard and should not be needed in practice.
skip-config-overrides = false
Default retention-policy, if a write is made to Kapacitor and
it does not have a retention policy associated with it,
then the retention policy will be set to this value
default-retention-policy = “”
[http]
HTTP API Server for Kapacitor
This server is always on,
it serves both as a write endpoint
and as the API endpoint for all other
Kapacitor calls.
bind-address = “:9092”
log-enabled = true
write-tracing = false
pprof-enabled = false
https-enabled = false
https-certificate = “/etc/ssl/kapacitor.pem”
Use a separate private key location.
https-private-key = “”
[config-override]
Enable/Disable the service for overridding configuration via the HTTP API.
enabled = true
[logging]
# Destination for logs
# Can be a path to a file or ‘STDOUT’, ‘STDERR’.
file = “/var/log/kapacitor/kapacitor.log”
# Logging level can be one of:
# DEBUG, INFO, ERROR
# HTTP logging can be disabled in the [http] config section.
level = “DEBUG”
[load]
Enable/Disable the service for loading tasks/templates/handlers
from a directory
enabled = true
Directory where task/template/handler files are set
dir = “/etc/kapacitor/load”
[replay]
Where to store replay files, aka recordings.
dir = “/var/lib/kapacitor/replay”
[task]
Where to store the tasks database
DEPRECATED: This option is not needed for new installations.
It is only used to determine the location of the task.db file
for migrating to the new storage
service.
dir = “/var/lib/kapacitor/tasks”
How often to snapshot running task state.
snapshot-interval = “60s”
[storage]
Where to store the Kapacitor boltdb database
boltdb = “/var/lib/kapacitor/kapacitor.db”
[deadman]
Configure a deadman’s switch
Globally configure deadman’s switches on all tasks.
NOTE: for this to be of use you must also globally configure at least one alerting method.
global = false
Threshold, if globally configured the alert will be triggered if the throughput in points/interval is <= threshold.
threshold = 0.0
Interval, if globally configured the frequency at which to check the throughput.
interval = “10s”
Id – the alert Id, NODE_NAME will be replaced with the name of the node being monitored.
id = “node ‘NODE_NAME’ in task ‘{{ .TaskName }}’”
The message of the alert. INTERVAL will be replaced by the interval.
message = “{{ .ID }} is {{ if eq .Level "OK" }}alive{{ else }}dead{{ end }}: {{ index .Fields "collected" | printf "%0.3f" }} points/INTERVAL.”
Multiple InfluxDB configurations can be defined.
Exactly one must be marked as the default.
Each one will be given a name and can be referenced in batch queries and InfluxDBOut nodes.
[[influxdb]]
Connect to an InfluxDB cluster
Kapacitor can subscribe, query and write to this cluster.
Using InfluxDB is not required and can be disabled.
enabled = true
default = true
name = “localhost”
urls = [“http://localhost:8086”]
username = “”
password = “”
timeout = 0
Absolute path to pem encoded CA file.
A CA can be provided without a key/cert pair
ssl-ca = “/etc/kapacitor/ca.pem”
Absolutes paths to pem encoded key and cert files.
ssl-cert = “/etc/kapacitor/cert.pem”
ssl-key = “/etc/kapacitor/key.pem”
Do not verify the TLS/SSL certificate.
This is insecure.
insecure-skip-verify = false
Maximum time to try and connect to InfluxDB during startup
startup-timeout = “5m”
Turn off all subscriptions
disable-subscriptions = false
Subscription mode is either “cluster” or “server”
subscription-mode = “cluster”
Which protocol to use for subscriptions
one of ‘udp’, ‘http’, or ‘https’.
subscription-protocol = “http”
Subscriptions resync time interval
Useful if you want to subscribe to new created databases
without restart Kapacitord
subscriptions-sync-interval = “1m0s”
Override the global hostname option for this InfluxDB cluster.
Useful if the InfluxDB cluster is in a separate network and
needs special config to connect back to this Kapacitor instance.
Defaults to hostname
if empty.
kapacitor-hostname = “”
Override the global http port option for this InfluxDB cluster.
Useful if the InfluxDB cluster is in a separate network and
needs special config to connect back to this Kapacitor instance.
Defaults to the port from [http] bind-address
if 0.
http-port = 0
Host part of a bind address for UDP listeners.
For example if a UDP listener is using port 1234
and udp-bind = "hostname_or_ip"
,
then the UDP port will be bound to hostname_or_ip:1234
The default empty value will bind to all addresses.
udp-bind = “”
Subscriptions use the UDP network protocl.
The following options of for the created UDP listeners for each subscription.
Number of packets to buffer when reading packets off the socket.
udp-buffer = 1000
The size in bytes of the OS read buffer for the UDP socket.
A value of 0 indicates use the OS default.
udp-read-buffer = 0
[influxdb.subscriptions]
# Set of databases and retention policies to subscribe to.
# If empty will subscribe to all, minus the list in
# influxdb.excluded-subscriptions
#
# Format
# db_name =
#
# Example:
# my_database = [ “default”, “longterm” ]
[influxdb.excluded-subscriptions]
# Set of databases and retention policies to exclude from the subscriptions.
# If influxdb.subscriptions is empty it will subscribe to all
# except databases listed here.
#
# Format
# db_name =
#
# Example:
# my_database = [ “default”, “longterm” ]
[kubernetes]
Enable/Disable the kubernetes service.
Needed by the k8sAutoscale TICKscript node.
enabled = false
There are several ways to connect to the kubernetes API servers:
Via the proxy, start the proxy via the kubectl proxy
command:
api-servers = [“http://localhost:8001”]
From within the cluster itself, in which case
kubernetes secrets and DNS services are used
to determine the needed configuration.
in-cluster = true
Direct connection, in which case you need to know
the URL of the API servers, the authentication token and
the path to the ca cert bundle.
These value can be found using the kubectl config view
command.
api-servers = [“http://192.168.99.100:8443”]
token = “…”
ca-path = “/path/to/kubernetes/ca.crt”
Kubernetes can also serve as a discoverer for scrape targets.
In that case the type of resources to discoverer must be specified.
Valid values are: “node”, “pod”, “service”, and “endpoint”.
resource = “pod”
[smtp]
Configure an SMTP email server
Will use TLS and authentication if possible
Only necessary for sending emails from alerts.
enabled = false
host = “localhost”
port = 25
username = “”
password = “”
From address for outgoing mail
from = “”
List of default To addresses.
to = [“oncall@example.com”]
Skip TLS certificate verify when connecting to SMTP server
no-verify = false
Close idle connections after timeout
idle-timeout = “30s”
If true the all alerts will be sent via Email
without explicitly marking them in the TICKscript.
global = false
Only applies if global is true.
Sets all alerts in state-changes-only mode,
meaning alerts will only be sent if the alert state changes.
state-changes-only = false
[snmptrap]
Configure an SNMP trap server
enabled = false
The host:port address of the SNMP trap server
addr = “localhost:162”
The community to use for traps
community = “kapacitor”
Number of retries when sending traps
retries = 1
[opsgenie]
# Configure OpsGenie with your API key
enabled = false
# Your OpsGenie API Key.
api-key = “”
# Default OpsGenie teams, can be overridden per alert.
# teams = [“team1”, “team2”]
# Default OpsGenie recipients, can be overridden per alert.
# recipients = [“recipient1”, “recipient2”]
# The OpsGenie API URL should not need to be changed.
url = “https://api.opsgenie.com/v1/json/alert”
# The OpsGenie Recovery URL, you can change this
# based on which behavior you want a recovery to
# trigger (Add Notes, Close Alert, etc.)
recovery_url = “https://api.opsgenie.com/v1/json/alert/note”
# If true then all alerts will be sent to OpsGenie
# without explicitly marking them in the TICKscript.
# The team and recipients can still be overridden.
global = false
[opsgenie2]
# Configure OpsGenie v2 with your API key
enabled = false
# Your OpsGenie API Key.
api-key = “”
# Default OpsGenie teams, can be overridden per alert.
# teams = [“team1”, “team2”]
# Default OpsGenie recipients, can be overridden per alert.
# recipients = [“recipient1”, “recipient2”]
# The OpsGenie API URL should not need to be changed.
url = “https://api.opsgenie.com/v2/alerts”
# The Recovery Action specifies which action to take when alerts recover.
# Valid values include:
# * notes – add a note to the alert
# * close – close the alert
#
recovery_action = “notes”
# If true then all alerts will be sent to OpsGenie
# without explicitly marking them in the TICKscript.
# The team and recipients can still be overridden.
global = false
[victorops]
Configure VictorOps with your API key and default routing key.
enabled = false
Your VictorOps API Key.
api-key = “”
Default VictorOps routing key, can be overridden per alert.
routing-key = “”
The VictorOps API URL should not need to be changed.
url = “https://alert.victorops.com/integrations/generic/20131114/alert”
If true the all alerts will be sent to VictorOps
without explicitly marking them in the TICKscript.
The routing key can still be overridden.
global = false
Use JSON for the “data” field
New installations will want to set this to true as it makes
the data that triggered the alert available within VictorOps.
The default is “false” for backwards compatibility reasons.
json-data = false
[pagerduty]
Configure PagerDuty.
enabled = false
Your PagerDuty Service Key.
service-key = “”
The PagerDuty API URL should not need to be changed.
url = “https://events.pagerduty.com/generic/2010-04-15/create_event.json”
If true the all alerts will be sent to PagerDuty
without explicitly marking them in the TICKscript.
global = false
[pagerduty2]
Configure PagerDuty API v2.
enabled = false
Your PagerDuty Routing Key.
routing-key = “”
The PagerDuty API v2 URL should not need to be changed.
url = “https://events.pagerduty.com/v2/enqueue”
If true the all alerts will be sent to PagerDuty
without explicitly marking them in the TICKscript.
global = false
[pushover]
Configure Pushover.
enabled = false
Your Pushover API token.
token = “”
Your Pushover USER_TOKEN.
user-key = “”
The URL for the Pushover API.
url = “https://api.pushover.net/1/messages.json”
##########################################
Configure Alert POST request Endpoints
As ENV variables:
KAPACITOR_HTTPPOST_0_ENDPOINT = “example”
KAPACITOR_HTTPPOST_0_URL = “http://example.com”
KAPACITOR_HTTPPOST_0_HEADERS_Example = “header”
[[httppost]]
endpoint = “example”
url = “http://example.com”
headers = { Example = “your-key” }
basic-auth = { username = “my-user”, password = “my-pass” }
# Provide an alert template for constructing a custom HTTP body.
# Alert templates are only used with post alert handlers as they consume alert data.
# The template uses template package - text/template - Go Packages and has access to the following fields:
# * .ID - The unique ID for this alert
# * .Message - The message of the alert
# * .Details - The details of the alert
# * .Time - The time the alert event occurred
# * .Duration - The duration of the alert event.
# * .Level - The level of the alert, i.e INFO, WARN, or CRITICAL.
# * .Data - The data that triggered the alert.
# Specify the template inline.
alert-template = “{{.Message}}:{{range .Data.Series}}{{.Tags}},{{range .Values}}{{.}}{{end}}{{end}}”
# Specify an absolute path to a template file.
alert-template-file = “/path/to/template/file”
# Provide a row template for constructing a custom HTTP body.
# Row templates are only used with httpPost pipeline nodes as they consume a row at a time.
# The template uses template package - text/template - Go Packages and has access to the following fields:
# * .Name - The measurement name of the data stream
# * .Tags - A map of tags on the data.
# * .Values - A list of values, each entry is a map containing a “time” key for the time of the point
# and keys for all other fields on the point.
# Specify the template inline.
row-template = “{{.Name}} host={{index .Tags "host"}}{{range .Values}} {{index . “time”}} {{index . “value”}}{{end}}”
# Specify an absolute path to a template file.
row-template-file = “/path/to/template/file”
Slack client configuration
Mutliple different clients may be configured by
repeating [[slack]] sections.
[[slack]]
Configure Slack.
enabled = false
identify one of the slack configurations as the default
default = true
workspace ID
This can be any string to identify this particular slack configuration
A logical choice is the name of the slack workspace, e.g. .slack.com
workspace = “”
The Slack webhook URL, can be obtained by adding
an Incoming Webhook integration.
Visit Slack
to add new webhook for Kapacitor.
url = “”
Default channel for messages
channel = “”
If true all the alerts will be sent to Slack
without explicitly marking them in the TICKscript.
global = false
Only applies if global is true.
Sets all alerts in state-changes-only mode,
meaning alerts will only be sent if the alert state changes.
state-changes-only = false
[telegram]
Configure Telegram.
enabled = false
The Telegram Bot URL should not need to be changed.
url = “https://api.telegram.org/bot”
Telegram Bot Token, can be obtained From @BotFather.
token = “”
Default recipient for messages, Contact @myidbot on Telegram to get an ID.
chat-id = “”
Send Markdown or HTML, if you want Telegram apps to show bold, italic, fixed-width text or inline URLs in your alert message.
#parse-mode = “Markdown”
Disable link previews for links in this message
disable-web-page-preview = false
Sends the message silently. iOS users will not receive a notification, Android users will receive a notification with no sound.
disable-notification = false
If true the all alerts will be sent to Telegram
without explicitly marking them in the TICKscript.
global = false
Only applies if global is true.
Sets all alerts in state-changes-only mode,
meaning alerts will only be sent if the alert state changes.
state-changes-only = false
[hipchat]
Configure HipChat.
enabled = false
The HipChat API URL. Replace subdomain with your
HipChat subdomain.
url = “https://subdomain.hipchat.com/v2/room”
Visit https://www.hipchat.com/docs/apiv2
for information on obtaining your room id and
authentication token.
Default room for messages
room = “”
Default authentication token
token = “”
If true then all alerts will be sent to HipChat
without explicitly marking them in the TICKscript.
global = false
Only applies if global is true.
Sets all alerts in state-changes-only mode,
meaning alerts will only be sent if the alert state changes.
state-changes-only = false
[[kafka]]
Configure Kafka
enabled = false
ID is a unique identifier for this kafka cluster.
id = “localhost”
Brokers is a list of host:port addresses of Kafka brokers.
brokers =
Timeout on network operations with the brokers.
If 0 a default of 10s will be used.
timeout = “10s”
BatchSize is the number of messages that are batched before being sent to Kafka
If 0 a default of 100 will be used.
batch-size = 100
BatchTimeout is the maximum amount of time to wait before flushing an incomplete batch.
If 0 a default of 1s will be used.
batch-timeout = “1s”
Use SSL enables ssl communication.
Must be true for the other ssl options to take effect.
use-ssl = false
Path to CA file
ssl-ca = “”
Path to host cert file
ssl-cert = “”
Path to cert key file
ssl-key = “”
Use SSL but skip chain & host verification
insecure-skip-verify = false
[alerta]
Configure Alerta.
enabled = false
The Alerta URL.
url = “”
Default authentication token.
token = “”
Default token prefix
If you are on older versions of alerta you may need to change this to “Key”
token-prefix = “Bearer”
Default environment.
environment = “”
Default origin.
origin = “kapacitor”
[sensu]
Configure Sensu.
enabled = false
The Sensu Client host:port address.
addr = “sensu-client:3030”
Default JIT source.
source = “Kapacitor”
[reporting]
Send usage statistics
every 12 hours to Enterprise.
enabled = true
url = “https://usage.influxdata.com”
[stats]
Emit internal statistics about Kapacitor.
To consume these stats create a stream task
that selects data from the configured database
and retention policy.
Example:
stream|from().database(‘_kapacitor’).retentionPolicy(‘autogen’)…
enabled = true
stats-interval = “10s”
database = “_kapacitor”
retention-policy= “autogen”
[udf]
Configuration for UDFs (User Defined Functions)
[udf.functions]
# Example go UDF.
# First compile example:
# go build -o avg_udf ./udf/agent/examples/moving_avg.go
#
# Use in TICKscript like:
# stream.goavg()
# .field(‘value’)
# .size(10)
# .as(‘m_average’)
#
# uncomment to enable
#[udf.functions.goavg]
# prog = “./avg_udf”
# args =
# timeout = “10s”
# Example python UDF.
# Use in TICKscript like:
# stream.pyavg()
# .field('value')
# .size(10)
# .as('m_average')
#
# uncomment to enable
#[udf.functions.pyavg]
# prog = "/usr/bin/python2"
# args = ["-u", "./udf/agent/examples/moving_avg.py"]
# timeout = "10s"
# [udf.functions.pyavg.env]
# PYTHONPATH = "./udf/agent/py"
# Example UDF over a socket
#[udf.functions.myCustomUDF]
# socket = "/path/to/socket"
# timeout = "10s"
[talk]
Configure Talk.
enabled = false
The Talk webhook URL.
url = “https://jianliao.com/v2/services/webhook/uuid”
The default authorName.
author_name = “Kapacitor”
MQTT client configuration.
Mutliple different clients may be configured by
repeating [[mqtt]] sections.
[[mqtt]]
enabled = false
Unique name for this broker configuration
name = “localhost”
Whether this broker configuration is the default
default = true
URL of the MQTT broker.
Possible protocols include:
tcp - Raw TCP network connection
ssl - TLS protected TCP network connection
ws - Websocket network connection
url = “tcp://localhost:1883”
TLS/SSL configuration
A CA can be provided without a key/cert pair
ssl-ca = “/etc/kapacitor/ca.pem”
Absolutes paths to pem encoded key and cert files.
ssl-cert = “/etc/kapacitor/cert.pem”
ssl-key = “/etc/kapacitor/key.pem”
Unique ID for this MQTT client.
If empty used the value of “name”
client-id = “”
Username
username = “”
Password
password = “”
[[swarm]]
Enable/Disable the Docker Swarm service.
Needed by the swarmAutoscale TICKscript node.
enabled = false
Unique ID for this Swarm cluster
NOTE: This is not the ID generated by Swarm rather a user defined
ID for this cluster since Kapacitor can communicate with multiple clusters.
id = “”
List of URLs for Docker Swarm servers.
servers = [“http://localhost:2376”]
TLS/SSL Configuration for connecting to secured Docker daemons
ssl-ca = “”
ssl-cert = “”
ssl-key = “”
insecure-skip-verify = false
##################################
Input Methods, same as InfluxDB
[collectd]
enabled = false
bind-address = “:25826”
database = “collectd”
retention-policy = “”
batch-size = 1000
batch-pending = 5
batch-timeout = “10s”
typesdb = “/usr/share/collectd/types.db”
[opentsdb]
enabled = false
bind-address = “:4242”
database = “opentsdb”
retention-policy = “”
consistency-level = “one”
tls-enabled = false
certificate = “/etc/ssl/influxdb.pem”
batch-size = 1000
batch-pending = 5
batch-timeout = “1s”
Service Discovery and metric scraping
[[scraper]]
enabled = false
name = “myscraper”
Specify the id of a discoverer service specified below
discoverer-id = “”
Specify the type of discoverer service being used.
discoverer-service = “”
db = “prometheus_raw”
rp = “autogen”
type = “prometheus”
scheme = “http”
metrics-path = “/metrics”
scrape-interval = “1m0s”
scrape-timeout = “10s”
username = “”
password = “”
bearer-token = “”
ssl-ca = “”
ssl-cert = “”
ssl-key = “”
ssl-server-name = “”
insecure-skip-verify = false
Supported discovery services
[[azure]]
enabled = false
id = “myazure”
port = 80
subscription-id = “”
tenant-id = “”
client-id = “”
client-secret = “”
refresh-interval = “5m0s”
[[consul]]
enabled = false
id = “myconsul”
address = “127.0.0.1:8500”
token = “”
datacenter = “”
tag-separator = “,”
scheme = “http”
username = “”
password = “”
ssl-ca = “”
ssl-cert = “”
ssl-key = “”
ssl-server-name = “”
insecure-skip-verify = false
[[dns]]
enabled = false
id = “mydns”
refresh-interval = “30s”
Type can be SRV, A, or AAAA
type = “SRV”
Port is the port to scrape for records returned by A or AAAA types
port = 80
[[ec2]]
enabled = false
id = “myec2”
region = “us-east-1”
access-key = “”
secret-key = “”
profile = “”
refresh-interval = “1m0s”
port = 80
[[file-discovery]]
enabled = false
id = “myfile”
refresh-interval = “5m0s”
files =
[[gce]]
enabled = false
id = “mygce”
project = “”
zone = “”
filter = “”
refresh-interval = “1m0s”
port = 80
tag-separator = “,”
[[marathon]]
enabled = false
id = “mymarathon”
timeout = “30s”
refresh-interval = “30s”
bearer-token = “”
ssl-ca = “”
ssl-cert = “”
ssl-key = “”
ssl-server-name = “”
insecure-skip-verify = false
[[nerve]]
enabled = false
id = “mynerve”
timeout = “10s”
[[serverset]]
enabled = false
id = “myserverset”
timeout = “10s”
[[static-discovery]]
enabled = false
id = “mystatic”
targets = [“localhost:9100”]
[static.labels]
region = “us-east-1”
[[triton]]
enabled = false
id = “mytriton”
account = “”
dns-suffix = “”
endpoint = “”
port = 9163
refresh-interval = “1m0s”
version = 1
ssl-ca = “”
ssl-cert = “”
ssl-key = “”
ssl-server-name = “”
insecure-skip-verify = false