Influxdb is not starting

#1

hii ,
i installed influxdb on ubuntu and enabled grafite but when i try ‘sudo service influxdb start’ i am getting like this.

#2

CAn you post the relevant part of the log file for this? /var/log/influxdb/influxdb.log

That should give more information about the error. There isn’t much in the screenshot to help diagnose so the log file would be better.

Usually fails due to a bad line in the config. Check you haven’t commented/uncommented the wrong line. I’ve not used Graphite so have never connected it to InfluxDB but the configs are at the bottom of this link

#4

gatling.conf file

gatling {

core {

#outputDirectoryBaseName = "" # The prefix for each simulation result folder (then suffixed by the report generation timestamp)

#runDescription = ""          # The description for this simulation run, displayed in each report

#encoding = "utf-8"           # Encoding to use throughout Gatling for file and string manipulation

#simulationClass = ""         # The FQCN of the simulation to run (when used in conjunction with noReports, the simulation for which assertions will be validated)

#mute = false                 # When set to true, don't ask for simulation name nor run description (currently only used by Gatling SBT plugin)

#elFileBodiesCacheMaxCapacity = 200        # Cache size for request body EL templates, set to 0 to disable

#rawFileBodiesCacheMaxCapacity = 200       # Cache size for request body Raw templates, set to 0 to disable

#rawFileBodiesInMemoryMaxSize = 1000       # Below this limit, raw file bodies will be cached in memory

extract {

  regex {

    #cacheMaxCapacity = 200 # Cache size for the compiled regexes, set to 0 to disable caching

  }

  xpath {

    #cacheMaxCapacity = 200 # Cache size for the compiled XPath queries,  set to 0 to disable caching

  }

  jsonPath {

    #cacheMaxCapacity = 200 # Cache size for the compiled jsonPath queries, set to 0 to disable caching

    #preferJackson = false  # When set to true, prefer Jackson over Boon for JSON-related operations

  }

  css {

    #cacheMaxCapacity = 200 # Cache size for the compiled CSS selectors queries,  set to 0 to disable caching

  }

}

directory {

  data = user-files/data               # Folder where user's data (e.g. files used by Feeders) is located

  bodies = user-files/bodies           # Folder where bodies are located

  simulations = user-files/simulations # Folder where the bundle's simulations are located

  reportsOnly = ""                     # If set, name of report folder to look for in order to generate its report

  binaries = ""                        # If set, name of the folder where compiles classes are located: Defaults to GATLING_HOME/target.

  results = results                    # Name of the folder where all reports folder are located

}

}

charting {

#noReports = false       # When set to true, don't generate HTML reports

#maxPlotPerSeries = 1000 # Number of points per graph in Gatling reports

#useGroupDurationMetric = false  # Switch group timings from cumulated response time to group duration.

indicators {

  #lowerBound = 800      # Lower bound for the requests' response time to track in the reports and the console summary

  #higherBound = 1200    # Higher bound for the requests' response time to track in the reports and the console summary

  #percentile1 = 50      # Value for the 1st percentile to track in the reports, the console summary and Graphite

  #percentile2 = 75      # Value for the 2nd percentile to track in the reports, the console summary and Graphite

  #percentile3 = 95      # Value for the 3rd percentile to track in the reports, the console summary and Graphite

  #percentile4 = 99      # Value for the 4th percentile to track in the reports, the console summary and Graphite

}

}

http {

#fetchedCssCacheMaxCapacity = 200          # Cache size for CSS parsed content, set to 0 to disable

#fetchedHtmlCacheMaxCapacity = 200         # Cache size for HTML parsed content, set to 0 to disable

#perUserCacheMaxCapacity = 200             # Per virtual user cache size, set to 0 to disable

#warmUpUrl = "[https://gatling.io](https://gatling.io/)"           # The URL to use to warm-up the HTTP stack (blank means disabled)

#enableGA = true                           # Very light Google Analytics, please support

ssl {

  keyStore {

    #type = ""      # Type of SSLContext's KeyManagers store

    #file = ""      # Location of SSLContext's KeyManagers store

    #password = ""  # Password for SSLContext's KeyManagers store

    #algorithm = "" # Algorithm used SSLContext's KeyManagers store

  }

  trustStore {

    #type = ""      # Type of SSLContext's TrustManagers store

    #file = ""      # Location of SSLContext's TrustManagers store

    #password = ""  # Password for SSLContext's TrustManagers store

    #algorithm = "" # Algorithm used by SSLContext's TrustManagers store

  }

}

ahc {

  keepAlive = true                                # Allow pooling HTTP connections (keep-alive header automatically added)

  connectTimeout = 10000                          # Timeout when establishing a connection

  handshakeTimeout = 10000                        # Timeout when performing TLS hashshake

  pooledConnectionIdleTimeout = 60000             # Timeout when a connection stays unused in the pool

  readTimeout = 60000                             # Timeout when a used connection stays idle

  maxRetry = 2                                    # Number of times that a request should be tried again

  requestTimeout = 60000                          # Timeout of the requests

  #acceptAnyCertificate = true                     # When set to true, doesn't validate SSL certificates

  #httpClientCodecMaxInitialLineLength = 4096      # Maximum length of the initial line of the response (e.g. "HTTP/1.0 200 OK")

  #httpClientCodecMaxHeaderSize = 8192             # Maximum size, in bytes, of each request's headers

  #httpClientCodecMaxChunkSize = 8192              # Maximum length of the content or each chunk

  #webSocketMaxFrameSize = 10240000                # Maximum frame payload size

  #sslEnabledProtocols = [TLSv1.2, TLSv1.1, TLSv1] # Array of enabled protocols for HTTPS, if empty use the JDK defaults

  #sslEnabledCipherSuites = []                     # Array of enabled cipher suites for HTTPS, if empty use the AHC defaults

  #sslSessionCacheSize = 0                         # SSLSession cache size, set to 0 to use JDK's default

  #sslSessionTimeout = 0                           # SSLSession timeout in seconds, set to 0 to use JDK's default (24h)

  #useOpenSsl = false                              # if OpenSSL should be used instead of JSSE (requires tcnative jar)

  #useNativeTransport = false                      # if native transport should be used instead of Java NIO (requires netty-transport-native-epoll, currently Linux only)

  #tcpNoDelay = true

  #soReuseAddress = false

  #soLinger = -1

  #soSndBuf = -1

  #soRcvBuf = -1

  #allocator = "pooled"                            # switch to unpooled for unpooled ByteBufAllocator

  #maxThreadLocalCharBufferSize = 200000           # Netty's default is 16k

}

dns {

  #queryTimeout = 5000                             # Timeout of each DNS query in millis

  #maxQueriesPerResolve = 6                        # Maximum allowed number of DNS queries for a given name resolution

}

}

jms {

#acknowledgedMessagesBufferSize = 5000             # size of the buffer used to tracked acknowledged messages and protect against duplicate receives

}

data {

writers = [console, file, graphite]      # The list of DataWriters to which Gatling write simulation data (currently supported : console, file, graphite, jdbc)

console {

  #light = false                # When set to true, displays a light version without detailed request stats

}

file {

  #bufferSize = 8192            # FileDataWriter's internal data buffer size, in bytes

}

leak {

  #noActivityTimeout = 30  # Period, in seconds, for which Gatling may have no activity before considering a leak may be happening

}

graphite {

  #light = false              # only send the all* stats

  host = "127.0.0.1"         # The host where the Carbon server is located

  port = "2003"                # The port to which the Carbon server listens to (2003 is default for plaintext, 2004 is default for pickle)

  protocol = "tcp"           # The protocol used to send data to Carbon (currently supported : "tcp", "udp")

  rootPathPrefix = "gatling" # The common prefix of all metrics sent to Graphite

  #bufferSize = 8192          # GraphiteDataWriter's internal data buffer size, in bytes

  #writeInterval = 1          # GraphiteDataWriter's write interval, in seconds

}

}

}

#5

Hii Philip

Thanks for the reply. now the issue is resolved. It was because i was restarting influxdb quickly after stopping it. However now I am facing a new issue that my ‘gatlingdb’ is not forming eventhough i enabled graphite in the configuration file. i will share you the influxdb configuration file here

Welcome to the InfluxDB configuration file.

The values in this file override the default values used by the system if

a config option is not specified. The commented out lines are the configuration

field and the default value used. Uncommenting a line and changing the value

will change the value used at runtime when the process is restarted.

Once every 24 hours InfluxDB will report usage data to usage.influxdata.com

The data includes a random ID, os, arch, version, the number of series and other

usage data. No data from user databases is ever transmitted.

Change this option to true to disable reporting.

reporting-disabled = false

Bind address to use for the RPC service for backup and restore.

bind-address = “127.0.0.1:8088”

[meta]

Controls the parameters for the Raft consensus group that stores metadata

about the InfluxDB cluster.

[meta]

Where the metadata/raft database is stored

dir = “/var/lib/influxdb/meta”

Automatically create a default retention policy when creating a database.

retention-autocreate = true

If log messages are printed for the meta service

logging-enabled = true

[data]

Controls where the actual shard data for InfluxDB lives and how it is

flushed from the WAL. “dir” may need to be changed to a suitable place

for your system, but the WAL settings are an advanced configuration. The

defaults should work for most systems.

[data]

The directory where the TSM storage engine stores TSM files.

dir = “/var/lib/influxdb/data”

The directory where the TSM storage engine stores WAL files.

wal-dir = “/var/lib/influxdb/wal”

The amount of time that a write will wait before fsyncing. A duration

greater than 0 can be used to batch up multiple fsync calls. This is useful for slower

disks or when WAL write contention is seen. A value of 0s fsyncs every write to the WAL.

Values in the range of 0-100ms are recommended for non-SSD disks.

wal-fsync-delay = “0s”

The type of shard index to use for new shards. The default is an in-memory index that is

recreated at startup. A value of “tsi1” will use a disk based index that supports higher

cardinality datasets.

index-version = “inmem”

Trace logging provides more verbose output around the tsm engine. Turning

this on can provide more useful output for debugging tsm engine issues.

trace-logging-enabled = false

Whether queries should be logged before execution. Very useful for troubleshooting, but will

log any sensitive data contained within a query.

query-log-enabled = true

Validates incoming writes to ensure keys only have valid unicode characters.

This setting will incur a small overhead because every key must be checked.

validate-keys = false

Settings for the TSM engine

CacheMaxMemorySize is the maximum size a shard’s cache can

reach before it starts rejecting writes.

Valid size suffixes are k, m, or g (case insensitive, 1024 = 1k).

Values without a size suffix are in bytes.

cache-max-memory-size = “1g”

CacheSnapshotMemorySize is the size at which the engine will

snapshot the cache and write it to a TSM file, freeing up memory

Valid size suffixes are k, m, or g (case insensitive, 1024 = 1k).

Values without a size suffix are in bytes.

cache-snapshot-memory-size = “25m”

CacheSnapshotWriteColdDuration is the length of time at

which the engine will snapshot the cache and write it to

a new TSM file if the shard hasn’t received writes or deletes

cache-snapshot-write-cold-duration = “10m”

CompactFullWriteColdDuration is the duration at which the engine

will compact all TSM files in a shard if it hasn’t received a

write or delete

compact-full-write-cold-duration = “4h”

The maximum number of concurrent full and level compactions that can run at one time. A

value of 0 results in 50% of runtime.GOMAXPROCS(0) used at runtime. Any number greater

than 0 limits compactions to that value. This setting does not apply

to cache snapshotting.

max-concurrent-compactions = 0

CompactThroughput is the rate limit in bytes per second that we

will allow TSM compactions to write to disk. Note that short bursts are allowed

to happen at a possibly larger value, set by CompactThroughputBurst

compact-throughput = “48m”

CompactThroughputBurst is the rate limit in bytes per second that we

will allow TSM compactions to write to disk.

compact-throughput-burst = “48m”

The threshold, in bytes, when an index write-ahead log file will compact

into an index file. Lower sizes will cause log files to be compacted more

quickly and result in lower heap usage at the expense of write throughput.

Higher sizes will be compacted less frequently, store more series in-memory,

and provide higher write throughput.

Valid size suffixes are k, m, or g (case insensitive, 1024 = 1k).

Values without a size suffix are in bytes.

max-index-log-file-size = “1m”

The maximum series allowed per database before writes are dropped. This limit can prevent

high cardinality issues at the database level. This limit can be disabled by setting it to

0.

max-series-per-database = 1000000

The maximum number of tag values per tag that are allowed before writes are dropped. This limit

can prevent high cardinality tag values from being written to a measurement. This limit can be

disabled by setting it to 0.

max-values-per-tag = 100000

If true, then the mmap advise value MADV_WILLNEED will be provided to the kernel with respect to

TSM files. This setting has been found to be problematic on some kernels, and defaults to off.

It might help users who have slow disks in some cases.

tsm-use-madv-willneed = false

[coordinator]

Controls the clustering service configuration.

[coordinator]

The default time a write request will wait until a “timeout” error is returned to the caller.

write-timeout = “10s”

The maximum number of concurrent queries allowed to be executing at one time. If a query is

executed and exceeds this limit, an error is returned to the caller. This limit can be disabled

by setting it to 0.

max-concurrent-queries = 0

The maximum time a query will is allowed to execute before being killed by the system. This limit

can help prevent run away queries. Setting the value to 0 disables the limit.

query-timeout = “0s”

The time threshold when a query will be logged as a slow query. This limit can be set to help

discover slow or resource intensive queries. Setting the value to 0 disables the slow query logging.

log-queries-after = “0s”

The maximum number of points a SELECT can process. A value of 0 will make

the maximum point count unlimited. This will only be checked every second so queries will not

be aborted immediately when hitting the limit.

max-select-point = 0

The maximum number of series a SELECT can run. A value of 0 will make the maximum series

count unlimited.

max-select-series = 0

The maxium number of group by time bucket a SELECT can create. A value of zero will max the maximum

number of buckets unlimited.

max-select-buckets = 0

[retention]

Controls the enforcement of retention policies for evicting old data.

[retention]

Determines whether retention policy enforcement enabled.

enabled = true

The interval of time when retention policy enforcement checks run.

check-interval = “30m”

[shard-precreation]

Controls the precreation of shards, so they are available before data arrives.

Only shards that, after creation, will have both a start- and end-time in the

future, will ever be created. Shards are never precreated that would be wholly

or partially in the past.

[shard-precreation]

Determines whether shard pre-creation service is enabled.

enabled = true

The interval of time when the check to pre-create new shards runs.

check-interval = “10m”

The default period ahead of the endtime of a shard group that its successor

group is created.

advance-period = “30m”

Controls the system self-monitoring, statistics and diagnostics.

The internal database for monitoring data is created automatically if

if it does not already exist. The target retention within this database

is called ‘monitor’ and is also created with a retention period of 7 days

and a replication factor of 1, if it does not exist. In all cases the

this retention policy is configured as the default for the database.

[monitor]

Whether to record statistics internally.

store-enabled = true

The destination database for recorded statistics

store-database = “_internal”

The interval at which to record statistics

store-interval = “10s”

[http]

Controls how the HTTP endpoints are configured. These are the primary

mechanism for getting data into and out of InfluxDB.

[http]

Determines whether HTTP endpoint is enabled.

enabled = true

The bind address used by the HTTP service.

bind-address = “127.0.0.1:8086”

Determines whether user authentication is enabled over HTTP/HTTPS.

auth-enabled = false

The default realm sent back when issuing a basic auth challenge.

realm = “InfluxDB”

Determines whether HTTP request logging is enabled.

log-enabled = true

Determines whether the HTTP write request logs should be suppressed when the log is enabled.

suppress-write-log = false

When HTTP request logging is enabled, this option specifies the path where

log entries should be written. If unspecified, the default is to write to stderr, which

intermingles HTTP logs with internal InfluxDB logging.

If influxd is unable to access the specified path, it will log an error and fall back to writing

the request log to stderr.

access-log-path = “”

Determines whether detailed write logging is enabled.

write-tracing = false

Determines whether the pprof endpoint is enabled. This endpoint is used for

troubleshooting and monitoring.

pprof-enabled = true

Enables a pprof endpoint that binds to localhost:6060 immediately on startup.

This is only needed to debug startup issues.

debug-pprof-enabled = false

Determines whether HTTPS is enabled.

https-enabled = false

The SSL certificate to use when HTTPS is enabled.

https-certificate = “/etc/ssl/influxdb.pem”

Use a separate private key location.

https-private-key = “”

The JWT auth shared secret to validate requests using JSON web tokens.

shared-secret = “”

The default chunk size for result sets that should be chunked.

max-row-limit = 0

The maximum number of HTTP connections that may be open at once. New connections that

would exceed this limit are dropped. Setting this value to 0 disables the limit.

max-connection-limit = 0

Enable http service over unix domain socket

unix-socket-enabled = false

The path of the unix domain socket.

bind-socket = “/var/run/influxdb.sock”

The maximum size of a client request body, in bytes. Setting this value to 0 disables the limit.

max-body-size = 25000000

The maximum number of writes processed concurrently.

Setting this to 0 disables the limit.

max-concurrent-write-limit = 0

The maximum number of writes queued for processing.

Setting this to 0 disables the limit.

max-enqueued-write-limit = 0

The maximum duration for a write to wait in the queue to be processed.

Setting this to 0 or setting max-concurrent-write-limit to 0 disables the limit.

enqueued-write-timeout = 0

[ifql]

Configures the ifql RPC API.

[ifql]

Determines whether the RPC service is enabled.

enabled = true

Determines whether additional logging is enabled.

#log-enabled = true

The bind address used by the ifql RPC service.

#bind-address = “:8082”

[logging]

Controls how the logger emits logs to the output.

[logging]

Determines which log encoder to use for logs. Available options

are auto, logfmt, and json. auto will use a more a more user-friendly

output format if the output terminal is a TTY, but the format is not as

easily machine-readable. When the output is a non-TTY, auto will use

logfmt.

format = “auto”

Determines which level of logs will be emitted. The available levels

are error, warn, info, and debug. Logs that are equal to or above the

specified level will be emitted.

level = “info”

Suppresses the logo output that is printed when the program is started.

The logo is always suppressed if STDOUT is not a TTY.

suppress-logo = false

[subscriber]

Controls the subscriptions, which can be used to fork a copy of all data

received by the InfluxDB host.

[subscriber]

Determines whether the subscriber service is enabled.

enabled = true

The default timeout for HTTP writes to subscribers.

http-timeout = “30s”

Allows insecure HTTPS connections to subscribers. This is useful when testing with self-

signed certificates.

insecure-skip-verify = false

The path to the PEM encoded CA certs file. If the empty string, the default system certs will be used

ca-certs = “”

The number of writer goroutines processing the write channel.

write-concurrency = 40

The number of in-flight writes buffered in the write channel.

write-buffer-size = 1000

[[graphite]]

Controls one or many listeners for Graphite data.

[[graphite]]

Determines whether the graphite endpoint is enabled.

enabled = true
database = “gatlingdb”
retention-policy = “”
bind-address = “127.0.0.1:2003”
protocol = “tcp”

consistency-level = “one”

templates = [
“gatling.... measurement.simulation.request.status.field”,
“gatling..users..* measurement.simulation.measurement.request.field”
]

These next lines control how batching works. You should have this enabled

otherwise you could get dropped metrics or poor performance. Batching

will buffer points in memory if you have many coming in.

Flush if this many points get buffered

batch-size = 5000

number of batches that may be pending in memory

batch-pending = 10

Flush at least this often even if we haven’t hit buffer limit

batch-timeout = “1s”

UDP Read buffer size, 0 means OS default. UDP listener will fail if set above OS max.

udp-read-buffer = 0

This string joins multiple matching ‘measurement’ values providing more control over the final measurement name.

separator = “.”

Default tags that will be added to all metrics. These can be overridden at the template level

or by tags extracted from metric

tags = [“region=us-east”, “zone=1c”]

Each template line requires a template pattern. It can have an optional

filter before the template and separated by spaces. It can also have optional extra

tags following the template. Multiple tags should be separated by commas and no spaces

similar to the line protocol format. There can be only one default template.

templates = [

“*.app env.service.resource.measurement”,

# Default template

“server.*”,

]

[collectd]

Controls one or many listeners for collectd data.

[[collectd]]

enabled = false

bind-address = “:25826”

database = “collectd”

retention-policy = “”

The collectd service supports either scanning a directory for multiple types

db files, or specifying a single db file.

typesdb = “/usr/local/share/collectd”

security-level = “none”

auth-file = “/etc/collectd/auth_file”

These next lines control how batching works. You should have this enabled

otherwise you could get dropped metrics or poor performance. Batching

will buffer points in memory if you have many coming in.

Flush if this many points get buffered

batch-size = 5000

Number of batches that may be pending in memory

batch-pending = 10

Flush at least this often even if we haven’t hit buffer limit

batch-timeout = “10s”

UDP Read buffer size, 0 means OS default. UDP listener will fail if set above OS max.

read-buffer = 0

Multi-value plugins can be handled two ways.

“split” will parse and store the multi-value plugin data into separate measurements

“join” will parse and store the multi-value plugin as a single multi-value measurement.

“split” is the default behavior for backward compatability with previous versions of influxdb.

parse-multivalue-plugin = “split”

[opentsdb]

Controls one or many listeners for OpenTSDB data.

[[opentsdb]]

enabled = false

bind-address = “:4242”

database = “opentsdb”

retention-policy = “”

consistency-level = “one”

tls-enabled = false

certificate= “/etc/ssl/influxdb.pem”

Log an error for every malformed point.

log-point-errors = true

These next lines control how batching works. You should have this enabled

otherwise you could get dropped metrics or poor performance. Only points

metrics received over the telnet protocol undergo batching.

Flush if this many points get buffered

batch-size = 1000

Number of batches that may be pending in memory

batch-pending = 5

Flush at least this often even if we haven’t hit buffer limit

batch-timeout = “1s”

[[udp]]

Controls the listeners for InfluxDB line protocol data via UDP.

[[udp]]

enabled = false

bind-address = “:8089”

database = “udp”

retention-policy = “”

InfluxDB precision for timestamps on received points ("" or “n”, “u”, “ms”, “s”, “m”, “h”)

precision = “”

These next lines control how batching works. You should have this enabled

otherwise you could get dropped metrics or poor performance. Batching

will buffer points in memory if you have many coming in.

Flush if this many points get buffered

batch-size = 5000

Number of batches that may be pending in memory

batch-pending = 10

Will flush at least this often even if we haven’t hit buffer limit

batch-timeout = “1s”

UDP Read buffer size, 0 means OS default. UDP listener will fail if set above OS max.

read-buffer = 0

[continuous_queries]

Controls how continuous queries are run within InfluxDB.

[continuous_queries]

Determines whether the continuous query service is enabled.

enabled = true

Controls whether queries are logged when executed by the CQ service.

log-enabled = true

Controls whether queries are logged to the self-monitoring data store.

query-stats-enabled = false

interval for how often continuous queries will be checked if they need to run

run-interval = “1s”

[tls]

Global configuration settings for TLS in InfluxDB.

[tls]

Determines the available set of cipher suites. See

for a list of available ciphers, which depends on the version of Go (use the query

SHOW DIAGNOSTICS to see the version of Go used to build InfluxDB). If not specified, uses

the default settings from Go’s crypto/tls package.

ciphers = [

“TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305”,

“TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256”,

]

Minimum version of the tls protocol that will be negotiated. If not specified, uses the

default settings from Go’s crypto/tls package.

min-version = “tls1.2”

Maximum version of the tls protocol that will be negotiated. If not specified, uses the

default settings from Go’s crypto/tls package.

max-version = “tls1.2”

#6

This seems to be a duplicate of a new topic you started on this subject so I’m closing this one.

closed #7