mirror of
https://github.com/nikdoof/vsphere-influxdb-go.git
synced 2025-12-19 21:49:21 +00:00
add vendoring with go dep
This commit is contained in:
1
vendor/github.com/influxdata/influxdb/etc/burn-in/.rvmrc
generated
vendored
Normal file
1
vendor/github.com/influxdata/influxdb/etc/burn-in/.rvmrc
generated
vendored
Normal file
@@ -0,0 +1 @@
|
||||
rvm use ruby-2.1.0@burn-in --create
|
||||
4
vendor/github.com/influxdata/influxdb/etc/burn-in/Gemfile
generated
vendored
Normal file
4
vendor/github.com/influxdata/influxdb/etc/burn-in/Gemfile
generated
vendored
Normal file
@@ -0,0 +1,4 @@
|
||||
source 'https://rubygems.org'
|
||||
|
||||
gem "colorize"
|
||||
gem "influxdb"
|
||||
14
vendor/github.com/influxdata/influxdb/etc/burn-in/Gemfile.lock
generated
vendored
Normal file
14
vendor/github.com/influxdata/influxdb/etc/burn-in/Gemfile.lock
generated
vendored
Normal file
@@ -0,0 +1,14 @@
|
||||
GEM
|
||||
remote: https://rubygems.org/
|
||||
specs:
|
||||
colorize (0.6.0)
|
||||
influxdb (0.0.16)
|
||||
json
|
||||
json (1.8.1)
|
||||
|
||||
PLATFORMS
|
||||
ruby
|
||||
|
||||
DEPENDENCIES
|
||||
colorize
|
||||
influxdb
|
||||
79
vendor/github.com/influxdata/influxdb/etc/burn-in/burn-in.rb
generated
vendored
Normal file
79
vendor/github.com/influxdata/influxdb/etc/burn-in/burn-in.rb
generated
vendored
Normal file
@@ -0,0 +1,79 @@
|
||||
require "influxdb"
|
||||
require "colorize"
|
||||
require "benchmark"
|
||||
|
||||
require_relative "log"
|
||||
require_relative "random_gaussian"
|
||||
|
||||
BATCH_SIZE = 10_000
|
||||
|
||||
Log.info "Starting burn-in suite"
|
||||
master = InfluxDB::Client.new
|
||||
master.delete_database("burn-in") rescue nil
|
||||
master.create_database("burn-in")
|
||||
master.create_database_user("burn-in", "user", "pass")
|
||||
|
||||
master.database = "burn-in"
|
||||
# master.query "select * from test1 into test2;"
|
||||
# master.query "select count(value) from test1 group by time(1m) into test2;"
|
||||
|
||||
influxdb = InfluxDB::Client.new "burn-in", username: "user", password: "pass"
|
||||
|
||||
Log.success "Connected to server #{influxdb.host}:#{influxdb.port}"
|
||||
|
||||
Log.log "Creating RandomGaussian(500, 25)"
|
||||
gaussian = RandomGaussian.new(500, 25)
|
||||
point_count = 0
|
||||
|
||||
while true
|
||||
Log.log "Generating 10,000 points.."
|
||||
points = []
|
||||
BATCH_SIZE.times do |n|
|
||||
points << {value: gaussian.rand.to_i.abs}
|
||||
end
|
||||
point_count += points.length
|
||||
|
||||
Log.info "Sending points to server.."
|
||||
begin
|
||||
st = Time.now
|
||||
foo = influxdb.write_point("test1", points)
|
||||
et = Time.now
|
||||
Log.log foo.inspect
|
||||
Log.log "#{et-st} seconds elapsed"
|
||||
Log.success "Write successful."
|
||||
rescue => e
|
||||
Log.failure "Write failed:"
|
||||
Log.log e
|
||||
end
|
||||
sleep 0.5
|
||||
|
||||
Log.info "Checking regular points"
|
||||
st = Time.now
|
||||
response = influxdb.query("select count(value) from test1;")
|
||||
et = Time.now
|
||||
|
||||
Log.log "#{et-st} seconds elapsed"
|
||||
|
||||
response_count = response["test1"].first["count"]
|
||||
if point_count == response_count
|
||||
Log.success "Point counts match: #{point_count} == #{response_count}"
|
||||
else
|
||||
Log.failure "Point counts don't match: #{point_count} != #{response_count}"
|
||||
end
|
||||
|
||||
# Log.info "Checking continuous query points for test2"
|
||||
# st = Time.now
|
||||
# response = influxdb.query("select count(value) from test2;")
|
||||
# et = Time.now
|
||||
|
||||
# Log.log "#{et-st} seconds elapsed"
|
||||
|
||||
# response_count = response["test2"].first["count"]
|
||||
# if point_count == response_count
|
||||
# Log.success "Point counts match: #{point_count} == #{response_count}"
|
||||
# else
|
||||
# Log.failure "Point counts don't match: #{point_count} != #{response_count}"
|
||||
# end
|
||||
end
|
||||
|
||||
|
||||
23
vendor/github.com/influxdata/influxdb/etc/burn-in/log.rb
generated
vendored
Normal file
23
vendor/github.com/influxdata/influxdb/etc/burn-in/log.rb
generated
vendored
Normal file
@@ -0,0 +1,23 @@
|
||||
module Log
|
||||
def self.info(msg)
|
||||
print Time.now.strftime("%r") + " | "
|
||||
puts msg.to_s.colorize(:yellow)
|
||||
end
|
||||
|
||||
def self.success(msg)
|
||||
print Time.now.strftime("%r") + " | "
|
||||
puts msg.to_s.colorize(:green)
|
||||
end
|
||||
|
||||
def self.failure(msg)
|
||||
print Time.now.strftime("%r") + " | "
|
||||
puts msg.to_s.colorize(:red)
|
||||
end
|
||||
|
||||
def self.log(msg)
|
||||
print Time.now.strftime("%r") + " | "
|
||||
puts msg.to_s
|
||||
end
|
||||
end
|
||||
|
||||
|
||||
31
vendor/github.com/influxdata/influxdb/etc/burn-in/random_gaussian.rb
generated
vendored
Normal file
31
vendor/github.com/influxdata/influxdb/etc/burn-in/random_gaussian.rb
generated
vendored
Normal file
@@ -0,0 +1,31 @@
|
||||
class RandomGaussian
|
||||
def initialize(mean, stddev, rand_helper = lambda { Kernel.rand })
|
||||
@rand_helper = rand_helper
|
||||
@mean = mean
|
||||
@stddev = stddev
|
||||
@valid = false
|
||||
@next = 0
|
||||
end
|
||||
|
||||
def rand
|
||||
if @valid then
|
||||
@valid = false
|
||||
return @next
|
||||
else
|
||||
@valid = true
|
||||
x, y = self.class.gaussian(@mean, @stddev, @rand_helper)
|
||||
@next = y
|
||||
return x
|
||||
end
|
||||
end
|
||||
|
||||
private
|
||||
def self.gaussian(mean, stddev, rand)
|
||||
theta = 2 * Math::PI * rand.call
|
||||
rho = Math.sqrt(-2 * Math.log(1 - rand.call))
|
||||
scale = stddev * rho
|
||||
x = mean + scale * Math.cos(theta)
|
||||
y = mean + scale * Math.sin(theta)
|
||||
return x, y
|
||||
end
|
||||
end
|
||||
29
vendor/github.com/influxdata/influxdb/etc/burn-in/random_points.rb
generated
vendored
Normal file
29
vendor/github.com/influxdata/influxdb/etc/burn-in/random_points.rb
generated
vendored
Normal file
@@ -0,0 +1,29 @@
|
||||
require "influxdb"
|
||||
|
||||
ONE_WEEK_IN_SECONDS = 7*24*60*60
|
||||
NUM_POINTS = 10_000
|
||||
BATCHES = 100
|
||||
|
||||
master = InfluxDB::Client.new
|
||||
master.delete_database("ctx") rescue nil
|
||||
master.create_database("ctx")
|
||||
|
||||
influxdb = InfluxDB::Client.new "ctx"
|
||||
influxdb.time_precision = "s"
|
||||
|
||||
names = ["foo", "bar", "baz", "quu", "qux"]
|
||||
|
||||
st = Time.now
|
||||
BATCHES.times do |m|
|
||||
points = []
|
||||
|
||||
puts "Writing #{NUM_POINTS} points, time ##{m}.."
|
||||
NUM_POINTS.times do |n|
|
||||
timestamp = Time.now.to_i - rand(ONE_WEEK_IN_SECONDS)
|
||||
points << {value: names.sample, time: timestamp}
|
||||
end
|
||||
|
||||
influxdb.write_point("ct1", points)
|
||||
end
|
||||
puts st
|
||||
puts Time.now
|
||||
431
vendor/github.com/influxdata/influxdb/etc/config.sample.toml
generated
vendored
Normal file
431
vendor/github.com/influxdata/influxdb/etc/config.sample.toml
generated
vendored
Normal file
@@ -0,0 +1,431 @@
|
||||
### Welcome to the InfluxDB configuration file.
|
||||
|
||||
# The values in this file override the default values used by the system if
|
||||
# a config option is not specified. The commented out lines are the configuration
|
||||
# field and the default value used. Uncommenting a line and changing the value
|
||||
# will change the value used at runtime when the process is restarted.
|
||||
|
||||
# Once every 24 hours InfluxDB will report usage data to usage.influxdata.com
|
||||
# The data includes a random ID, os, arch, version, the number of series and other
|
||||
# usage data. No data from user databases is ever transmitted.
|
||||
# Change this option to true to disable reporting.
|
||||
# reporting-disabled = false
|
||||
|
||||
# Bind address to use for the RPC service for backup and restore.
|
||||
# bind-address = "127.0.0.1:8088"
|
||||
|
||||
###
|
||||
### [meta]
|
||||
###
|
||||
### Controls the parameters for the Raft consensus group that stores metadata
|
||||
### about the InfluxDB cluster.
|
||||
###
|
||||
|
||||
[meta]
|
||||
# Where the metadata/raft database is stored
|
||||
dir = "/var/lib/influxdb/meta"
|
||||
|
||||
# Automatically create a default retention policy when creating a database.
|
||||
# retention-autocreate = true
|
||||
|
||||
# If log messages are printed for the meta service
|
||||
# logging-enabled = true
|
||||
|
||||
###
|
||||
### [data]
|
||||
###
|
||||
### Controls where the actual shard data for InfluxDB lives and how it is
|
||||
### flushed from the WAL. "dir" may need to be changed to a suitable place
|
||||
### for your system, but the WAL settings are an advanced configuration. The
|
||||
### defaults should work for most systems.
|
||||
###
|
||||
|
||||
[data]
|
||||
# The directory where the TSM storage engine stores TSM files.
|
||||
dir = "/var/lib/influxdb/data"
|
||||
|
||||
# The directory where the TSM storage engine stores WAL files.
|
||||
wal-dir = "/var/lib/influxdb/wal"
|
||||
|
||||
# The amount of time that a write will wait before fsyncing. A duration
|
||||
# greater than 0 can be used to batch up multiple fsync calls. This is useful for slower
|
||||
# disks or when WAL write contention is seen. A value of 0s fsyncs every write to the WAL.
|
||||
# Values in the range of 0-100ms are recommended for non-SSD disks.
|
||||
# wal-fsync-delay = "0s"
|
||||
|
||||
|
||||
# The type of shard index to use for new shards. The default is an in-memory index that is
|
||||
# recreated at startup. A value of "tsi1" will use a disk based index that supports higher
|
||||
# cardinality datasets.
|
||||
# index-version = "inmem"
|
||||
|
||||
# Trace logging provides more verbose output around the tsm engine. Turning
|
||||
# this on can provide more useful output for debugging tsm engine issues.
|
||||
# trace-logging-enabled = false
|
||||
|
||||
# Whether queries should be logged before execution. Very useful for troubleshooting, but will
|
||||
# log any sensitive data contained within a query.
|
||||
# query-log-enabled = true
|
||||
|
||||
# Settings for the TSM engine
|
||||
|
||||
# CacheMaxMemorySize is the maximum size a shard's cache can
|
||||
# reach before it starts rejecting writes.
|
||||
# cache-max-memory-size = 1048576000
|
||||
|
||||
# CacheSnapshotMemorySize is the size at which the engine will
|
||||
# snapshot the cache and write it to a TSM file, freeing up memory
|
||||
# cache-snapshot-memory-size = 26214400
|
||||
|
||||
# CacheSnapshotWriteColdDuration is the length of time at
|
||||
# which the engine will snapshot the cache and write it to
|
||||
# a new TSM file if the shard hasn't received writes or deletes
|
||||
# cache-snapshot-write-cold-duration = "10m"
|
||||
|
||||
# CompactFullWriteColdDuration is the duration at which the engine
|
||||
# will compact all TSM files in a shard if it hasn't received a
|
||||
# write or delete
|
||||
# compact-full-write-cold-duration = "4h"
|
||||
|
||||
# The maximum number of concurrent full and level compactions that can run at one time. A
|
||||
# value of 0 results in runtime.GOMAXPROCS(0) used at runtime. This setting does not apply
|
||||
# to cache snapshotting.
|
||||
# max-concurrent-compactions = 0
|
||||
|
||||
# The maximum series allowed per database before writes are dropped. This limit can prevent
|
||||
# high cardinality issues at the database level. This limit can be disabled by setting it to
|
||||
# 0.
|
||||
# max-series-per-database = 1000000
|
||||
|
||||
# The maximum number of tag values per tag that are allowed before writes are dropped. This limit
|
||||
# can prevent high cardinality tag values from being written to a measurement. This limit can be
|
||||
# disabled by setting it to 0.
|
||||
# max-values-per-tag = 100000
|
||||
|
||||
###
|
||||
### [coordinator]
|
||||
###
|
||||
### Controls the clustering service configuration.
|
||||
###
|
||||
|
||||
[coordinator]
|
||||
# The default time a write request will wait until a "timeout" error is returned to the caller.
|
||||
# write-timeout = "10s"
|
||||
|
||||
# The maximum number of concurrent queries allowed to be executing at one time. If a query is
|
||||
# executed and exceeds this limit, an error is returned to the caller. This limit can be disabled
|
||||
# by setting it to 0.
|
||||
# max-concurrent-queries = 0
|
||||
|
||||
# The maximum time a query will is allowed to execute before being killed by the system. This limit
|
||||
# can help prevent run away queries. Setting the value to 0 disables the limit.
|
||||
# query-timeout = "0s"
|
||||
|
||||
# The time threshold when a query will be logged as a slow query. This limit can be set to help
|
||||
# discover slow or resource intensive queries. Setting the value to 0 disables the slow query logging.
|
||||
# log-queries-after = "0s"
|
||||
|
||||
# The maximum number of points a SELECT can process. A value of 0 will make
|
||||
# the maximum point count unlimited. This will only be checked every 10 seconds so queries will not
|
||||
# be aborted immediately when hitting the limit.
|
||||
# max-select-point = 0
|
||||
|
||||
# The maximum number of series a SELECT can run. A value of 0 will make the maximum series
|
||||
# count unlimited.
|
||||
# max-select-series = 0
|
||||
|
||||
# The maxium number of group by time bucket a SELECT can create. A value of zero will max the maximum
|
||||
# number of buckets unlimited.
|
||||
# max-select-buckets = 0
|
||||
|
||||
###
|
||||
### [retention]
|
||||
###
|
||||
### Controls the enforcement of retention policies for evicting old data.
|
||||
###
|
||||
|
||||
[retention]
|
||||
# Determines whether retention policy enforcement enabled.
|
||||
# enabled = true
|
||||
|
||||
# The interval of time when retention policy enforcement checks run.
|
||||
# check-interval = "30m"
|
||||
|
||||
###
|
||||
### [shard-precreation]
|
||||
###
|
||||
### Controls the precreation of shards, so they are available before data arrives.
|
||||
### Only shards that, after creation, will have both a start- and end-time in the
|
||||
### future, will ever be created. Shards are never precreated that would be wholly
|
||||
### or partially in the past.
|
||||
|
||||
[shard-precreation]
|
||||
# Determines whether shard pre-creation service is enabled.
|
||||
# enabled = true
|
||||
|
||||
# The interval of time when the check to pre-create new shards runs.
|
||||
# check-interval = "10m"
|
||||
|
||||
# The default period ahead of the endtime of a shard group that its successor
|
||||
# group is created.
|
||||
# advance-period = "30m"
|
||||
|
||||
###
|
||||
### Controls the system self-monitoring, statistics and diagnostics.
|
||||
###
|
||||
### The internal database for monitoring data is created automatically if
|
||||
### if it does not already exist. The target retention within this database
|
||||
### is called 'monitor' and is also created with a retention period of 7 days
|
||||
### and a replication factor of 1, if it does not exist. In all cases the
|
||||
### this retention policy is configured as the default for the database.
|
||||
|
||||
[monitor]
|
||||
# Whether to record statistics internally.
|
||||
# store-enabled = true
|
||||
|
||||
# The destination database for recorded statistics
|
||||
# store-database = "_internal"
|
||||
|
||||
# The interval at which to record statistics
|
||||
# store-interval = "10s"
|
||||
|
||||
###
|
||||
### [http]
|
||||
###
|
||||
### Controls how the HTTP endpoints are configured. These are the primary
|
||||
### mechanism for getting data into and out of InfluxDB.
|
||||
###
|
||||
|
||||
[http]
|
||||
# Determines whether HTTP endpoint is enabled.
|
||||
# enabled = true
|
||||
|
||||
# The bind address used by the HTTP service.
|
||||
# bind-address = ":8086"
|
||||
|
||||
# Determines whether user authentication is enabled over HTTP/HTTPS.
|
||||
# auth-enabled = false
|
||||
|
||||
# The default realm sent back when issuing a basic auth challenge.
|
||||
# realm = "InfluxDB"
|
||||
|
||||
# Determines whether HTTP request logging is enabled.
|
||||
# log-enabled = true
|
||||
|
||||
# Determines whether detailed write logging is enabled.
|
||||
# write-tracing = false
|
||||
|
||||
# Determines whether the pprof endpoint is enabled. This endpoint is used for
|
||||
# troubleshooting and monitoring.
|
||||
# pprof-enabled = true
|
||||
|
||||
# Determines whether HTTPS is enabled.
|
||||
# https-enabled = false
|
||||
|
||||
# The SSL certificate to use when HTTPS is enabled.
|
||||
# https-certificate = "/etc/ssl/influxdb.pem"
|
||||
|
||||
# Use a separate private key location.
|
||||
# https-private-key = ""
|
||||
|
||||
# The JWT auth shared secret to validate requests using JSON web tokens.
|
||||
# shared-secret = ""
|
||||
|
||||
# The default chunk size for result sets that should be chunked.
|
||||
# max-row-limit = 0
|
||||
|
||||
# The maximum number of HTTP connections that may be open at once. New connections that
|
||||
# would exceed this limit are dropped. Setting this value to 0 disables the limit.
|
||||
# max-connection-limit = 0
|
||||
|
||||
# Enable http service over unix domain socket
|
||||
# unix-socket-enabled = false
|
||||
|
||||
# The path of the unix domain socket.
|
||||
# bind-socket = "/var/run/influxdb.sock"
|
||||
|
||||
###
|
||||
### [subscriber]
|
||||
###
|
||||
### Controls the subscriptions, which can be used to fork a copy of all data
|
||||
### received by the InfluxDB host.
|
||||
###
|
||||
|
||||
[subscriber]
|
||||
# Determines whether the subscriber service is enabled.
|
||||
# enabled = true
|
||||
|
||||
# The default timeout for HTTP writes to subscribers.
|
||||
# http-timeout = "30s"
|
||||
|
||||
# Allows insecure HTTPS connections to subscribers. This is useful when testing with self-
|
||||
# signed certificates.
|
||||
# insecure-skip-verify = false
|
||||
|
||||
# The path to the PEM encoded CA certs file. If the empty string, the default system certs will be used
|
||||
# ca-certs = ""
|
||||
|
||||
# The number of writer goroutines processing the write channel.
|
||||
# write-concurrency = 40
|
||||
|
||||
# The number of in-flight writes buffered in the write channel.
|
||||
# write-buffer-size = 1000
|
||||
|
||||
|
||||
###
|
||||
### [[graphite]]
|
||||
###
|
||||
### Controls one or many listeners for Graphite data.
|
||||
###
|
||||
|
||||
[[graphite]]
|
||||
# Determines whether the graphite endpoint is enabled.
|
||||
# enabled = false
|
||||
# database = "graphite"
|
||||
# retention-policy = ""
|
||||
# bind-address = ":2003"
|
||||
# protocol = "tcp"
|
||||
# consistency-level = "one"
|
||||
|
||||
# These next lines control how batching works. You should have this enabled
|
||||
# otherwise you could get dropped metrics or poor performance. Batching
|
||||
# will buffer points in memory if you have many coming in.
|
||||
|
||||
# Flush if this many points get buffered
|
||||
# batch-size = 5000
|
||||
|
||||
# number of batches that may be pending in memory
|
||||
# batch-pending = 10
|
||||
|
||||
# Flush at least this often even if we haven't hit buffer limit
|
||||
# batch-timeout = "1s"
|
||||
|
||||
# UDP Read buffer size, 0 means OS default. UDP listener will fail if set above OS max.
|
||||
# udp-read-buffer = 0
|
||||
|
||||
### This string joins multiple matching 'measurement' values providing more control over the final measurement name.
|
||||
# separator = "."
|
||||
|
||||
### Default tags that will be added to all metrics. These can be overridden at the template level
|
||||
### or by tags extracted from metric
|
||||
# tags = ["region=us-east", "zone=1c"]
|
||||
|
||||
### Each template line requires a template pattern. It can have an optional
|
||||
### filter before the template and separated by spaces. It can also have optional extra
|
||||
### tags following the template. Multiple tags should be separated by commas and no spaces
|
||||
### similar to the line protocol format. There can be only one default template.
|
||||
# templates = [
|
||||
# "*.app env.service.resource.measurement",
|
||||
# # Default template
|
||||
# "server.*",
|
||||
# ]
|
||||
|
||||
###
|
||||
### [collectd]
|
||||
###
|
||||
### Controls one or many listeners for collectd data.
|
||||
###
|
||||
|
||||
[[collectd]]
|
||||
# enabled = false
|
||||
# bind-address = ":25826"
|
||||
# database = "collectd"
|
||||
# retention-policy = ""
|
||||
#
|
||||
# The collectd service supports either scanning a directory for multiple types
|
||||
# db files, or specifying a single db file.
|
||||
# typesdb = "/usr/local/share/collectd"
|
||||
#
|
||||
# security-level = "none"
|
||||
# auth-file = "/etc/collectd/auth_file"
|
||||
|
||||
# These next lines control how batching works. You should have this enabled
|
||||
# otherwise you could get dropped metrics or poor performance. Batching
|
||||
# will buffer points in memory if you have many coming in.
|
||||
|
||||
# Flush if this many points get buffered
|
||||
# batch-size = 5000
|
||||
|
||||
# Number of batches that may be pending in memory
|
||||
# batch-pending = 10
|
||||
|
||||
# Flush at least this often even if we haven't hit buffer limit
|
||||
# batch-timeout = "10s"
|
||||
|
||||
# UDP Read buffer size, 0 means OS default. UDP listener will fail if set above OS max.
|
||||
# read-buffer = 0
|
||||
|
||||
###
|
||||
### [opentsdb]
|
||||
###
|
||||
### Controls one or many listeners for OpenTSDB data.
|
||||
###
|
||||
|
||||
[[opentsdb]]
|
||||
# enabled = false
|
||||
# bind-address = ":4242"
|
||||
# database = "opentsdb"
|
||||
# retention-policy = ""
|
||||
# consistency-level = "one"
|
||||
# tls-enabled = false
|
||||
# certificate= "/etc/ssl/influxdb.pem"
|
||||
|
||||
# Log an error for every malformed point.
|
||||
# log-point-errors = true
|
||||
|
||||
# These next lines control how batching works. You should have this enabled
|
||||
# otherwise you could get dropped metrics or poor performance. Only points
|
||||
# metrics received over the telnet protocol undergo batching.
|
||||
|
||||
# Flush if this many points get buffered
|
||||
# batch-size = 1000
|
||||
|
||||
# Number of batches that may be pending in memory
|
||||
# batch-pending = 5
|
||||
|
||||
# Flush at least this often even if we haven't hit buffer limit
|
||||
# batch-timeout = "1s"
|
||||
|
||||
###
|
||||
### [[udp]]
|
||||
###
|
||||
### Controls the listeners for InfluxDB line protocol data via UDP.
|
||||
###
|
||||
|
||||
[[udp]]
|
||||
# enabled = false
|
||||
# bind-address = ":8089"
|
||||
# database = "udp"
|
||||
# retention-policy = ""
|
||||
|
||||
# These next lines control how batching works. You should have this enabled
|
||||
# otherwise you could get dropped metrics or poor performance. Batching
|
||||
# will buffer points in memory if you have many coming in.
|
||||
|
||||
# Flush if this many points get buffered
|
||||
# batch-size = 5000
|
||||
|
||||
# Number of batches that may be pending in memory
|
||||
# batch-pending = 10
|
||||
|
||||
# Will flush at least this often even if we haven't hit buffer limit
|
||||
# batch-timeout = "1s"
|
||||
|
||||
# UDP Read buffer size, 0 means OS default. UDP listener will fail if set above OS max.
|
||||
# read-buffer = 0
|
||||
|
||||
###
|
||||
### [continuous_queries]
|
||||
###
|
||||
### Controls how continuous queries are run within InfluxDB.
|
||||
###
|
||||
|
||||
[continuous_queries]
|
||||
# Determines whether the continuous query service is enabled.
|
||||
# enabled = true
|
||||
|
||||
# Controls whether queries are logged when executed by the CQ service.
|
||||
# log-enabled = true
|
||||
|
||||
# interval for how often continuous queries will be checked if they need to run
|
||||
# run-interval = "1s"
|
||||
Reference in New Issue
Block a user