Personal tools
Skip to content. | Skip to navigation
This project adds a basic high availability layer to InfluxDB. With the right architecture and disaster recovery processes, this achieves a highly available setup. The architecture is fairly simple and consists of a load balancer, two or more InfluxDB Relay processes and two or more InfluxDB processes. The load balancer should point UDP traffic and HTTP POST requests with the path /write to the two relays while pointing GET requests with the path /query to the two InfluxDB servers. Buffering The relay can be configured to buffer failed requests for HTTP backends. The intent of this logic is reduce the number of failures during short outages or periodic network issues. This retry logic is NOT sufficient for for long periods of downtime as all data is buffered in RAM Buffering has the following configuration options (configured per HTTP backend): buffer-size-mb -- An upper limit on how much point data to keep in memory (in MB) max-batch-kb -- A maximum size on the aggregated batches that will be submitted (in KB) max-delay-interval -- the max delay between retry attempts per backend. The initial retry delay is 500ms and is doubled after every failure. If the buffer is full then requests are dropped and an error is logged. If a requests makes it into the buffer it is retried until success. Retries are serialized to a single backend. In addition, writes will be aggregated and batched as long as the body of the request will be less than max-batch-kb If buffered requests succeed then there is no delay between subsequent attempts. If the relay stays alive the entire duration of a downed backend server without filling that server's allocated buffer, and the relay can stay online until the entire buffer is flushed, it would mean that no operator intervention would be required to "recover" the data. The data will simply be batched together and written out to the recovered server in the order it was received. NOTE: The limits for buffering are not hard limits on the memory usage of the application, and there will be additional overhead that would be much more challenging to account for. The limits listed are just for the amount of point line protocol (including any added timestamps, if applicable). Factors such as small incoming batch sizes and a smaller max batch size will increase the overhead in the buffer. There is also the general application memory overhead to account for. This means that a machine with 2GB of memory should not have buffers that sum up to almost 2GB. Recovery InfluxDB organizes its data on disk into logical blocks of time called shards. We can use this to create a hot recovery process with zero downtime. The length of time that shards represent in InfluxDB are typically 1 hour, 1 day, or 7 days, depending on the retention duration, but can be explicitly set when creating the retention policy. For the sake of our example, let's assume shard durations of 1 day. Let's say one of the InfluxDB servers goes down for an hour on 2016-03-10. Once midnight UTC rolls over, all InfluxDB processes are now writing data to the shard for 2016-03-11 and the file(s) for 2016-03-10 have gone cold for writes. We can then restore things using these steps: Tell the load balancer to stop sending query traffic to the server that was down (this should be done as soon as an outage is detected to prevent partial or inconsistent query returns.) Create backup of 2016-03-10 shard from a server that was up the entire day Restore the backup of the shard from the good server to the server that had downtime Tell the load balancer to resume sending queries to the previously downed server During this entire process the Relays should be sending current writes to all servers, including the one with downtime. Sharding It's possible to add another layer on top of this kind of setup to shard data. Depending on your needs you could shard on the measurement name or a specific tag like customer_id. The sharding layer would have to service both queries and writes. As this relay does not handle queries, it will not implement any sharding logic. Any sharding would have to be done externally to the relay. Caveats While influxdb-relay does provide some level of high availability, there are a few scenarios that need to be accounted for: influxdb-relay will not relay the /query endpoint, and this includes schema modification (create database, DROPs, etc). This means that databases must be created before points are written to the backends. Continuous queries will still only write their results locally. If a server goes down, the continuous query will have to be backfilled after the data has been recovered for that instance. Overwriting points is potentially unpredictable. For example, given servers A and B, if B is down, and point X is written (we'll call the value X1) just before B comes back online, that write is queued behind every other write that occurred while B was offline. Once B is back online, the first buffered write succeeds, and all new writes are now allowed to pass-through. At this point (before X1 is written to B), X is written again (with value X2 this time) to both A and B. When the relay reaches the end of B's buffered writes, it will write X (with value X1) to B... At this point A now has X2, but B has X1. It is probably best to avoid re-writing points (if possible). Otherwise, please be aware that overwriting the same field for a given point can lead to data differences. This could potentially be mitigated by waiting for the buffer to flush before opening writes back up to being passed-through.
Poll network devices via SNMP and save the data in InfluxDB (version 0.12.x) It uses github.com/paulstuart/snmputil for snmp processing, and therefore has the following functionality: SNMP versions 1, 2/2c, 3 Bulk polling of all tabular data Regexp filtering by name of resulting data Auto conversion of INTEGER and BIT formats to their named types Auto generating OID lookup for names (if net-snmp-utils is installed) Optional processing of counter data (deltas and differentials) Overide column aliases with custom labels Auto throttling of requests - never poll faster than device can respond influxsnmp uses a datafile of parsed MIB objects in order to use symbolic names and to do automated formatting of polled data. If a previously saved file is not available, it will generate and same one automatically. The resulting file of such actions may be quite large (all OIDs included). To create a MIB file of only the OIDs that will be used, run the following command: influxsnmp -dump -filter > mibFile.json As it is using snmptranslate to create the dump file, one can export MIBDIRS to point to the directories containing mib files
Performance Co-Pilot (PCP) front-end tools for exporting metric values to InfluxDB (https://influxdata.com/time-series-platform/influxdb).
statsd_exporter receives StatsD-style metrics and exports them as Prometheus metrics. Overview With StatsD To pipe metrics from an existing StatsD environment into Prometheus, configure StatsD's repeater backend to repeat all received metrics to a statsd_exporter process. This exporter translates StatsD metrics to Prometheus metrics via configured mapping rules. +----------+ +-------------------+ +--------------+ | StatsD |---(UDP/TCP repeater)--->| statsd_exporter |<---(scrape /metrics)---| Prometheus | +----------+ +-------------------+ +--------------+ Without StatsD Since the StatsD exporter uses the same line protocol as StatsD itself, you can also configure your applications to send StatsD metrics directly to the exporter. In that case, you don't need to run a StatsD server anymore. We recommend this only as an intermediate solution and recommend switching to native Prometheus instrumentation in the long term. Tagging Extensions The exporter supports Librato, InfluxDB, and DogStatsD-style tags, which will be converted into Prometheus labels. For Librato-style tags, they must be appended to the metric name with a delimiting #, as so: metric.name#tagName=val,tag2Name=val2:0|c See the statsd-librato-backend README for a more complete description. For InfluxDB-style tags, they must be appended to the metric name with a delimiting comma, as so: metric.name,tagName=val,tag2Name=val2:0|c See this InfluxDB blog post for a larger overview. For DogStatsD-style tags, they're appended as a |# delimited section at the end of the metric, as so: metric.name:0|c|#tagName=val,tag2Name=val2 See Tags in the DogStatsD documentation for the concept description and Datagram Format. If you encounter problems, note that this tagging style is incompatible with the original statsd implementation. Be aware: If you mix tag styles (e.g., Librato/InfluxDB with DogStatsD), the exporter will consider this an error and the sample will be discarded. Also, tags without values (#some_tag) are not supported and will be ignored.
This is a metapackage bringing in influxdb extras requires for python3-apache-airflow. It makes sure the dependencies are installed.
Apache/Airflow influxdb provider
Package apache-airflow-providers-influxdb Release: 2.5.1 InfluxDB
InfluxDBPython is a client for interacting with InfluxDB_. Maintained by @aviau (https://github.com/aviau).**Help needed:** Development of this library is made by the community and help is needed. A comaintainer would be welcome. To contribute, take a look at the issues list of simply contact @aviau... image:: https://travisci.org/influxdata/influxdbpython.svg?branchmaster :target: ...
A collecting tool of system metrics (CPU, memory, load, disks I/Os, network traffic) to an InfluxDB server. This project mainly relies on gosigar, so it's compatible with GNU/Linux and MacOS system, but not with Windows yet.