Documentation

Identify write methods

Many different tools are available for writing data into your InfluxDB cluster. Based on your use case, you should identify the most appropriate tools and methods to use. Below is a summary of some of the tools that are available (this list is not exhaustive).

Telegraf

Telegraf is a data collection agent that collects data from various sources, parses the data into line protocol, and then writes the data to InfluxDB. Telegraf is plugin-based and provides hundreds of plugins that collect, aggregate, process, and write data.

If you need to collect data from well-established systems and technologies, Telegraf likely already supports a plugin for collecting that data. Some of the most common use cases are:

  • Monitoring system metrics (memory, CPU, disk usage, etc.)
  • Monitoring Docker containers
  • Monitoring network devices via SNMP
  • Collecting data from a Kafka queue
  • Collecting data from an MQTT broker
  • Collecting data from HTTP endpoints
  • Scraping data from a Prometheus exporter
  • Parsing logs

For more information about using Telegraf with InfluxDB Clustered, see Use Telegraf to write data to InfluxDB Clustered.

InfluxDB client libraries

InfluxDB client libraries are language-specific packages that integrate with InfluxDB APIs. They simplify integrating InfluxDB with your own custom application and standardize interactions between your application and your InfluxDB cluster. With client libraries, you can collect and write whatever time series data is useful for your application.

InfluxDB Clustered includes backwards compatible write APIs, so if you are currently using an InfluxDB v1 or v2 client library, you can continue to use the same client library to write data to your cluster.

InfluxDB HTTP write APIs

InfluxDB Clustered provides backwards-compatible HTTP write APIs for writing data to your cluster. The InfluxDB client libraries use these APIs, but if you choose not to use a client library, you can integrate directly with the API. Because these APIs are backwards compatible, you can use existing InfluxDB API integrations with your InfluxDB cluster.

Write optimizations

As you decide on and integrate tooling to write data to your InfluxDB cluster, there are things you can do to ensure your write pipeline is as performant as possible. The list below provides links to more detailed descriptions of these optimizations in the Optimize writes documentation:

Telegraf and InfluxDB client libraries leverage many of these optimizations by default.


Was this page helpful?

Thank you for your feedback!


InfluxDB OSS 2.9.0: API tokens are hashed by default

Stronger token security in InfluxDB OSS 2.9.0 — tokens are hashed on disk by default. Existing tokens are hashed on first startup and can’t be recovered afterward. Capture any plaintext tokens you still need before you upgrade.

View InfluxDB OSS 2.9.0 release notes

Hashed tokens authenticate exactly like unhashed tokens — clients and integrations keep working.

Also new in 2.9.0:

  • Configurable backup compression
  • Restore support for backups containing hashed tokens
  • Tighter Edge Data Replication queue validation
  • Flux upgrade
  • Compaction reliability improvements

Key enhancements in Explorer 1.8

Explorer 1.8 is now available with streaming data subscriptions (beta), line protocol preview, and query history & saved queries.

View Explorer 1.8 release notes

Explorer 1.8 includes new features and improvements that make it easier to ingest, explore, and manage data.

Highlights:

  • Streaming data subscriptions (beta): Stream data into Explorer from MQTT, Kafka, and AMQP sources.
  • Line protocol preview: Preview line protocol, schema, and parse errors before data is written.
  • Custom sample data: Generate custom sample datasets with line protocol and schema preview.
  • Query history and saved queries: Browse query history and save/re-run named queries.
  • Retention period management: Set, update, or clear retention periods on databases and tables.

For more details, see Explorer 1.8 release notes

InfluxDB 3.9: Performance upgrade preview

InfluxDB 3 Enterprise 3.9 includes a beta of major performance upgrades with faster single-series queries, wide-and-sparse table support, and more.

InfluxDB 3 Enterprise 3.9 includes a beta of major performance and feature updates.

Key improvements:

  • Faster single-series queries
  • Consistent resource usage
  • Wide-and-sparse table support
  • Automatic distinct value caches for reduced latency with metadata queries

Preview features are subject to breaking changes.

For more information, see:

Telegraf Enterprise now in public beta

Get early access to the Telegraf Controller and provide feedback to help shape the future of Telegraf Enterprise.

See the Blog Post

The upcoming Telegraf Enterprise offering is for organizations running Telegraf at scale and is comprised of two key components:

  • Telegraf Controller: A control plane (UI + API) that centralizes Telegraf configuration management and agent health visibility.
  • Telegraf Enterprise Support: Official support for Telegraf Controller and Telegraf plugins.

Join the Telegraf Enterprise beta to get early access to the Telegraf Controller and provide feedback to help shape the future of Telegraf Enterprise.

For more information:

Telegraf Controller v0.0.7-beta now available

Telegraf Controller v0.0.7-beta is now available with new features, improvements, bug fixes, and an important breaking change.

View the release notes
Download Telegraf Controller v0.0.7-beta

InfluxDB Docker latest tag changing to InfluxDB 3 Core

On May 27, 2026, the latest tag for InfluxDB Docker images will point to InfluxDB 3 Core. To avoid unexpected upgrades, use specific version tags in your Docker deployments.

If using Docker to install and run InfluxDB, the latest tag will point to InfluxDB 3 Core. To avoid unexpected upgrades, use specific version tags in your Docker deployments. For example, if using Docker to run InfluxDB v2, replace the latest version tag with a specific version tag in your Docker pull command–for example:

docker pull influxdb:2