SNMP and Cortex Integration

Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.

info

This is not the recommended configuration for real-time query at scale. For query and compression optimization, high-speed ingest, and high availability, you may want to consider SNMP and InfluxDB.

5B+

Telegraf downloads

#1

Time series database
Source: DB Engines

1B+

Downloads of InfluxDB

2,800+

Contributors

Table of Contents

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Input and output integration overview

The SNMP plugin allows you to collect a variety of metrics from SNMP (Simple Network Management Protocol) agents. It provides flexibility in how data is retrieved, whether collecting single metrics or entire tables.

This plugin enables Telegraf to send metrics to Cortex using the Prometheus remote write protocol, allowing seamless ingestion into Cortex’s scalable, multi-tenant time series storage.

Integration details

SNMP

This plugin uses polling to gather metrics from SNMP agents, supporting retrieval of individual OIDs and complete SNMP tables. It can be configured to handle multiple SNMP versions, authentication, and other features.

Cortex

With Telegraf’s HTTP output plugin and the prometheusremotewrite data format you can send metrics directly to Cortex, a horizontally scalable, long-term storage backend for Prometheus. Cortex supports multi-tenancy and accepts remote write requests using the Prometheus protobuf format. By using Telegraf as the collection agent and Remote Write as the transport mechanism, organizations can extend observability into sources not natively supported by Prometheus—such as Windows hosts, SNMP-enabled devices, or custom application metrics—while leveraging Cortex’s high-availability and long-retention capabilities.

Configuration

SNMP


[[inputs.snmp]]
  agents = ["udp://127.0.0.1:161"]

  [[inputs.snmp.field]]
    oid = "RFC1213-MIB::sysUpTime.0"
    name = "sysUptime"
    conversion = "float(2)"

  [[inputs.snmp.field]]
    oid = "RFC1213-MIB::sysName.0"
    name = "sysName"
    is_tag = true

  [[inputs.snmp.table]]
    oid = "IF-MIB::ifTable"
    name = "interface"
    inherit_tags = ["sysName"]

    [[inputs.snmp.table.field]]
      oid = "IF-MIB::ifDescr"
      name = "ifDescr"
      is_tag = true

Cortex

[[outputs.http]]
  ## Cortex Remote Write endpoint
  url = "http://cortex.example.com/api/v1/push"

  ## Use POST to send data
  method = "POST"

  ## Send metrics using Prometheus remote write format
  data_format = "prometheusremotewrite"

  ## Optional HTTP headers for authentication
  # [outputs.http.headers]
  #   X-Scope-OrgID = "your-tenant-id"
  #   Authorization = "Bearer YOUR_API_TOKEN"

  ## Optional TLS configuration
  # tls_ca = "/path/to/ca.pem"
  # tls_cert = "/path/to/cert.pem"
  # tls_key = "/path/to/key.pem"
  # insecure_skip_verify = false

  ## Request timeout
  timeout = "10s"

Input and output integration examples

SNMP

  1. Basic SNMP Configuration: Collect metrics from a local SNMP agent using typical SNMP community string settings. This setup is ideal for local monitoring of device performance.
  2. Advanced SNMPv3 Setup: Securely collect metrics using SNMPv3 with authentication and encryption to enhance security. This configuration is recommended for production environments.
  3. Collect Interface Metrics: Configure the plugin to collect interface metrics from the device’s SNMP table. Utilize fields to capture specific data points for traffic analysis.
  4. Join Two SNMP Tables: By using translation fields, join data from two SNMP tables for a comprehensive view of correlated performance metrics.

Cortex

  1. Unified Multi-Tenant Monitoring: Use Telegraf to collect metrics from different teams or environments and push them to Cortex with separate X-Scope-OrgID headers. This enables isolated data ingestion and querying per tenant, ideal for managed services and platform teams.

  2. Extending Prometheus Coverage to Edge Devices: Deploy Telegraf on edge or IoT devices to collect system metrics and send them to a centralized Cortex cluster. This approach ensures consistent observability even for environments without local Prometheus scrapers.

  3. Global Service Observability with Federated Tenants: Aggregate metrics from global infrastructure by configuring Telegraf agents to push data into regional Cortex clusters, each tagged with tenant identifiers. Cortex handles deduplication and centralized access across regions.

  4. Custom App Telemetry Pipeline: Collect app-specific telemetry via Telegraf’s exec or http input plugins and forward it to Cortex. This allows DevOps teams to monitor app-specific KPIs in a scalable, query-efficient format while keeping metrics logically grouped by tenant or service.

Feedback

Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Related Integrations

HTTP and InfluxDB Integration

The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.

View Integration

Kafka and InfluxDB Integration

This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.

View Integration

Kinesis and InfluxDB Integration

The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.

View Integration