OpenStack and Prometheus Integration

Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.

info

This is not the recommended configuration for real-time query at scale. For query and compression optimization, high-speed ingest, and high availability, you may want to consider OpenStack and InfluxDB.

5B+

Telegraf downloads

#1

Time series database
Source: DB Engines

1B+

Downloads of InfluxDB

2,800+

Contributors

Table of Contents

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Input and output integration overview

This plugin collects metrics from essential OpenStack services, facilitating the monitoring and management of cloud infrastructures.

The Prometheus Output Plugin enables Telegraf to expose metrics at an HTTP endpoint for scraping by a Prometheus server. This integration allows users to collect and aggregate metrics from various sources in a format that Prometheus can process efficiently.

Integration details

OpenStack

The OpenStack plugin allows users to collect performance metrics from various OpenStack services such as CINDER, GLANCE, HEAT, KEYSTONE, NEUTRON, and NOVA. It supports multiple OpenStack APIs to fetch critical metrics related to these services, enabling comprehensive monitoring and management of cloud resources. As organizations increasingly adopt OpenStack for their cloud infrastructure, this plugin plays a vital role in providing insights into resource usage, availability, and performance across the cloud environment. Configuration options allow for customized polling intervals and filtering unwanted tags to optimize performance and cardinals.

Prometheus

This plugin for facilitates the integration with Prometheus, a well-known open-source monitoring and alerting toolkit designed for reliability and efficiency in large-scale environments. By working as a Prometheus client, it allows users to expose a defined set of metrics via an HTTP server that Prometheus can scrape at specified intervals. This plugin plays a crucial role in monitoring diverse systems by allowing them to publish performance metrics in a standardized format, enabling extensive visibility into system health and behavior. Key features include support for configuring various endpoints, enabling TLS for secure communication, and options for HTTP basic authentication. The plugin also integrates seamlessly with global Telegraf configuration settings, supporting extensive customization to fit specific monitoring needs. This promotes interoperability in environments where different systems must communicate performance data effectively. Leveraging Prometheus’s metric format, it allows for flexible metric management through advanced configurations such as metric expiration and collectors control, offering a sophisticated solution for monitoring and alerting workflows.

Configuration

OpenStack

[[inputs.openstack]]
  ## The recommended interval to poll is '30m'

  ## The identity endpoint to authenticate against and get the service catalog from.
  authentication_endpoint = "https://my.openstack.cloud:5000"

  ## The domain to authenticate against when using a V3 identity endpoint.
  # domain = "default"

  ## The project to authenticate as.
  # project = "admin"

  ## User authentication credentials. Must have admin rights.
  username = "admin"
  password = "password"

  ## Available services are:
  ## "agents", "aggregates", "cinder_services", "flavors", "hypervisors",
  ## "networks", "nova_services", "ports", "projects", "servers",
  ## "serverdiagnostics", "services", "stacks", "storage_pools", "subnets",
  ## "volumes"
  # enabled_services = ["services", "projects", "hypervisors", "flavors", "networks", "volumes"]

  ## Query all instances of all tenants for the volumes and server services
  ## NOTE: Usually this is only permitted for administrators!
  # query_all_tenants = true

  ## output secrets (such as adminPass(for server) and UserID(for volume)).
  # output_secrets = false

  ## Amount of time allowed to complete the HTTP(s) request.
  # timeout = "5s"

  ## HTTP Proxy support
  # http_proxy_url = ""

  ## Optional TLS Config
  # tls_ca = /path/to/cafile
  # tls_cert = /path/to/certfile
  # tls_key = /path/to/keyfile
  ## Use TLS but skip chain & host verification
  # insecure_skip_verify = false

  ## Options for tags received from Openstack
  # tag_prefix = "openstack_tag_"
  # tag_value = "true"

  ## Timestamp format for timestamp data received from Openstack.
  ## If false format is unix nanoseconds.
  # human_readable_timestamps = false

  ## Measure Openstack call duration
  # measure_openstack_requests = false

Prometheus

[[outputs.prometheus_client]]
  ## Address to listen on.
  ##   ex:
  ##     listen = ":9273"
  ##     listen = "vsock://:9273"
  listen = ":9273"

  ## Maximum duration before timing out read of the request
  # read_timeout = "10s"
  ## Maximum duration before timing out write of the response
  # write_timeout = "10s"

  ## Metric version controls the mapping from Prometheus metrics into Telegraf metrics.
  ## See "Metric Format Configuration" in plugins/inputs/prometheus/README.md for details.
  ## Valid options: 1, 2
  # metric_version = 1

  ## Use HTTP Basic Authentication.
  # basic_username = "Foo"
  # basic_password = "Bar"

  ## If set, the IP Ranges which are allowed to access metrics.
  ##   ex: ip_range = ["192.168.0.0/24", "192.168.1.0/30"]
  # ip_range = []

  ## Path to publish the metrics on.
  # path = "/metrics"

  ## Expiration interval for each metric. 0 == no expiration
  # expiration_interval = "60s"

  ## Collectors to enable, valid entries are "gocollector" and "process".
  ## If unset, both are enabled.
  # collectors_exclude = ["gocollector", "process"]

  ## Send string metrics as Prometheus labels.
  ## Unless set to false all string metrics will be sent as labels.
  # string_as_label = true

  ## If set, enable TLS with the given certificate.
  # tls_cert = "/etc/ssl/telegraf.crt"
  # tls_key = "/etc/ssl/telegraf.key"

  ## Set one or more allowed client CA certificate file names to
  ## enable mutually authenticated TLS connections
  # tls_allowed_cacerts = ["/etc/telegraf/clientca.pem"]

  ## Export metric collection time.
  # export_timestamp = false

  ## Specify the metric type explicitly.
  ## This overrides the metric-type of the Telegraf metric. Globbing is allowed.
  # [outputs.prometheus_client.metric_types]
  #   counter = []
  #   gauge = []

Input and output integration examples

OpenStack

  1. Cross-Cloud Management: Leverage the OpenStack plugin to monitor and manage multiple OpenStack clouds from a single Telegraf instance. By aggregating metrics across different clouds, organizations can gain insights into resource utilization and optimize their cloud architecture for cost and performance.

  2. Automated Scaling Based on Metrics: Integrate the metrics gathered from OpenStack into an automated scaling solution. For example, if the plugin detects that a specific service’s performance is degraded, it can trigger auto-scaling rules to launch additional instances, ensuring that system performance remains optimal under varying workloads.

  3. Performance Monitoring Dashboard: Use data collected by the OpenStack Telegraf plugin to power real-time monitoring dashboards. This setup provides visualizations of key metrics from OpenStack services, enabling stakeholders to quickly identify trends, pinpoint issues, and make data-driven decisions in managing their cloud infrastructure.

  4. Reporting and Analysis of Service Availability: By utilizing the metrics collected from various OpenStack services, teams can generate detailed reports on service availability and performance over time. This information can help identify recurring issues, improve service delivery, and make informed decisions regarding changes in infrastructure or service configuration.

Prometheus

  1. Monitoring Multi-cloud Deployments: Utilize the Prometheus plugin to collect metrics from applications running across multiple cloud providers. This scenario allows teams to centralize monitoring through a single Prometheus instance that scrapes metrics from different environments, providing a unified view of performance metrics across hybrid infrastructures. It streamlines reporting and alerting, enhancing operational efficiency without needing complex integrations.

  2. Enhancing Microservices Visibility: Implement the plugin to expose metrics from various microservices within a Kubernetes cluster. Using Prometheus, teams can visualize service metrics in real time, identify bottlenecks, and maintain system health checks. This setup supports adaptive scaling and resource utilization optimization based on insights generated from the collected metrics. It enhances the ability to troubleshoot service interactions, significantly improving the resilience of the microservice architecture.

  3. Real-time Anomaly Detection in E-commerce: By leveraging this plugin alongside Prometheus, an e-commerce platform can monitor key performance indicators such as response times and error rates. Integrating anomaly detection algorithms with scraped metrics allows the identification of unexpected patterns indicating potential issues, such as sudden traffic spikes or backend service failure. This proactive monitoring empowers business continuity and operational efficiency, minimizing potential downtimes while ensuring service reliability.

  4. Performance Metrics Reporting for APIs: Utilize the Prometheus Output Plugin to gather and report API performance metrics, which can then be visualized in Grafana dashboards. This use case enables detailed analysis of API response times, throughput, and error rates, promoting continuous improvement of API services. By closely monitoring these metrics, teams can quickly react to degradation, ensuring optimal API performance and maintaining a high level of service availability.

Feedback

Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Related Integrations

HTTP and InfluxDB Integration

The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.

View Integration

Kafka and InfluxDB Integration

This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.

View Integration

Kinesis and InfluxDB Integration

The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.

View Integration