Hashicorp Nomad and Prometheus Integration
Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.
5B+
Telegraf downloads
#1
Time series database
Source: DB Engines
1B+
Downloads of InfluxDB
2,800+
Contributors
Table of Contents
Powerful Performance, Limitless Scale
Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.
See Ways to Get Started
Input and output integration overview
This plugin allows users to collect metrics from Hashicorp Nomad agents in distributed environments.
The Prometheus Output Plugin enables Telegraf to expose metrics at an HTTP endpoint for scraping by a Prometheus server. This integration allows users to collect and aggregate metrics from various sources in a format that Prometheus can process efficiently.
Integration details
Hashicorp Nomad
The Hashicorp Nomad input plugin is designed to gather metrics from every Nomad agent within a cluster. By deploying Telegraf on each node, it can connect to the local Nomad agent, typically available at ‘http://127.0.0.1:4646’. With this setup, users can systematically collect and monitor metrics related to the performance and status of their Nomad environment, ensuring they maintain a healthy and efficient cluster operational state. This plugin enables visibility into the operational aspects of Nomad, which is essential for maintaining reliable cloud infrastructure.
Prometheus
This plugin for facilitates the integration with Prometheus, a well-known open-source monitoring and alerting toolkit designed for reliability and efficiency in large-scale environments. By working as a Prometheus client, it allows users to expose a defined set of metrics via an HTTP server that Prometheus can scrape at specified intervals. This plugin plays a crucial role in monitoring diverse systems by allowing them to publish performance metrics in a standardized format, enabling extensive visibility into system health and behavior. Key features include support for configuring various endpoints, enabling TLS for secure communication, and options for HTTP basic authentication. The plugin also integrates seamlessly with global Telegraf configuration settings, supporting extensive customization to fit specific monitoring needs. This promotes interoperability in environments where different systems must communicate performance data effectively. Leveraging Prometheus’s metric format, it allows for flexible metric management through advanced configurations such as metric expiration and collectors control, offering a sophisticated solution for monitoring and alerting workflows.
Configuration
Hashicorp Nomad
[[inputs.nomad]]
## URL for the Nomad agent
# url = "http://127.0.0.1:4646"
## Set response_timeout (default 5 seconds)
# response_timeout = "5s"
## Optional TLS Config
# tls_ca = /path/to/cafile
# tls_cert = /path/to/certfile
# tls_key = /path/to/keyfile
Prometheus
[[outputs.prometheus_client]]
## Address to listen on.
## ex:
## listen = ":9273"
## listen = "vsock://:9273"
listen = ":9273"
## Maximum duration before timing out read of the request
# read_timeout = "10s"
## Maximum duration before timing out write of the response
# write_timeout = "10s"
## Metric version controls the mapping from Prometheus metrics into Telegraf metrics.
## See "Metric Format Configuration" in plugins/inputs/prometheus/README.md for details.
## Valid options: 1, 2
# metric_version = 1
## Use HTTP Basic Authentication.
# basic_username = "Foo"
# basic_password = "Bar"
## If set, the IP Ranges which are allowed to access metrics.
## ex: ip_range = ["192.168.0.0/24", "192.168.1.0/30"]
# ip_range = []
## Path to publish the metrics on.
# path = "/metrics"
## Expiration interval for each metric. 0 == no expiration
# expiration_interval = "60s"
## Collectors to enable, valid entries are "gocollector" and "process".
## If unset, both are enabled.
# collectors_exclude = ["gocollector", "process"]
## Send string metrics as Prometheus labels.
## Unless set to false all string metrics will be sent as labels.
# string_as_label = true
## If set, enable TLS with the given certificate.
# tls_cert = "/etc/ssl/telegraf.crt"
# tls_key = "/etc/ssl/telegraf.key"
## Set one or more allowed client CA certificate file names to
## enable mutually authenticated TLS connections
# tls_allowed_cacerts = ["/etc/telegraf/clientca.pem"]
## Export metric collection time.
# export_timestamp = false
## Specify the metric type explicitly.
## This overrides the metric-type of the Telegraf metric. Globbing is allowed.
# [outputs.prometheus_client.metric_types]
# counter = []
# gauge = []
Input and output integration examples
Hashicorp Nomad
-
Cluster Health Monitoring: Use the Hashicorp Nomad plugin to aggregate metrics across all nodes in a Nomad deployment. By monitoring health metrics such as allocation status, job performance, and resource utilization, operations teams can gain insights into the overall health of their deployment, quickly identify and resolve issues, and optimize resource allocation based on real-time data.
-
Performance Analytics for Job Execution: Leverage the metrics provided by Nomad to analyze job execution times and resource consumption. This use case enables developers to adjust job parameters effectively, optimize task performance, and illustrate trends over time, ultimately leading to increased efficiency and reduced costs in resource allocation.
-
Alerting on Critical Conditions: Implement alerting mechanisms based on metrics scraped from Nomad agents. By setting thresholds for critical metrics like CPU usage or failed job allocations, teams can proactively respond to potential issues before they escalate, ensuring higher uptime and reliability for applications running on the Nomad platform.
-
Integration with Visualization Tools: Use the data collected by the Hashicorp Nomad plugin to feed into visualization tools for real-time dashboards. This setup allows teams to monitor cluster workloads, job states, and system performance at a glance, facilitating better decision-making and strategic planning based on visual insights into the Nomad environment.
Prometheus
-
Monitoring Multi-cloud Deployments: Utilize the Prometheus plugin to collect metrics from applications running across multiple cloud providers. This scenario allows teams to centralize monitoring through a single Prometheus instance that scrapes metrics from different environments, providing a unified view of performance metrics across hybrid infrastructures. It streamlines reporting and alerting, enhancing operational efficiency without needing complex integrations.
-
Enhancing Microservices Visibility: Implement the plugin to expose metrics from various microservices within a Kubernetes cluster. Using Prometheus, teams can visualize service metrics in real time, identify bottlenecks, and maintain system health checks. This setup supports adaptive scaling and resource utilization optimization based on insights generated from the collected metrics. It enhances the ability to troubleshoot service interactions, significantly improving the resilience of the microservice architecture.
-
Real-time Anomaly Detection in E-commerce: By leveraging this plugin alongside Prometheus, an e-commerce platform can monitor key performance indicators such as response times and error rates. Integrating anomaly detection algorithms with scraped metrics allows the identification of unexpected patterns indicating potential issues, such as sudden traffic spikes or backend service failure. This proactive monitoring empowers business continuity and operational efficiency, minimizing potential downtimes while ensuring service reliability.
-
Performance Metrics Reporting for APIs: Utilize the Prometheus Output Plugin to gather and report API performance metrics, which can then be visualized in Grafana dashboards. This use case enables detailed analysis of API response times, throughput, and error rates, promoting continuous improvement of API services. By closely monitoring these metrics, teams can quickly react to degradation, ensuring optimal API performance and maintaining a high level of service availability.
Feedback
Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.
Powerful Performance, Limitless Scale
Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.
See Ways to Get Started
Related Integrations
Related Integrations
HTTP and InfluxDB Integration
The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.
View IntegrationKafka and InfluxDB Integration
This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.
View IntegrationKinesis and InfluxDB Integration
The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.
View Integration