ActiveMQ and Cortex Integration

Powerful performance with an easy integration, powered by Telegraf, the open source data connector built by InfluxData.

info

This is not the recommended configuration for real-time query at scale. For query and compression optimization, high-speed ingest, and high availability, you may want to consider ActiveMQ and InfluxDB.

5B+

Telegraf downloads

#1

Time series database
Source: DB Engines

1B+

Downloads of InfluxDB

2,800+

Contributors

Table of Contents

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Input and output integration overview

The ActiveMQ Input Plugin collects metrics from the ActiveMQ message broker through its Console API, providing insights into the performance and status of message queues, topics, and subscribers.

This plugin enables Telegraf to send metrics to Cortex using the Prometheus remote write protocol, allowing seamless ingestion into Cortex’s scalable, multi-tenant time series storage.

Integration details

ActiveMQ

The ActiveMQ Input Plugin interfaces with the ActiveMQ Console API to gather metrics related to queues, topics, and subscribers. ActiveMQ, a widely-used open-source message broker, supports various messaging protocols and provides a robust Web Console for management and monitoring. This plugin allows users to track essential metrics including queue sizes, consumer counts, and message counts across different ActiveMQ entities, thereby enhancing observability within messaging systems. Users can configure various parameters such as the WebConsole URL and basic authentication credentials to tailor the plugin to their environment. The metrics collected can be used for monitoring the health and performance of messaging applications, facilitating proactive management and troubleshooting.

Cortex

With Telegraf’s HTTP output plugin and the prometheusremotewrite data format you can send metrics directly to Cortex, a horizontally scalable, long-term storage backend for Prometheus. Cortex supports multi-tenancy and accepts remote write requests using the Prometheus protobuf format. By using Telegraf as the collection agent and Remote Write as the transport mechanism, organizations can extend observability into sources not natively supported by Prometheus—such as Windows hosts, SNMP-enabled devices, or custom application metrics—while leveraging Cortex’s high-availability and long-retention capabilities.

Configuration

ActiveMQ

[[inputs.activemq]]
  ## ActiveMQ WebConsole URL
  url = "http://127.0.0.1:8161"

  ## Required ActiveMQ Endpoint
  ##   deprecated in 1.11; use the url option
  # server = "192.168.50.10"
  # port = 8161

  ## Credentials for basic HTTP authentication
  # username = "admin"
  # password = "admin"

  ## Required ActiveMQ webadmin root path
  # webadmin = "admin"

  ## Maximum time to receive response.
  # response_timeout = "5s"

  ## Optional TLS Config
  # tls_ca = "/etc/telegraf/ca.pem"
  # tls_cert = "/etc/telegraf/cert.pem"
  # tls_key = "/etc/telegraf/key.pem"
  ## Use TLS but skip chain & host verification
  # insecure_skip_verify = false

Cortex

[[outputs.http]]
  ## Cortex Remote Write endpoint
  url = "http://cortex.example.com/api/v1/push"

  ## Use POST to send data
  method = "POST"

  ## Send metrics using Prometheus remote write format
  data_format = "prometheusremotewrite"

  ## Optional HTTP headers for authentication
  # [outputs.http.headers]
  #   X-Scope-OrgID = "your-tenant-id"
  #   Authorization = "Bearer YOUR_API_TOKEN"

  ## Optional TLS configuration
  # tls_ca = "/path/to/ca.pem"
  # tls_cert = "/path/to/cert.pem"
  # tls_key = "/path/to/key.pem"
  # insecure_skip_verify = false

  ## Request timeout
  timeout = "10s"

Input and output integration examples

ActiveMQ

  1. Proactive Queue Monitoring: Use the ActiveMQ plugin to monitor queue sizes in real-time for a high-volume trading application. This implementation allows teams to receive alerts when queue sizes exceed a certain threshold, enabling rapid response to potential downtime caused by backlogs, thereby ensuring continuous availability of trading operations.

  2. Performance Baselines and Anomaly Detection: Integrate this plugin with machine learning frameworks to establish performance baselines for message throughput. By analyzing historical data collected through this plugin, teams can flag anomalies in processing rates, leading to quicker identification of issues impacting service reliability and performance.

  3. Cross-Messaging System Analytics: Combine metrics from ActiveMQ with those from other messaging systems in a centralized dashboard. Users can visualize and compare performance data, such as enqueue and dequeue rates, providing valuable insights into the overall messaging architecture and assisting in optimizing the message flow between different brokers.

  4. Subscriber Performance Insights: Leverage the subscriber metrics collected by this plugin to analyze behavior patterns and optimize configuration for consumer applications. Understanding metrics such as dispatched queue size and counter values can guide adjustments to improve processing efficiency and resource allocation.

Cortex

  1. Unified Multi-Tenant Monitoring: Use Telegraf to collect metrics from different teams or environments and push them to Cortex with separate X-Scope-OrgID headers. This enables isolated data ingestion and querying per tenant, ideal for managed services and platform teams.

  2. Extending Prometheus Coverage to Edge Devices: Deploy Telegraf on edge or IoT devices to collect system metrics and send them to a centralized Cortex cluster. This approach ensures consistent observability even for environments without local Prometheus scrapers.

  3. Global Service Observability with Federated Tenants: Aggregate metrics from global infrastructure by configuring Telegraf agents to push data into regional Cortex clusters, each tagged with tenant identifiers. Cortex handles deduplication and centralized access across regions.

  4. Custom App Telemetry Pipeline: Collect app-specific telemetry via Telegraf’s exec or http input plugins and forward it to Cortex. This allows DevOps teams to monitor app-specific KPIs in a scalable, query-efficient format while keeping metrics logically grouped by tenant or service.

Feedback

Thank you for being part of our community! If you have any general feedback or found any bugs on these pages, we welcome and encourage your input. Please submit your feedback in the InfluxDB community Slack.

Powerful Performance, Limitless Scale

Collect, organize, and act on massive volumes of high-velocity data. Any data is more valuable when you think of it as time series data. with InfluxDB, the #1 time series platform built to scale with Telegraf.

See Ways to Get Started

Related Integrations

HTTP and InfluxDB Integration

The HTTP plugin collects metrics from one or more HTTP(S) endpoints. It supports various authentication methods and configuration options for data formats.

View Integration

Kafka and InfluxDB Integration

This plugin reads messages from Kafka and allows the creation of metrics based on those messages. It supports various configurations including different Kafka settings and message processing options.

View Integration

Kinesis and InfluxDB Integration

The Kinesis plugin allows for reading metrics from AWS Kinesis streams. It supports multiple input data formats and offers checkpointing features with DynamoDB for reliable message processing.

View Integration