Skip to content

HiveMQ Data Hub

IoT Stream Governance

Integrated policy and data transformation engine that validates, enforces, and manipulates data in motion to ensure data integrity and quality across your MQTT deployment.

Business Impact of Poor Data Quality

$1M Revenue Loss

Organizations lose millions annually due to poor data quality (Gartner)

3% Lower Revenue

Bad data costs enterprises 30% of their revenues (Ovum Research)

5% Data Mistrust

Over two-thirds of C-suite execs don’t trust the data coming from their systems (Deloitte)

Why IoT Stream Governance?

With the rapid growth of IoT and real-time data, the need for trusted, high-quality data across systems is becoming essential. While tools for managing IoT connections and ensuring basic data flow exist, guaranteeing that data is accurate, reliable, and actionable across teams has been a persistent challenge. This is especially true for businesses deploying distributed IoT ecosystems supported by diverse devices and complex infrastructures.

HiveMQ Data Hub Key Benefits

Maximize the business value of data being transported by defining data policies, transformations, and validation requirements to ensure the data is accurate and meets the standards your organization requires.

Read Data Sheet
Faster Business Insights with HiveMQ MQTT Platform

Faster Business Insights

Provide faster business insights on validated data - stop acting on rogue data and generating more noise than insights.

Increased Data Quality with HiveMQ MQTT Platform

Increased Data Quality

Ensure data quality standards are centrally defined and enforced across all devices and messages.

Operational Efficiency with HiveMQ MQTT Platform

Operational Efficiency

Stop bad-acting devices from misusing MQTT connections, sending bad data, and monopolizing resources.

/sb-assets/f/243938/134x134/905daaa987/cost-savings.png

Reduced Costs

Reduce redundant processing and storage costs by only acting on good data. Reduce the impact of acting on bad data.

Improve Data Integrity with HiveMQ MQTT Platform

Improve Data Integrity

Quarantine and further investigate bad data to prevent it from contaminating your systems and ultimately ensuring your data is accurate and reliable.

Seamless Data Management with HiveMQ MQTT Platform

Data Management

Manage everything in a single system, allowing data to be processed faster and negating the need to manage another standalone system to ensure data quality.

HiveMQ Data Hub allows us to automatically validate and standardize data to build a more reliable data pipeline and further grow our business

Klaas Jan Koopman

/

Director, Liberty Global

Read Case Study

HiveMQ Data Hub Architecture

HiveMQ Data Hub provides mechanisms to define how MQTT data is handled in the HiveMQ broker. This ensures that data quality is assessed at an early stage in the data supply chain, eliminating the need for subscribers to perform resource-intense validation before data reaches downstream devices or upstream services.

Key Data Hub Components

/sb-assets/f/243938/150x150/77d8751de5/operability.svg

Policy Engine

Ensure high data integrity by setting and enforcing rules that govern how the entire system operates, controlling what data is allowed in and what is filtered out.

AI Training

Transformers

Guarantee trusted data by applying transformation rules that standardize, clean, and enrich data, ensuring it meets required standards and structure.

Number of Plants

Industry Modules

Faster time to solution using industry specific no-code modules like SparkPlug and incorporate business-specific modules without complex configurations.

HiveMQ Data Hub Capabilities

The HiveMQ Data Hub provides the following capabilities to help you enforce data integrity and quality.

Enforce Behavior Policies

Define behavior policies to manage device interactions with the MQTT broker, including logging, stopping, or transforming actions. Flow-control validates message flow patterns, while the scripting engine enables custom behaviors using JavaScript functions.

Transform Data

Convert or manipulate data formats as it moves through the MQTT broker before reaching consumers. These data transformations enable more processing at the edge, streamlining data integration. For example, convert Fahrenheit to Celsius.

Define Data Policies

Set structured blueprints for data formats and enforce specific rules for incoming messages. Support for JSON and Protobuf formats ensures data integrity, with options to specify requirements, like temperature readings within defined ranges.

Simplify Workflow via Modules

Use specialized Data Hub modules that simplify data transformation and policy enforcement without coding. These modules bundle schemas, scripts, and policies into ready-to-use solutions for real-time automation.

Visualize Data

Utilize the simple user interface to manage schemas, data, and behavioral policies. The dashboard provides an overview of overall quality metrics, making it easy to locate bad actors and bad data sources. Visualize the MQTT data further in tools like Grafana.

Interact with RestAPI

Use HiveMQ’s REST API for programmatic interactions with the HiveMQ Enterprise MQTT broker, enabling seamless integration with other applications. The API supports sending and receiving data as JSON objects.

Data Schema

Create the schema of how data should be formatted. Both JSON and Protobuf formats are currently supported. These can vary from simple to complex. The example on the right is a schema that validates GPS coordinates.

{
  "$id": "https://example.com/geographical-location.schema.json",
  "$schema": "https://json-schema.org/draft/2020-12/schema",
  "title": "Longitude and Latitude Values",
  "description": "A geographical coordinate.",
  "required": ["latitude", "longitude"],
  "type": "object",
  "properties": {
    "latitude": {
      "type": "number",
      "minimum": -90,
      "maximum": 90
    },
    "longitude": {
      "type": "number",
      "minimum": -180,
      "maximum": 180
    }
  }
}
policy-gps.json

Data Policy

An appropriate policy tells HiveMQ how to handle incoming MQTT messages to enforce the rules and structure that the schema outlines. The example on the right drops the message and logs the result, but these can be arbitrarily complex and could re-queue the message. On the re-queue, the system could also apply transformation functions to fix bad data.

{
  "id": "com.hivemq.policy.coordinates",
  "matching": {
    "topicFilter": "coordinates"
  },
  "validation": {
    "validators": [
      {
        "type": "schema",
        "arguments": {
          "strategy": "ALL_OF",
          "schemas": [
            {
              "schemaId": "gps_coordinates",
              "version": "latest"
            }
          ]
        }
      }
    ]
  },
  "onFailure": {
    "pipeline": [
      {
        "id": "logFailure",
        "functionId": "log",
        "arguments": {
          "level": "WARN",
          "message": "${clientId} sent invalid coordinates on topic '${topic}' with result '${validationResult}'"
        }
      }
    ]
  }
}
workflow.json
Data Hub Modules screenshot

Modules

A look inside the HiveMQ Control Center which hosts a library of modules which can help quickly implement functionality like fanning out Sparkplug metrics, or dropping duplicate messages. The modules also helps easily transform data without requiring custom code or complex configuration. This is a growing collection with the potential of customers adding their own modules.

Grafana Dashboard to Visualize MQTT Data

Visualization

A demo use case introduces a quality metric and visualizes it in a Grafana dashboard. An addition to the quality metric is the list of bad clients queried from a PostgreSQL database. A screenshot of the dashboard includes the data quality on the left hand side and list of the top 10 bad clients on the right hand side.

Data Hub More Reading

HiveMQ at Hannover Messe 2025: Unlock the Value of Your Industrial Data

Join HiveMQ at Hannover Messe 2025! Discover how MQTT, UNS, and Distributed Data Intelligence unlock the full value of your industrial data in real time.

Blog

Improve IT-OT Collaboration with HiveMQ’s Custom Modules for Data Hub: Part 1

Bridge IT-OT gaps and optimize industrial data with HiveMQ’s Custom Modules for Data Hub—streamline transformations and reduce operational overhead.

Blog

Solving Real-Time Smart Meter Data Inconsistencies via HiveMQ Data Hub

Discover how HiveMQ Data Hub resolves smart meter data inconsistencies, enabling seamless integration, real-time processing, and operational efficiency in the energy sector.

Blog

HiveMQ Data Sheet - Data Hub

Download the HiveMQ Data Hub data sheet and learn how to enhance the value of your IoT data.

Resource

Overcoming MQTT Sparkplug Challenges for Smarter Manufacturing

Explore how HiveMQ's Sparkplug Module removes data bottlenecks, enforces compliance, & streamlines real-time IIoT data handling for a robust data pipeline.

Blog

Leveraging Open Standards like MQTT to Manage Data at the Industrial Edge

Explore why use open standards, like MQTT, and specifications, like Sparkplug, at the Industrial Edge to manage data.

Blog

Enhancing Axis Network Camera Capabilities with MQTT and HiveMQ

Discover how Axis network cameras & HiveMQ MQTT Broker enable real-time data processing & edge intelligence for revolutionary network camera capabilities.

Blog

Sending Filtered MQTT Sparkplug Messages to MongoDB for IIoT Data Transformation: Part 2

Learn how to filter unnecessary Sparkplug metrics using a data policy method, and send the refined data to MongoDB via HiveMQ MongoDB Extension.

Blog

Sending Filtered MQTT Sparkplug Messages to MongoDB for IIoT Data Transformation: Part 1

Learn how to transform Sparkplug data to JSON, optimize IIoT data processing, & integrate OT with IT using HiveMQ Data Hub and MongoDB Extension.

Blog

HiveMQ Data Hub Frequently Asked Questions

Recommended Next Steps

Schedule Demo

Contact Us

Get HiveMQ

Free Download

Read the Docs

Documentation
HiveMQ logo
Review HiveMQ on G2