Rediger

Del via


Process data with data flow graphs

Important

This page includes instructions for managing Azure IoT Operations components using Kubernetes deployment manifests, which is in PREVIEW. This feature is provided with several limitations, and shouldn't be used for production workloads.

See the Supplemental Terms of Use for Microsoft Azure Previews for legal terms that apply to Azure features that are in beta, preview, or otherwise not yet released into general availability.

A data flow graph is a composable processing pipeline that transforms data as it moves between sources and destinations. A standard data flow follows a fixed enrich, filter, map sequence. A data flow graph lets you chain transforms in any order, branch into parallel paths, and aggregate data over time windows.

This article walks through creating a data flow graph step by step. For an overview of data flow graphs and the available transforms, see Data flow graphs overview.

Important

Data flow graphs currently support only MQTT, Kafka, and OpenTelemetry endpoints. Other endpoint types like Data Lake, Microsoft Fabric OneLake, Azure Data Explorer, and Local Storage aren't supported. For more information, see Known issues.

Prerequisites

Create a data flow graph

A data flow graph contains three types of elements: sources that bring data in, transforms that process it, and destinations that send it out. Connect them in the order you want data to flow.

  1. In the Operations experience, go to your Azure IoT Operations instance.

  2. Select Data flow graph > Create data flow graph.

    Screenshot of operations experience showing data flow graph.

  3. Enter a name for the data flow graph and select a data flow profile. The default profile is selected by default.

    Screenshot of the operations experience create dialog showing the name field and profile dropdown.

  4. Build your pipeline by adding elements to the canvas:

    1. Add a source: Select the source endpoint and configure the topics to subscribe to for incoming messages.

      Screenshot of the operations experience source configuration panel showing endpoint dropdown and topic input.

      1. Add transforms: Select one or more transforms to process the data. Available transforms include map, filter, branch, concat, and window. For details on each transform type, see Data flow graphs overview.

      Screenshot of the operations experience transform selection menu showing available transform types.

      Screenshot of the operations experience showing a branch transform configuration example.Screenshot of the operations experience showing a branch transform configuration example.

    2. Add a destination: Select the destination endpoint and configure the topic or path to send processed data to.

      Screenshot of the operations experience showing a destination example.

  5. Connect the elements in the order you want data to flow.

    Screenshot of the operations experience canvas showing a connected source, transform, and destination pipeline.

  6. Select Save to deploy the data flow graph.

Configure the source

The source defines where data enters the pipeline. Specify an endpoint reference and one or more topics.

In the data flow graph editor, select the source element and configure:

Setting Description
Endpoint The data flow endpoint to use. Select default for the local MQTT broker.
Topics One or more topics to subscribe to for incoming messages.

Add transforms

Transforms process data between the source and destination. Each transform references a built-in artifact and is configured with rules.

The available built-in transforms are:

Transform Artifact Description
Map azureiotoperations/graph-dataflow-map:1.0.0 Rename, restructure, compute, and copy fields
Filter azureiotoperations/graph-dataflow-filter:1.0.0 Drop messages that match a condition
Branch azureiotoperations/graph-dataflow-branch:1.0.0 Route messages to a true or false path
Concat azureiotoperations/graph-dataflow-concatenate:1.0.0 Merge branched paths back together
Window azureiotoperations/graph-dataflow-window:1.0.0 Aggregate data over a time interval

For detailed configuration of each transform type, see:

In the data flow graph editor, select Add transform and choose the transform type. Configure the rules in the visual editor.

Chain multiple transforms

You can chain any number of transforms. Connect them in the nodeConnections section in the order you want data to flow:

Drag connections between transforms on the canvas to define the processing order.

Configure the destination

The destination defines where processed data is sent. Specify an endpoint reference and a topic or path.

Select the destination element and configure:

Setting Description
Endpoint The data flow endpoint to send data to.
Topic The topic or path to publish processed data to.

For dynamic topic routing based on message content, see Route messages to different topics.

Verify the data flow graph is working

After you deploy a data flow graph, verify it's running:

In the Operations experience, select your data flow graph to view its status. A healthy graph shows a Running state.

Next steps