Bemærk
Adgang til denne side kræver godkendelse. Du kan prøve at logge på eller ændre mapper.
Adgang til denne side kræver godkendelse. Du kan prøve at ændre mapper.
Important
This page includes instructions for managing Azure IoT Operations components using Kubernetes deployment manifests, which is in PREVIEW. This feature is provided with several limitations, and shouldn't be used for production workloads.
See the Supplemental Terms of Use for Microsoft Azure Previews for legal terms that apply to Azure features that are in beta, preview, or otherwise not yet released into general availability.
A data flow graph is a composable processing pipeline that transforms data as it moves between sources and destinations. A standard data flow follows a fixed enrich, filter, map sequence. A data flow graph lets you chain transforms in any order, branch into parallel paths, and aggregate data over time windows.
This article walks through creating a data flow graph step by step. For an overview of data flow graphs and the available transforms, see Data flow graphs overview.
Important
Data flow graphs currently support only MQTT, Kafka, and OpenTelemetry endpoints. Other endpoint types like Data Lake, Microsoft Fabric OneLake, Azure Data Explorer, and Local Storage aren't supported. For more information, see Known issues.
Prerequisites
- An instance of Azure IoT Operations, version 1.2 or later.
- A data flow profile. You can use the default profile.
- A data flow endpoint for your source and destination. The default MQTT broker endpoint works for getting started.
Create a data flow graph
A data flow graph contains three types of elements: sources that bring data in, transforms that process it, and destinations that send it out. Connect them in the order you want data to flow.
In the Operations experience, go to your Azure IoT Operations instance.
Select Data flow graph > Create data flow graph.
Enter a name for the data flow graph and select a data flow profile. The default profile is selected by default.
Build your pipeline by adding elements to the canvas:
Add a source: Select the source endpoint and configure the topics to subscribe to for incoming messages.
- Add transforms: Select one or more transforms to process the data. Available transforms include map, filter, branch, concat, and window. For details on each transform type, see Data flow graphs overview.
Add a destination: Select the destination endpoint and configure the topic or path to send processed data to.
Connect the elements in the order you want data to flow.
Select Save to deploy the data flow graph.
Configure the source
The source defines where data enters the pipeline. Specify an endpoint reference and one or more topics.
In the data flow graph editor, select the source element and configure:
| Setting | Description |
|---|---|
| Endpoint | The data flow endpoint to use. Select default for the local MQTT broker. |
| Topics | One or more topics to subscribe to for incoming messages. |
Add transforms
Transforms process data between the source and destination. Each transform references a built-in artifact and is configured with rules.
The available built-in transforms are:
| Transform | Artifact | Description |
|---|---|---|
| Map | azureiotoperations/graph-dataflow-map:1.0.0 |
Rename, restructure, compute, and copy fields |
| Filter | azureiotoperations/graph-dataflow-filter:1.0.0 |
Drop messages that match a condition |
| Branch | azureiotoperations/graph-dataflow-branch:1.0.0 |
Route messages to a true or false path |
| Concat | azureiotoperations/graph-dataflow-concatenate:1.0.0 |
Merge branched paths back together |
| Window | azureiotoperations/graph-dataflow-window:1.0.0 |
Aggregate data over a time interval |
For detailed configuration of each transform type, see:
In the data flow graph editor, select Add transform and choose the transform type. Configure the rules in the visual editor.
Chain multiple transforms
You can chain any number of transforms. Connect them in the nodeConnections section in the order you want data to flow:
Drag connections between transforms on the canvas to define the processing order.
Configure the destination
The destination defines where processed data is sent. Specify an endpoint reference and a topic or path.
Select the destination element and configure:
| Setting | Description |
|---|---|
| Endpoint | The data flow endpoint to send data to. |
| Topic | The topic or path to publish processed data to. |
For dynamic topic routing based on message content, see Route messages to different topics.
Verify the data flow graph is working
After you deploy a data flow graph, verify it's running:
In the Operations experience, select your data flow graph to view its status. A healthy graph shows a Running state.