Edit

Share via


Known issues in Azure Local

This article identifies critical known issues and their workarounds in Azure Local.

These release notes are continuously updated, and as critical issues requiring a workaround are discovered, they're added. Before you deploy your Azure Local instance, carefully review the information contained here.

Important

For information about supported update paths for this release, see Release information.

For more information about new features in this release, see What's new for Azure Local.

Known issues for version 2602

For the 2602 release of Azure Local, Microsoft released the following update:

Solution version OS build
12.2602.1002.501 26100.32370

Important

The new deployments of this software use the 12.2602.1002.501 build. Release notes for this version include the issues fixed in this release, known issues in this release, and known issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Comments
Deployment Fixed issue in the Azure portal where the Azure Local deployment wizard wasn't loading.
Deployment Fixed issue in Azure portal to block proceeding through Azure Local deployment wizard when inputs aren't specified.
Deployment Added validation to instance and machine names in Azure Local deployment wizard in Azure portal.
Update Fixed issue with SBE update failing due to invalid argument.
Azure Local VMs Fixed issue where VM start or stop operation was blocked with an error after attaching or detaching GPU to the VM.

Known issues

There's no known issue in this release. Any previously known issues have been fixed in subsequent releases.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment In rare instances, deployment fails with errors during validation that state that the mandatory Arc extensions are not yet installed. If you face this issue, retry the deployment.
Security If the Windows Defender attack surface reduction rule Block Process Creations originating from PSExec & WMI commands is configured to Block, the Azure Local Solution Update will fail to run. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Add server, Repair server Add node and repair node operations fail when running on 11.2510.1002.87 or 12.2510.1002.88, as these images were recalled and don't exist. Upgrade your environment to 11.2510.1002.93 or 12.2510.1002.94.

If you need to run add node or repair node operations during the update from 1.2510.1002.87/12.2510.1002.88 to 11.2510.1002.93/12.2510.1002.94, open a support case to overwrite the image validation.
Azure Local VMs VM start, stop, or delete operations may fail due to the wssdagent node agent crashing. To check if wssdagent has crashed, run the following command:

$ServerList = (Get-Clusternode).name
foreach ($Server in $ServerList) {
Write-Output "Cluster Node: $Server..."
Invoke-Command -ComputerName $Server -ScriptBlock {
get-service wssdagent
}
}

If the wssdagent status shows "Stopped", run the following command to restart the agent from that node:

start-service wssdagent

This should get the node agent running again and unblock the VMs. If any VMs are deleted while the node agent is down, open a support case to get the issue resolved.
Update Fetching the secret rotation action plan status fails. The secret rotation completes successfully, so the failure message can be ignored.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2601

For the 2601 release of Azure Local, Microsoft released the following update:

Solution version OS build
12.2601.1002.503 26100.32230

Important

The new deployments of this software use the 12.2601.1002.503 build. Release notes for this version include the issues fixed in this release, known issues in this release, and known issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Comments
Update Update may fail when the cloud management group is running on a different node than the owner node with the error: Type 'RegisterCloudManagementUpdatesExtension' of Role 'CloudManagementConfig' raised an exception: Exception occurred in Get-ClusterExtension'
Update Secret rotation changes WinRM certificate thumbprint for users who aren't using a Microsoft issued certificate.
Azure Local VMs When creating or deleting large numbers of network interfaces, the infrastructure can face out of memory issues causing failures. Improved validation logic to handle large numbers of operations on network interfaces.
Azure Local VMs Added operation to set the cluster functional level for upgrade to 24H2.
Deployment Deployment using Local Identity doesn't support manual secret rotation.
Deployment Enable SMB protocol for all cluster nodes during deployment.
Deployment Added step to make sure auto-mount is enabled.
Deployment Added check for null values before resuming failed deployment.
Deployment Improved validation to prevent upgrading an already deployed instance.
Deployment Updated validation for cluster resource name.
Deployment Improved readability of physical disk test output.
Deployment Improved validation performed by DNS resolution test.
Deployment Added retry to help cluster resource come online when rerunning deployment.
Deployment Implemented maximum length check for cluster name.
Deployment Removed unused step to add DSC certificate from deployment.
Deployment Updated logic to fetch the cluster IP.
Deployment Fixed issue with stuck deployment after domain join step.
Deployment Added a wait step to ensure cluster DNS names are successfully resolved.
Deployment When using 3 or more intents in the system, PowerShell Invoke-Command call crashes with the error: Type 'SetAzureStackHostsPreConfiguration' of Role 'HostNetwork' raised an exception: Cannot get intent status of Storage. Exception: No intent statuses found for intent Storage.
Update Added fix for failed updates due to SBE helper module.
Update Fixed issue with downloading SBE during update.
Update Added cleanup for CAU reports during SBE update.
Update Reduced duration of SBE steps for deployment and update.
Update Improved health check results for SBE test.
Update Fixed issue where SBE files may be blocked after performing robocopy to each machine.
Update Fixed issue with solution update failing due to incorrectly detecting presence of SBE.
Update SBE update fails with the error: Cannot validate argument on parameter 'DeployADLess'. The argument "[DEPLOYADLESS]" does not belong to the set "true,false," specified by the ValidateSet attribute.
Update Update fails due to CLI extension connectedmachine exception causing deployment disruption.
Update Solution Builder Extension (SBE) download fails with the error:
CloudEngine.Actions.InterfaceInvocationFailedException: Type 'SBEPartnerDownloadConnectorCheckHealth' of Role 'SBE' raised an exception:

[SBEPartnerDownloadConnectorCheckHealth] SBE download connector failure getting version '0.0.0000.0000' files. The download connector health check failed. Exception Message : The term 'Assert-SBEResponseSchema' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again.
Upgrade Improved logic to find storage cluster group.
Upgrade When upgrading from 2510 to 2511, 2512, or 2601, AKS Arc cluster creation fails.

Known issues

The following table lists the known issues in this release:

Feature Issue Workaround
Update Fetching the secret rotation action plan status fails. The secret rotation completes successfully, so the failure message can be ignored.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment In rare instances, deployment fails with errors during validation that state that the mandatory Arc extensions are not yet installed. If you face this issue, retry the deployment.
Security If the Windows Defender attack surface reduction rule Block Process Creations originating from PSExec & WMI commands is configured to Block, the Azure Local Solution Update will fail to run. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Add server, Repair server Add node and repair node operations fail when running on 11.2510.1002.87 or 12.2510.1002.88, as these images were recalled and don't exist. Upgrade your environment to 11.2510.1002.93 or 12.2510.1002.94.

If you need to run add node or repair node operations during the update from 1.2510.1002.87/12.2510.1002.88 to 11.2510.1002.93/12.2510.1002.94, open a support case to overwrite the image validation.
Azure Local VMs VM start, stop, or delete operations may fail due to the wssdagent node agent crashing. To check if wssdagent has crashed, run the following command:

$ServerList = (Get-Clusternode).name
foreach ($Server in $ServerList) {
Write-Output "Cluster Node: $Server..."
Invoke-Command -ComputerName $Server -ScriptBlock {
get-service wssdagent
}
}

If the wssdagent status shows "Stopped", run the following command to restart the agent from that node:

start-service wssdagent

This should get the node agent running again and unblock the VMs. If any VMs are deleted while the node agent is down, open a support case to get the issue resolved.
Azure Local VMs VM start or stop operation is blocked with an error after attaching or detaching GPU to the VM. Update the size of the VM (add/remove vCPU, memory) before attempting a start or stop operation. Once the size is updated, proceed with VM start or stop.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.

Known issues for version 2512

For the 2512 release of Azure Local, Microsoft released the following update:

Solution version OS build
12.2512.1002.16 26100.7462

Important

The new deployments of this software use the 12.2512.1002.16 build. Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Comments
Azure Local VMs Azure Local VM create flow allows adding a data disk with a name that already exists in the same resource group, leading to provisioning errors. Validation to ensure the data disk name is unique within the resource group before creating the Azure Local VM.

Known issues

The following table lists the known issues in this release:

Feature Issue Workaround
Deployment Deployment using Local Identity doesn't support manual secret rotation. There's no known workaround in this release.
Deployment When upgrading from 2510 to 2511, 2512, or 2601, cluster creation fails due to CSI provisioning. For detailed steps on how to resolve this issue, see AKS Arc cluster creation fails on Azure Local 2511, 2512, or 2601 after upgrade from 2510.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment After Azure portal deployment, SConfig network settings shows the error: Set-SCfNetworksetting : Cannot bind argument to parameter 'Value' because it is null. There's no known workaround in this release.
Security If the Windows Defender attack surface reduction rule Block Process Creations originating from PSExec & WMI commands is configured to Block, the Azure Local Solution Update will fail to run. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Add server, Repair server Add node and repair node operations fail when running on 11.2510.1002.87 or 12.2510.1002.88, as these images were recalled and don't exist. Upgrade your environment to 11.2510.1002.93 or 12.2510.1002.531.

If you need to run add node or repair node operations during the update from 1.2510.1002.87/12.2510.1002.88 to 11.2510.1002.93/12.2510.1002.531, open a support case to overwrite the image validation.
Azure Local VMs VM start, stop, or delete operations may fail due to the wssdagent node agent crashing. To check if wssdagent has crashed, run the following command:

$ServerList = (Get-Clusternode).name
foreach ($Server in $ServerList) {
Write-Output "Cluster Node: $Server..."
Invoke-Command -ComputerName $Server -ScriptBlock {
get-service wssdagent
}
}

If the wssdagent status shows "Stopped", run the following command to restart the agent from that node:

start-service wssdagent

This should get the node agent running again and unblock the VMs. If any VMs are deleted while the node agent is down, open a support case to get the issue resolved.
Update Update may fail when the cloud management group is running on a different node than the owner node with the error: Type 'RegisterCloudManagementUpdatesExtension' of Role 'CloudManagementConfig' raised an exception: Exception occurred in Get-ClusterExtension' Move the management group to the owner node manually and proceed with the update:

# Get the owner node of the group matching '*orch*'
$orchOwner = (Get-ClusterGroup -Name '*orch*').OwnerNode.Name

# Move the "Cloud Management" group to that node
Move-ClusterGroup -Name 'Cloud Management' -Node $orchOwner
Azure Local VMs VM start or stop operation is blocked with an error after attaching or detaching GPU to the VM. Update the size of the VM (add/remove vCPU, memory) before attempting a start or stop operation. Once the size is updated, proceed with VM start or stop.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2511

For the 2511 release of Azure Local, Microsoft released two security updates, each aligned with a specific OS build. The following table provides the specific versions and their OS builds:

Solution version OS build
12.2511.1002.502 26100.7171

Important

The new deployments of this software use the 12.2511.1002.502 build. Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Comments
Deployment Updated logic to reduce dependence on DNS servers during cluster DNS name resolution.
Deployment Fixed issue where deployment is stuck post domain join by resuming the deployment.
Deployment Add a wait step to ensure the cluster DNS names are resolved.
Deployment Remove add DSC certificate step since it's not required.
Azure Local VMs Added operation to set the cluster functional level for upgrade to 24H2.
Update When installing cumulative updates using Azure Update Manager, only the latest update for version 2507 is installed. If earlier update options (for versions 2505, 2506) are selected, they are not installed.
Update Update health check results not shown when it's been longer than 3 hours after health check completed
Update When updating the Azure Local instance via the Azure Update Manager, the update progress and results may not be visible in the Azure portal.
Deployment Deployment, add node, and repair node operations may fail with the error:
Type 'EncryptClusterSharedVolumes' of Role 'AzureStackBitlocker' raised an exception: The job running on xxx failed due to: System.Management.Automation.RemoteException: -> Failed enabling bitlocker for C:\ClusterStorage\UserStorage_13 (F:)

Known issues

The following table lists the known issues in this release:

Feature Issue Workaround
Update Update may fail when the cloud management group is running on a different node than the owner node with the error: Type 'RegisterCloudManagementUpdatesExtension' of Role 'CloudManagementConfig' raised an exception: Exception occurred in Get-ClusterExtension' Move the management group to the owner node manually and proceed with the update:

# Get the owner node of the group matching '*orch*'
$orchOwner = (Get-ClusterGroup -Name '*orch*').OwnerNode.Name

# Move the "Cloud Management" group to that node
Move-ClusterGroup -Name 'Cloud Management' -Node $orchOwner
Deployment When upgrading from 2510 to 2511, 2512, or 2601, cluster creation fails due to CSI provisioning. For detailed steps on how to resolve this issue, see AKS Arc cluster creation fails on Azure Local 2511, 2512, or 2601 after upgrade from 2510.
Azure Local VMs VM start or stop operation is blocked with an error after attaching or detaching GPU to the VM. Update the size of the VM (add/remove vCPU, memory) before attempting a start or stop operation. Once the size is updated, proceed with VM start or stop.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment After Azure portal deployment, SConfig network settings shows the error: Set-SCfNetworksetting : Cannot bind argument to parameter 'Value' because it is null. There's no known workaround in this release.
Deployment In rare instances, deployment fails with errors during validation that state that the mandatory Arc extensions are not yet installed. If you face this issue, retry the deployment.
Security If the Windows Defender attack surface reduction rule Block Process Creations originating from PSExec & WMI commands is configured to Block, the Azure Local Solution Update will fail to run. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Add server, Repair server Add node and repair node operations fail when running on 11.2510.1002.87 or 12.2510.1002.88, as these images were recalled and don't exist. Upgrade your environment to 11.2510.1002.93 or 12.2510.1002.531.

If you need to run add node or repair node operations during the update from 1.2510.1002.87/12.2510.1002.88 to 11.2510.1002.93/12.2510.1002.531, open a support case to overwrite the image validation.
Azure Local VMs VM start, stop, or delete operations may fail due to the wssdagent node agent crashing. To check if wssdagent has crashed, run the following command:

$ServerList = (Get-Clusternode).name
foreach ($Server in $ServerList) {
Write-Output "Cluster Node: $Server..."
Invoke-Command -ComputerName $Server -ScriptBlock {
get-service wssdagent
}
}

If the wssdagent status shows "Stopped", run the following command to restart the agent from that node:

start-service wssdagent

This should get the node agent running again and unblock the VMs. If any VMs are deleted while the node agent is down, open a support case to get the issue resolved.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2510

For the 2510 release of Azure Local, Microsoft released two security updates, each aligned with a specific OS build. The following table provides the specific versions and their OS builds:

Solution version OS build
11.2510.1002.93 25398.1913
12.2510.1002.531 26100.6899

Important

The new deployments of this software use the 12.2510.1002.531 build. You can also update an existing deployment from 2509 by using 11.2510.1002.93.

Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Comments
Deployment Improved calculation of maximum volume size for express storage provisioning.
Deployment Adjusted allocation unit size for thin provisioned volumes.
Deployment Fixed issue with importing modules for storage.
Deployment Enabled Rack Level Nested Mirror for rack aware clusters.
Deployment Fixed issue with restarting performance history volume.
Deployment Added mitigation for Cluster DNS resolution issues.
Upgrade Refined the version check to get the correct latest updates.
Upgrade Improved check for local administrator.
Security When fixing the compliance for the minimum password length rule, even after you've changed the minimum password length on the Azure Local host to 14, you continue to see it as noncompliant in Azure policy.
Azure Local VMs Azure Local VMs have empty DVD drives attached after deployment. Newly created Azure Local VMs will no longer include empty DVD drives after deployment.
Azure Local VMs In some cases, Azure Local VMs fail to create with error: Failed to cleanup seed iso disk from the file system for vm. This build addresses additional corner cases that were not resolved in the previous fix (2503).
Azure Local VMs In rare cases, the operator would enter a crash loop, preventing Azure Local VM management.
Azure Local VMs In some cases, deleting logical networks could fail when network interfaces created with those logical networks existed in other resource groups or locations. Updated deletion logic to ensure safe deletion of logical network.
Update Update fails with VSR registry not found error.
Deployment, update Deployment and updates crash with a blue screen and error code 0x00000139.
Update Updates failing for deployments between 2504 and 2511.
Update After upgrade, the subsequent update preparation fails.
Deployment Deployment fails with error: The following error with errorcode 0x80090308 occurred while using Negotiate authentication: The parameter is incorrect.
Deployment Deployment fails with error: The parameter is incorrect.
Update OS upgrade doesn't perform Update-ClusterFunctionalLevel and Update-StoragePool after the OS upgrade completes.
Deployment Log collection doesn't start when it starts it the first time, and then shows 404 errors subsequent times. There's no known workaround for this issue in this release.

Known issues

The following table lists the known issues in this release:

Feature Issue Workaround
Add server, Repair server Add node and repair node operations fail when running on 11.2510.1002.87 or 12.2510.1002.88, as these images were recalled and don't exist. Upgrade your environment to 11.2510.1002.93 or 12.2510.1002.531.

If you need to run add node or repair node operations during the update from 1.2510.1002.87/12.2510.1002.88 to 11.2510.1002.93/12.2510.1002.531, open a support case to overwrite the image validation.
Azure Local VMs VM start, stop, or delete operations may fail due to the wssdagent node agent crashing. To check if wssdagent has crashed, run the following command:

$ServerList = (Get-Clusternode).name
foreach ($Server in $ServerList) {
Write-Output "Cluster Node: $Server..."
Invoke-Command -ComputerName $Server -ScriptBlock {
get-service wssdagent
}
}

If the wssdagent status shows "Stopped", run the following command to restart the agent from that node:

start-service wssdagent

This should get the node agent running again and unblock the VMs. If any VMs are deleted while the node agent is down, open a support case to get the issue resolved.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Update When updating the Azure Local instance via the Azure Update Manager, the update progress and results may not be visible in the Azure portal. To work around this issue, on each node, add the following registry key (no value needed):

New-Item -Path "HKLM:\SYSTEM\CurrentControlSet\Services\HciCloudManagementSvc\Parameters" -force

Then on one of the nodes, restart the Cloud Management cluster group.

Stop-ClusterGroup "Cloud Management"

Start-ClusterGroup "Cloud Management"

This won't fully remediate the issue as the progress details may still not be displayed for a duration of the update process. To get the latest update details, you can Retrieve the update progress with PowerShell.
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment After Azure portal deployment, SConfig network settings shows the error: Set-SCfNetworksetting : Cannot bind argument to parameter 'Value' because it is null. There's no known workaround in this release.
Deployment In rare instances, deployment fails with errors during validation that state that the mandatory Arc extensions are not yet installed. If you face this issue, retry the deployment.
Update When installing cumulative updates using Azure Update Manager, only the latest update for version 2507 is installed. If earlier update options (for versions 2505, 2506) are selected, they are not installed. There's no workaround in this release.
Update Update health check results not shown when it's been longer than 3 hours after health check completed Rerun the update health checks to restart the 3 hour expiration clock.
Deployment Deployment, add node, and repair node operations may fail with the error:
Type 'EncryptClusterSharedVolumes' of Role 'AzureStackBitlocker' raised an exception: The job running on xxx failed due to: System.Management.Automation.RemoteException: -> Failed enabling bitlocker for C:\ClusterStorage\UserStorage_13 (F:)
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Security If the Windows Defender attack surface reduction rule Block Process Creations originating from PSExec & WMI commands is configured to Block, the Azure Local Solution Update will fail to run. For detailed steps on how to resolve this issue, see the Troubleshooting guide.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2509

For the 2509 release of Azure Local, Microsoft released two security updates, each aligned with a specific OS build. The following table provides the specific versions and their OS builds:

Solution version OS build
11.2509.1001.21 25398.1849
12.2509.1001.22 26100.6584

Important

The new deployments of this software use the 12.2509.1001.22 build. You can also update an existing deployment from 2508 by using 11.2509.1001.21.

Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Comments
Azure Local VMs, Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. VMs are now activated properly.
Azure Local VMs Recreating an image with the same name could fail due to remnants of a previous failed deployment, occasionally resulting on the creating of a zero-byte image. Automatic proper cleanup of failed image artifacts, allowing subsequent image creation attempts to succeed.
Deployment Disk types aren't being determined correctly resulting in false negative results on storage pool size determination. Improved validation for resolving disk type when determining storage pool size.
Deployment True max volume size should be calculated and not defaulted to 64 TB. Improved calculation for maximum volume size.
Deployment Adjusted minimum allocation unit size for thin provisioned volumes.
Azure Local VMs Increased default disk size from 100 GB to 200 GB.
Azure Local VMs AKS Arc created cluster isn't working in locked environment with proxy enabled.
Azure portal Improved the associate NSG dropdown menu in the Azure portal.
Deployment In the 2506 release, KIR registries were added to the incorrect path. Added registries to the correct path.
Upgrade Fixed an issue where upgrade fails on environment validator.
Deployment Changed the default OS size for thin provisioned volumes.
Security Mitigation for security vulnerability CVE-2025-55316 was implemented.
Registration, deployment, Add-server, Upgrade, Update In this release, connectivity tests can take longer than 20 minutes in some environments. There's no known workaround for this issue in this release.

Known issues

The following table lists the known issues in this release:

Feature Issue Workaround
Deployment Log collection doesn't start when it starts it the first time, and then shows 404 errors subsequent times. There's no known workaround for this issue in this release.
Deployment Deployment, add node, and repair node operations may fail with the error:
Type 'EncryptClusterSharedVolumes' of Role 'AzureStackBitlocker' raised an exception: The job running on xxx failed due to: System.Management.Automation.RemoteException: -> Failed enabling bitlocker for C:\ClusterStorage\UserStorage_13 (F:)
For detailed steps on how to resolve this issue, see the Troubleshooting guide.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Update When updating the Azure Local instance via the Azure Update Manager, the update progress and results may not be visible in the Azure portal. To work around this issue, on each node, add the following registry key (no value needed):

New-Item -Path "HKLM:\SYSTEM\CurrentControlSet\Services\HciCloudManagementSvc\Parameters" -force

Then on one of the nodes, restart the Cloud Management cluster group.

Stop-ClusterGroup "Cloud Management"

Start-ClusterGroup "Cloud Management"

This won't fully remediate the issue as the progress details may still not be displayed for a duration of the update process. To get the latest update details, you can Retrieve the update progress with PowerShell.
Security When fixing the compliance for the minimum password length rule, even after you've changed the minimum password length on the Azure Local host to 14, you continue to see it as noncompliant in Azure policy. You can verify the length of the password using the net accounts cmdlet. In the output, find Minimum password length to see the value.
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment After Azure portal deployment, SConfig network settings shows the error: Set-SCfNetworksetting : Cannot bind argument to parameter 'Value' because it is null. There's no known workaround in this release.
Deployment In rare instances, deployment fails with errors during validation that state that the mandatory Arc extensions are not yet installed. If you face this issue, retry the deployment.
Update When installing cumulative updates using Azure Update Manager, only the latest update for version 2507 is installed. If earlier update options (for versions 2505, 2506) are selected, they are not installed. There's no workaround in this release.
Update Update health check results not shown when it's been longer than 3 hours after health check completed Rerun the update health checks to restart the 3 hour expiration clock.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2508

For the 2508 release of Azure Local, Microsoft released two security updates, each aligned with a specific OS build. The following table provides the specific versions and their OS builds:

Solution version OS build
11.2508.1001.51 25398.1791
12.2508.1001.52 26100.4946

Important

The new deployments of this software use the 12.2508.1001.52 build. You can also update an existing deployment from 2507 by using 11.2508.1001.51.

Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Comments
Azure Local VMs For Disconnected operations (preview) with Azure Local VMs, the Arc agent download will fail. Corrected end point and now Arc agent download succeeds.
Deployment Fixed issue for importing paths with spaces.
Update Fixed issue where active CAU runs get canceled prematurely.
Deployment Disk types not being determined correctly, resulting in false negative results on storage pool size determination.

Known issues

The following table lists the known issues in this release:

Feature Issue Workaround
Registration, deployment, Add-server, Upgrade, Update In this release, connectivity tests can take longer than 20 minutes in some environments. There's no known workaround for this issue in this release.
Update Update health check results not shown when it's been longer than 3 hours after health check completed Rerun the update health checks to restart the 3 hour expiration clock.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Update When updating the Azure Local instance via the Azure Update Manager, the update progress and results may not be visible in the Azure portal. To work around this issue, on each node, add the following registry key (no value needed):

New-Item -Path "HKLM:\SYSTEM\CurrentControlSet\Services\HciCloudManagementSvc\Parameters" -force

Then on one of the nodes, restart the Cloud Management cluster group.

Stop-ClusterGroup "Cloud Management"

Start-ClusterGroup "Cloud Management"

This won't fully remediate the issue as the progress details may still not be displayed for a duration of the update process. To get the latest update details, you can Retrieve the update progress with PowerShell.
Security When fixing the compliance for the minimum password length rule, even after you've changed the minimum password length on the Azure Local host to 14, you continue to see it as non-compliant in Azure policy. You can verify the length of the password using the net accounts cmdlet. In the output, find Minimum password length to see the value.
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment After Azure portal deployment, SConfig network settings shows the error: Set-SCfNetworksetting : Cannot bind argument to parameter 'Value' because it is null. There's no known workaround in this release.
Deployment In rare instances, deployment fails with errors during validation that state that the mandatory Arc extensions are not yet installed. If you face this issue, retry the deployment.
Update When installing cumulative updates using Azure Update Manager, only the latest update for version 2507 is installed. If earlier update options (for versions 2505, 2506) are selected, they are not installed. There's no workaround in this release.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2507

For the 2507 release of Azure Local, Microsoft released two security updates, each aligned with a specific OS build. The following table provides the specific versions and their OS builds:

Solution version OS build
11.2507.1001.9 25398.1732
12.2507.1001.10 26100.4652

Important

The new deployments of this software use the 12.2507.1001.10 build. You can also update an existing deployment from 2506 by using 11.2507.1001.9.

Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Comments
Update Solution update fails with the error: Unable to install solution update "11.2506.1001.24" - Type 'UpdateFOD' of Role 'ComposedImageUpdate' raised an exception.
Upgrade The upgrade banner is currently available for users using the Azure Government cloud. However, the environment checker fails, suggesting that Azure Government clouds are not supported.
Azure Local VMs When no storage path is specified during deployment, resources (VMs, data disks, and images) are automatically placed on the first storage path of the cluster, even when other storage paths are also available. Over time, this might cause insufficient disk space on that path, potentially resulting in deployment failures.
Update Cluster-Aware Updating runs might fail with the error:
Type 'SBEPartnerConfirmCauDone' of Role 'SBE' raised an exception:<br>SBE_MsftCIOnlyCommon_CommonForTesting_4.2.2504.16: ErrorID: SBE-CAU-RUNNING-AFTER-DONE -- CAU run is still in progress when it should be done. See https://aka.ms/AzureLocal/SBE/CauHelp for help. Review full Get-CauRun output it identify if it is progressing or stuck. Wait for it to complete if progressing.

Known issues

The following is a known issue with this release.

Feature Issue Workaround
Update When installing cumulative updates using Azure Update Manager, only the latest update for version 2507 is installed. If earlier update options (for versions 2505, 2506) are selected, they are not installed. There's no workaround in this release.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Update When updating the Azure Local instance via the Azure Update Manager, the update progress and results may not be visible in the Azure portal. To work around this issue, on each node, add the following registry key (no value needed):

New-Item -Path "HKLM:\SYSTEM\CurrentControlSet\Services\HciCloudManagementSvc\Parameters" -force

Then on one of the nodes, restart the Cloud Management cluster group.

Stop-ClusterGroup "Cloud Management"

Start-ClusterGroup "Cloud Management"

This won't fully remediate the issue as the progress details may still not be displayed for a duration of the update process. To get the latest update details, you can Retrieve the update progress with PowerShell.
Security When fixing the compliance for the minimum password length rule, even after you've changed the minimum password length on the Azure Local host to 14, you continue to see it as non-compliant in Azure policy. You can verify the length of the password using the net accounts cmdlet. In the output, find Minimum password length to see the value.
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment After Azure portal deployment, SConfig network settings shows the error: Set-SCfNetworksetting : Cannot bind argument to parameter 'Value' because it is null. There's no known workaround in this release.
Deployment In rare instances, deployment fails with errors during validation that state that the mandatory Arc extensions are not yet installed. If you face this issue, retry the deployment.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2506

For the 2506 release of Azure Local, Microsoft released two security updates, each aligned with a specific OS build. The following table provides the specific versions and their OS builds:

Solution version OS build
11.2506.1001.28 25398.1665
12.2506.1001.29 26100.4349

Important

The new deployments of this software use the 12.2506.1001.29 build. You can also update an existing deployment from 2505 using 11.2506.1001.28.

Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Workaround/Comments
Azure Local VMs When creating a VM, there's no option to prompt for password. Password is prompted when --password isn't explicitly passed. This provides customers a more secure way to input password.
Azure Local VMs When creating a VM, --admin-username and --password are required for guest management enablement. Unbound requirement for guest management enablement. Now, when creating a VM, --admin-username and --password are required unless the OS disk is provided or the --authentication-type is ssh.
Azure Local VMs In some cases, VM update operations were incorrectly shown as successful on the Azure portal even though they had failed in the underlying infrastructure. Improved error reporting to ensure the Azure portal accurately reflects the true status of VM update operations.
Azure Local VMs Deleting a logical network with an attached AKS cluster would leave behind orphaned on-premises resources. Properly detects and handles AKS cluster dependencies during logical network deletion to prevent resource leaks.
Azure Local VMs Unable to create VHDs using --download-url Correctly initializes VHD property before assigning the download URL, ensuring successful disk creation.
Azure Local VMs In some cases, deleting a VM from multiple platforms (Azure portal, Windows Admin Center, Failover Cluster Manager) could cause its high availability (HA) status to remain stuck in a pending state. Correctly updates the HA status when the VM is removed from the cluster, preventing it from getting stuck.
Azure Local VMs After creating a VM, a temporary ISO file used for guest agent setup was sometimes left attached. ISO file is automatically removed after VM creation.
Azure Local VMs In some cases, VM deletions did not properly clean up associated network interfaces or disks, leaving behind orphaned resources that blocked reuse of IP addresses and storage. All dependent resources are correctly detached and deleted when a VM is deleted, preventing resource leaks.
Deployment Improved validation for the IP address of the infrastructure range.
Deployment Improved behavior for loading PowerShell modules.
Deployment Added flag to enable decryption of volumes when re-attaching them.

Known issues in this release

The following table lists the known issues in this release:

Feature Issue Workaround
Security Azure Local might face an issue during normal operations (for example, Update, Repair) while using Defender for Endpoint and when the Restrict App Execution setting is enabled for one or more servers in the deployment. Disable the Restrict App Execution setting in the Defender portal and reboot. If the issue persists, open a support case.
Deployment In rare instances, deployment fails with errors during validation that state that the mandatory Arc extensions are not yet installed. If you face this issue, retry the deployment.
Update Solution update fails with the error: Unable to install solution update "11.2506.1001.24" - Type 'UpdateFOD' of Role 'ComposedImageUpdate' raised an exception. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Deployment After Azure portal deployment, SConfig network settings shows the error: Set-SCfNetworksetting : Cannot bind argument to parameter 'Value' because it is null. There's no known workaround in this release.
Update Cluster-Aware Updating runs might fail with the error:
Type 'SBEPartnerConfirmCauDone' of Role 'SBE' raised an exception:<br>SBE_MsftCIOnlyCommon_CommonForTesting_4.2.2504.16: ErrorID: SBE-CAU-RUNNING-AFTER-DONE -- CAU run is still in progress when it should be done. See https://aka.ms/AzureLocal/SBE/CauHelp for help. Review full Get-CauRun output it identify if it is progressing or stuck. Wait for it to complete if progressing.
Wait for CAU run to complete (wait for Get-CauRun to report RunNotInProgress) and resume the update.
Azure Local VMs When no storage path is specified during deployment, resources (VMs, data disks, and images) are automatically placed on the first storage path of the cluster, even when other storage paths are also available. Over time, this might cause insufficient disk space on that path, potentially resulting in deployment failures. Update to 2507 as this build contains a fix for the issue. Or, create resources with a specified storage path. For more information, see Troubleshoot Azure Local Virtual Machines enabled by Azure Arc.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Azure Verification VMs on Azure Local running Windows Server Azure Edition, Windows 10, or Windows 11 multi-session OS may not activate properly. A pop-up message or a watermark may display, indicating that Windows isn't activated. The VM will function, but the watermark will persist. There's no known workaround in this release.
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release when the Azure portal incorrectly reports the update status as Failed to update or In progress though the update is complete. Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. The Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Update When updating the Azure Local instance via the Azure Update Manager, the update progress and results may not be visible in the Azure portal. To work around this issue, on each node, add the following registry key (no value needed):

New-Item -Path "HKLM:\SYSTEM\CurrentControlSet\Services\HciCloudManagementSvc\Parameters" -force

Then on one of the nodes, restart the Cloud Management cluster group.

Stop-ClusterGroup "Cloud Management"

Start-ClusterGroup "Cloud Management"

This won't fully remediate the issue as the progress details may still not be displayed for a duration of the update process. To get the latest update details, you can Retrieve the update progress with PowerShell.
Security When fixing the compliance for the minimum password length rule, even after you've changed the minimum password length on the Azure Local host to 14, you continue to see it as non-compliant in Azure policy. You can verify the length of the password using the net accounts cmdlet. In the output, find Minimum password length to see the value.
Upgrade The upgrade banner is currently available for users using the Azure Government cloud. However, the environment checker fails, suggesting that Azure Government clouds are not supported. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

This article identifies critical known issues and their workarounds in Azure Local.

These release notes are continuously updated, and as critical issues requiring a workaround are discovered, they're added. Before you deploy your Azure Local instance, carefully review the information contained here.

Important

For information about supported update paths for this release, see Release information.

For more information about new features in this release, see What's new for Azure Local.

Known issues for version 2505

For the 2505 release of Azure Local, Microsoft released two security updates: one for existing deployments and another for new deployments. The following table provides information about different deployment types, their corresponding security updates, and OS builds:

Deployment type Solution version OS build
Existing deployments 11.2505.1001.22 25398.1611
New deployments 12.2505.1001.23 26100.4061

Important

The new deployments of this software use the 12.2505.1001.23 build. You can also update an existing deployment from 2504 using 11.2505.1001.22.

Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Workaround/Comments
Azure Local VMs Timeout errors occur when creating a disk and gallery image. Extended the timeout period to support larger disk and image size.
Azure Local VMs Mocguestagent gets stuck in a start/stop loop and can't connect to the host agent. Improved Mocguestagent connectivity.
Azure Local VMs Cannot delete a VM in the Azure portal or CLI after deleting it in Hyper-V. If you delete the VM in Hyper-V instead of the Azure portal or CLI, delete the resource it in the Azure portal or CLI to clean it up.
Azure Local VMs Unable to delete a data disk that failed to attach during VM creation. Fixed the attachment state of data disk to be accurate. You can now delete a data disk if it fails to attach during VM creation.
Azure Local VMs OS updates can automatically install on the third Tuesday of the month (for example, May 20, June 17, or July 15). This can cause unexpected outages or future issues because of mismatched OS versions for Azure Local. Removed logic that set up a scheduled CAU as part of some SBE updates.
Deployment Deployment via Azure Resource Manager (ARM) template fails with the following error during validation:
Type 'DeployArb' of Role 'MocArb' raised an exception: [DeployArb:Fetching SP in CloudDeployment Scenario] Exception calling 'GetCredential' with '1' argument(s): 'Exception of type 'CloudEngine.Configurations.SecretNotFoundException' was thrown.' at at Get-CloudDeploymentServicePrincipal.
The error is because of a missing SPN during deployment.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Solution extension Changed the severity level for the health check that validates the plugin's presence.
Azure portal Added a fix to fetch Storage account properties during redeployment.
Azure portal Added an error message about subscription access rights during Azure Local download.
Azure portal Added a warning message if the subscription isn't registered during the Azure Local download.
Azure portal Improved logic to show Azure Arc-enabled machines for selection during deployment.
Azure portal Enhanced the setting of machines' state during deployment.
Azure portal Improved error handling and data processing.
Azure portal Added a fix to disable the validation button until extension installations are complete during deployment.
Deployment Added validation to check for required permissions to change access control lists (ACL).
Deployment Added a server-side filter to the WMI call to improve performance and reduce call time.
Deployment Added fix to dispose of the PowerShell objects after use to improve performance when creating new PowerShell instances.
Deployment Fixed the time zone issue when running the Install Azure Stack HCI wizard and selecting a time zone. The Install Azure Stack HCI wizard now correctly applies the selected time zone during deployment.
Security management Fixed Security defaults, Application control, and Data protections pages showing as Unknown in the security compliance report.
Add server
Repair server
Fixed running Add-server and Repair-server cmdlets with a customized storage adapter IP in Azure Local which resulted in the error: Type 'ConfigureAzureStackHostStorageAdpaterIPAddressesInAddRepairNode' of Role 'HostNetwork' raised an exception: Connecting to remote server <MACHINE> failed with the following error message: Access is denied..

Known issues in this release

The following table lists the known issues in this release:

Feature Issue Workaround
Update A critical VM operational status not OK alert is shown in the Azure portal under Update readiness and in the Alerts pane after the update has completed successfully. Additionally, the alert appears when running the Get-HealthFault cmdlet. No action is required on your part. This alert will resolve automatically in a few days.
Deployment Updating Azure Arc extensions manually from the Azure Local Machine page via the Azure portal will result in issues during deployment. The extensions that shouldn't be updated manually are: AzureEdgeDeviceManagement, AzureEdgeLifecycleManager, and AzureEdgeAKVBackupForWindows. Installing extensions manually from the Azure Local machine page isn't supported.
Upgrade The upgrade banner is currently available for users using the Azure Government cloud. However, the environment checker fails, suggesting that Azure Government clouds are not supported. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release where the Azure portal may incorrectly display the update status as Failed to update or In progress, even though the update has actually completed successfully. This behavior is particularly observed when updating Azure Local instances via Azure Update Manager, where the update progress and results may not be visible in the portal. You might need to wait up to 30 minutes or more to see the updated status. If the status still isn't refreshed after that time, follow these steps: Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Security When fixing the compliance for the minimum password length rule, even after you've changed the minimum password length on the Azure Local host to 14, you continue to see it as non-compliant in Azure policy. You can verify the length of the password using the net accounts cmdlet. In the output, find Minimum password length to see the value.
Azure Arc registration When registering a new machine with Azure Arc, registration fails during ImageRecipeValidationTests with the following error:
"Responses": [ { "Name": "ImageRecipeValidation", "Status": "Failed", "Errors": [ { "ErrorMessage": "Diagnostics failed for the test category: ImageRecipeValidation.", "StackTrace": null, "ExceptionType": "DiagnosticsTestFailedException", "RecommendedActions": [ "Please contact Microsoft support." ] } ] }
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Upgrade Failed to upgrade cluster with Get-AzureStackHCI ConnectionStatus in RepairRegistration due to the Virtualization-Based Security (VBS) master key lost during Secure Boot certificate installation. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Registration After installing certain updates (including BIOS), clusters may report RepairRegistrationRequired and show a stale connection state. For detailed steps on how to resolve this issue, see the Troubleshooting guide.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2504

For the 2504 release of Azure Local, Microsoft released two security updates: one for existing deployments and another for new deployments. The following table provides information about different deployment types, their corresponding security updates, and OS builds:

Deployment type Solution version OS build
Existing deployments 11.2504.1001.21 25398.1551
New deployments 12.2504.1001.20 26100.3775

Important

The new deployments of this software use the 12.2504.1001.20 build. You can also update an existing deployment from 2503 using 11.2504.1001.21.

Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Workaround/Comments
Azure Local VMs Subsequent deployments or operations on source image fail if a VM is created with OS disk and source image located on the same CSV and an error occurred during a previous copy of the source image.
Azure Local VMs Multiple Hyper-V VMs for a single Azure Local VM that failed to be created are left running on Azure Local.
Azure Local VMs Storage path deletion times out if AKS node disks are present. No timeout and will error out with proper error message listing the resources on the storage path.
Azure Local VMs Issues with deleting resources that are in use by other resources that no longer exist on the cluster. Enhanced validation to check if resources are present on the cluster before attempting to delete.
Azure Local VMs Issues with deleting and resizing an Azure Local VM that has checkpoint(s). You can checkpoint an Azure Local VM with on-premises tools. The VM will remain manageable from the Azure portal.
Solution extension Improved error message to fix firewall blocking access to solution extension manifest endpoints.
Solution extension Improved reliability of copying solution extension content locally to each machine.
Solution extension Added specification of plug-in name in the solution extension.
Solution extension Fixed issue where system was unable to get available solution extension updates.
Update Simplified the Azure portal experience for viewing the progress and history of update runs.
Update When monitoring update progress in the Azure Update Management portal, the progress might appear to not have updated for several hours. Run Get-SolutionUpdate on one of the cluster nodes. If an update object is returned, the update might be taking longer than expected but it is progressing. If an update object isn't returned, the update may be stalled. For detailed steps on how to resolve this issue, see the Troubleshooting guide.

Known issues in this release

The following table lists the known issues in this release:

Feature Issue Workaround
Add server
Repair server
The Add-server and Repair-server cmdlets fail with the error:
Cluster Build ID matches node to add's Build ID.
Use the OS image of the same solution version as that running on the existing cluster. To get the OS image, identify and download the image version from this Release table.
Add server
Repair server
If running Add-server and Repair-server cmdlets with customized storage adapter IP configured in your Azure Local instance, the operation might fail with the error:
Type 'ConfigureAzureStackHostStorageAdpaterIPAddressesInAddRepairNode' of Role 'HostNetwork' raised an exception: Connecting to remote server <MACHINE> failed with the following error message : Access is denied..
Contact Microsoft Support if you experience this issue.
Security management The Security defaults, Application control, and Data protections pages show Unknown for all security settings. This issue is only in the security compliance report. The states of the security settings are unaffected. Use PowerShell to verify the compliance status of the security settings.

For more information, see Manage secure baseline via PowerShell cmdlets
Deployment In this release and previous releases, registration fails with the following error when you try to register Azure Local machines with Azure Arc:
AZCMAgent command failed with error: >> exitcode: 42. Additional Info: See https://aka.ms/arc/azcmerror.
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Azure Arc registration When registering a new machine with Azure Arc, registration fails during ImageRecipeValidationTests with the following error:
"Responses": [ { "Name": "ImageRecipeValidation", "Status": "Failed", "Errors": [ { "ErrorMessage": "Diagnostics failed for the test category: ImageRecipeValidation.", "StackTrace": null, "ExceptionType": "DiagnosticsTestFailedException", "RecommendedActions": [ "Please contact Microsoft support." ] } ] }
For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Deployment In the 2504 release, when you run the Install Azure Stack HCI wizard and select a time zone, the system continues to use Pacific Standard Time (PST), regardless of your selection. This isn't a blocking issue, and deployment will still complete successfully.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release where the Azure portal may incorrectly display the update status as Failed to update or In progress, even though the update has actually completed successfully. This behavior is particularly observed when updating Azure Local instances via Azure Update Manager, where the update progress and results may not be visible in the portal. You might need to wait up to 30 minutes or more to see the updated status. If the status still isn't refreshed after that time, follow these steps: Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Security When fixing the compliance for the minimum password length rule, even after you've changed the minimum password length on the Azure Local host to 14, you continue to see it as non-compliant in Azure policy. You can verify the length of the password using the net accounts cmdlet. In the output, find Minimum password length to see the value.
Upgrade Failed to upgrade cluster with Get-AzureStackHCI ConnectionStatus in RepairRegistration due to the Virtualization-Based Security (VBS) master key lost during Secure Boot certificate installation. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Registration After installing certain updates (including BIOS), clusters may report RepairRegistrationRequired and show a stale connection state. For detailed steps on how to resolve this issue, see the Troubleshooting guide.

Known and expected behaviors

The following table lists the known and expected system behaviors that shouldn't be considered as bugs or limitations.

Feature Behavior Workaround
Operating system Restoring the registry using RegBack isn't supported on Azure Local. This operation can remove the Lifecycle Manager (LCM) and Microsoft On-premises Cloud (MOC) settings on your Azure Local instance, which can corrupt the solution.
Azure Local VM management Using an exported Azure VM OS disk as a VHD to create a gallery image for provisioning an Azure Local VM is unsupported. Run the command restart-service mochostagent to restart the mochostagent service.

Known issues for version 2503

This software release maps to software version number 2503.0.13.

Important

The new deployments of this software use the 2503.0.13 build. You can also update from 2411.3.

Release notes for this version include the issues fixed in this release, known issues in this release, and release note issues carried over from previous versions.

Note

For detailed remediation for common known issues, see the Azure Local Supportability GitHub repository.

Fixed issues

The following table lists the fixed issues in this release:

Feature Issue Workaround/Comments
Azure Local VMs Intermittent failures or timeouts when creating or deleting Azure Local VMs due to internal service entering a deadlocked state. You can now deploy Azure Local VMs without facing timeout issues.
Azure Local VMs Unable to delete a gallery image if the source Azure Local VM gallery image was created from, was deleted. You can now delete a gallery image even if the source Azure Local VM the gallery image was created from, was deleted.
Azure Local VMs After Azure Arc resource bridge disaster recovery, some Azure Local VMs may lose network connectivity. Updated Azure Arc resource bridge disaster recovery logic to ensure network configurations of Azure Local VMs persists.
Azure Local VMs Inaccurate power state reporting on Azure for Azure Local VMs when the source image for the VM is deleted from the cluster. Power operations shows consistent and accurate power state representation on Azure for Azure Local VMs with deleted source images.
Azure Local VMs Inaccurate power state reporting on Azure for Azure Local VMs when attempting power operations. Improved the accuracy of power state reconciliation by reducing latency and fixing a bug that impacted power operations visibility to Azure.
Azure Local VMs Unable to create a gallery image with specified storage path. You can now create a gallery image with specified storage path.
Azure Local VMs Running azcmagent disconnect deletes the Azure Local VM. Don't use azcmagent disconnect on Azure Local VMs. Use only azcmagent disconnect -f or azcmagent disconnect --force-local-only. This command disconnects the connected machine agent and keeps the VM running (no longer deletes).
Azure Local VM management The Mochostagent service might appear to be running but can get stuck without updating logs for over a month. You can identify this issue by checking the service logs in C:\programdata\mochostagent\logs to see if logs are being updated. Run the following command to restart the mochostagent service: restart-service mochostagent.
Azure Local VMs/Azure Migrate Migration of Gen 1 (non-sysprep) VMs using Azure Migrate fails with the error: Failed to clean up seed ISO disk from the file system for VM. Migration of Gen 1 (non-sysprep) VMs using Azure Migrate will no longer fail with this error.
Azure Local VMs/Azure Migrate VM creation from OS disk fails due to incorrect storage precheck. VM creation succeeds regardless if the source is gallery image or OS disk.
Deployment When trying to deploy via the Azure portal, Azure Local machine nodes aren't visible in the Azure portal. Azure Local deployments via the Azure portal are only supported for 2503 and later. For previous versions, deploy via Azure Resource Manager (ARM) template.
Deployment During the Azure Local deployment, DeviceManagementExtension fails to install when a proxy is configured. Install previous DeviceManangementExtension version 1.2502.0.3012 when using a proxy.
Update When updating from version 2408.2.7 to 2411.0.24, the update process could fail with the following error message: Type 'CauPreRequisites' of Role 'CAU' raised an exception: Could not finish cau prerequisites due to error 'Cannot remove item C:\UpdateDistribution\<any_file_name>: Access to the path is denied.' This issue is now fixed.
Update Ensure that Solution Builder Extension updates of type Notify that have been imported get copied correctly before validation.
Update Improved the Solution Extension Secret Location cmdlet help to provide better examples.
Update Added retry logic to Cluster-Aware Updating runs and health checks for cluster nodes.
Update Increased system stability during the .NET 8 updates.
Update With the 2411 release, solution and Solution Builder Extension update aren't combined in a single update run. To apply a Solution Builder Extension package, you need a separate update run.
Upgrade Disable the Carbon PowerShell module if detected and load the known modules.
Upgrade Optimized the current Carbon PowerShell module solution.
Upgrade Added a check to validate enough free memory to start an Azure Arc resource bridge VM.
Security Mitigation for security vulnerability CVE-2024-21302 was implemented. See the Guidance for blocking rollback of Virtualization-based Security (VBS) related security updates
Deployment During Azure Local deployment via portal, Validate selected machines fails with this error message: Mandatory extension [Lcm controller] installed version [30.2503.0.907] is not equal to the required version [30.2411.2.789] for Arc machine [Name of the machine]. Please create EdgeDevice resource again for this machine to fix the issue. Reinstall the correct version of AzureEdgeLifecycleManager extension. Follow these steps:
1. Select the machine and then select Install extensions.

Screenshot of extension installation on Azure Local machines.

2. Repeat this step for each machine you intend to cluster. It takes roughly 15 minutes for the installation to complete.
3. Verify that the AzureEdgeLifecycleManager extension version is 30.2411.2.789.

Screenshot of extension version installed on Azure Local machines that can be validated.

4. After the extensions are installed on all the machines in the list, select Add machines to refresh the list.
5. Select Validate selected machines. The validation should succeed.
Deployment During the Azure Local deployment and update on OEM-licensed devices, ConfigureSecurityBaseline fails at the Apply security settings on servers step. This issue is now fixed.

Known issues in this release

The following table lists the known issues in this release:

Feature Issue Workaround/Comments
Azure Local VMs Unable to delete an Azure Local VM enabled by Arc when one of the nodes is down. Contact Microsoft Support if this issue occurs.
Azure Local VMs Checkpoints on Azure Local VMs cause sync issues between Azure Local VMs and Portal. Remove any checkpoints on the VM.
Updates Updating to 2503.0.13 failed Update PreRequisites for Role 'MocArb' with exception: SyntaxWarning: invalid escape sequence '\W' at CheckAndInstall-CliExtensions. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Security vulnerability There's a known security vulnerability in this release that may affect security scans for the updates. For more information, see Azure Arc Installer vulnerability CVE-2025-26627 - Host
Azure Local VMs There's a known security vulnerability in this release that may affect security scans for the updates. For more information, see Azure Arc Installer vulnerability CVE-2025-26627 - Guest
Microsoft Defender for Cloud

Azure Government
In the Azure Government cloud, Microsoft Defender for Cloud recommendations for Azure Local do not show up in the Microsoft Defender for Cloud portal.
Metrics

Azure Government
Metrics from Azure Local clusters in the Azure Government cloud fail to reach Azure. As a result, metrics don't show up in the Monitoring, Metrics, or workbook graphs. Metrics based alerts aren't triggered and new alerts can't be set up.
Upgrade Failed to upgrade cluster with Get-AzureStackHCI ConnectionStatus in RepairRegistration due to the Virtualization-Based Security (VBS) master key lost during Secure Boot certificate installation. For detailed steps on how to resolve this issue, see the Troubleshooting guide.
Registration After installing certain updates (including BIOS), clusters may report RepairRegistrationRequired and show a stale connection state. For detailed steps on how to resolve this issue, see the Troubleshooting guide.

Known issues from previous releases

The following table lists the known issues from previous releases:

Feature Issue Workaround
Update When viewing the readiness check results for an Azure Local instance via the Azure Update Manager, there might be multiple readiness checks with the same name. There's no known workaround in this release. Select View details to view specific information about the readiness check.
Update There's an intermittent issue in this release where the Azure portal may incorrectly display the update status as Failed to update or In progress, even though the update has actually completed successfully. This behavior is particularly observed when updating Azure Local instances via Azure Update Manager, where the update progress and results may not be visible in the portal. You might need to wait up to 30 minutes or more to see the updated status. If the status still isn't refreshed after that time, follow these steps: Connect to your Azure Local instance via a remote PowerShell session. To confirm the update status, run the following PowerShell cmdlets:

$Update = get-solutionupdate| ? version -eq "<version string>"

Replace the version string with the version you're running. For example, "10.2405.0.23".

$Update.state

If the update status is Installed, no further action is required on your part. Azure portal refreshes the status correctly within 24 hours.
To refresh the status sooner, follow these steps on one of the nodes.
Restart the Cloud Management cluster group.
Stop-ClusterGroup "Cloud Management"
Start-ClusterGroup "Cloud Management"
Add server In this release and previous releases, when adding a machine to the system, isn't possible to update the proxy bypass list string to include the new machine. Updating environment variables proxy bypass list on the hosts won't update the proxy bypass list on Azure resource bridge or AKS. There's no workaround in this release. If you encounter this issue, contact Microsoft Support to determine next steps.
Azure portal In some instances, the Azure portal might take a while to update and the view might not be current. You might need to wait for 30 minutes or more to see the updated view.
Security When fixing the compliance for the minimum password length rule, even after you've changed the minimum password length on the Azure Local host to 14, you continue to see it as non-compliant in Azure policy. You can verify the length of the password using the net accounts cmdlet. In the output, find Minimum password length to see the value.

Next steps