Share via

VMSS still failing after Feb 2 incident — VMSS control‑plane operations broken, only instance APIs work

Mike Jones 20 Reputation points Microsoft Employee
2026-02-03T19:48:22.6166667+00:00

I have a Virtual Machine Scale Set running in East US 2 that was impacted by the Feb 2 service‑management outage. During the outage, the VMSS failed to download extensions from Microsoft‑managed storage. Autoscaling attempted to scale out, and I manually deleted several instances while the outage was active (I did not know at the time that a platform incident was occurring).

After the outage was marked “mitigated,” three of my VMSS recovered normally. One VMSS did not recover and remains in a broken state.

Azure Virtual Machine Scale Sets
Azure Virtual Machine Scale Sets

Azure compute resources that are used to create and manage groups of heterogeneous load-balanced virtual machines.

{count} votes

Answer accepted by question author
  1. Manish Deshpande 4,450 Reputation points Microsoft External Staff Moderator
    2026-02-03T20:50:20.8166667+00:00

    Hello Mike

    This is not an instance-level issue and not a configuration problem in your VMSS.

    Your VM Scale Set is in a partially corrupted control‑plane state due to the February 2, 2026 Azure service‑management incident. During that incident, Azure confirmed a platform issue where VM and VMSS management operations failed because access to Microsoft‑managed extension storage was disrupted. As a result, VM lifecycle and scale‑set model updates were intermittently failing in affected regions, including East US 2.

    In your case:

    • Instance‑level operations succeed (instance view shows ProvisioningState/succeeded)
    • Control‑plane operations fail with 500 InternalOperationError
      • vmss show
        • scale
          • resource tag write
          • This indicates that the VMSS model persisted in an inconsistent state after instance deletions occurred while the platform outage was active

    Microsoft confirms that manual instance deletes during a VMSS control‑plane outage can leave the scale‑set model unrecoverable without backend repair, even after the incident is marked mitigated.

    There is no customer‑side fix for this condition.

    The only supported resolution is:

    • Microsoft backend repair of the VMSS model, or
    • Recreation of the VMSS from a clean configuration

    Microsoft requires a support‑initiated backend action to restore control‑plane consistency when this condition occurs. [learn.microsoft.com]

    If an Azure support case can be opened, provide:

    • VMSS name
    • Region
    • Correlation IDs from failing operations
    • Confirmation that the VMSS was modified during the Feb 2 incident window

    Microsoft engineering teams use this to perform a model reconciliation or forced repair.

    Summary :

    Your VMSS is healthy at the instance level but broken at the control plane due to a known Azure platform incident. This is a Microsoft‑side condition, and recovery requires backend intervention or VMSS re‑creation. Customer‑initiated retries, redeploys, or scaling changes cannot fix this state.
    Manual patching is needed from the backend.

    Thanks,
    Manish Deshpande.

    0 comments No comments

0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.