Information
Active - Virtual Machines and dependent services - Service management issues in multiple regions
Impact statement: As early as 19:46 UTC on 2 February 2026, we are aware of an ongoing issue causing customers to receive error notifications when performing service management operations - such as create, delete, update, scaling, start, stop - for Virtual Machines (VMs) across multiple regions. These issues are also causing impact to services with dependencies on these service management operations - including Azure Arc Enabled Servers, Azure Batch, Azure DevOps, Azure Load Testing, and GitHub. For details on the latter, please see https://www.githubstatus.com.
Current status: We have determined that these issues were caused by a recent configuration change that affected public access to certain Microsoft‑managed storage accounts, used to host extension packages. We are actively working on mitigation, including updating configuration to restore relevant access permissions. We have applied this update in one region so far, and are assessing the extent to which this mitigates customer issues. Our next update will be provided by 22:30 UTC, approximately 60 minutes from now.
Since there is no GitHub CEO, (Satya is not bothered anymore) and human employees not looking, Tay and Zoe are at the helm ruining GitHub with their broken AI generated fixes.
Which is again even worse.
"Impact statement: As early as 19:46 UTC on 2 February 2026, we are aware of an ongoing issue causing customers to receive error notifications when performing service management operations - such as create, delete, update, scaling, start, stop - for Virtual Machines (VMs) across multiple regions. These issues are also causing impact to services with dependencies on these service management operations - including Azure Arc Enabled Servers, Azure Batch, Azure DevOps, Azure Load Testing, and GitHub. For details on the latter, please see https://www.githubstatus.com."
I don't get how Microsoft views this level of service as acceptable.
Ran into an issue upgrading an AKS cluster last week. It completely stalled and broke the entire cluster in a way where our hands were tied as we can't see the control plane at all...
I submit a severity A ticket and 5 hours later I get told there was a known issue with the latest VM image that would create issues with the control plane leaving any cluster that was updated in that window to essentially kill itself and require manual intervention. Did they notify anyone? Nope, did they stop anyone from killing their own clusters. Nope.
It seems like every time I'm forced to touch the Azure environment I'm basically playing Russian roulette hoping that something's not broken on the backend.
Must be nice to be a monopoly that has most of the businesses in the world as their hostages.
There’s a bunch of hardware, and they can’t run more servers than they have hardware. I don’t see a way around that.
De-risk yourself from Microsoft
Or, if part of a future plan: how?
Resolved - On February 2, 2026, between 18:35 UTC and 22:15 UTC, GitHub Actions hosted runners were unavailable, with service degraded until full recovery at 23:10 UTC for standard runners and at February 3, 2026 00:30 UTC for larger runners. During this time, Actions jobs queued and timed out while waiting to acquire a hosted runner. Other GitHub features that leverage this compute infrastructure were similarly impacted, including Copilot Coding Agent, Copilot Code Review, CodeQL, Dependabot, GitHub Enterprise Importer, and Pages. All regions and runner types were impacted. Self-hosted runners on other providers were not impacted.
This outage was caused by a backend storage access policy change in our underlying compute provider that blocked access to critical VM metadata, causing all VM create, delete, reimage, and other operations to fail. More information is available at https://azure.status.microsoft/en-us/status/history/?trackingId=FNJ8-VQZ. This was mitigated by rolling back the policy change, which started at 22:15 UTC. As VMs came back online, our runners worked through the backlog of requests that hadn’t timed out.
We are working with our compute provider to improve our incident response and engagement time, improve early detection before they impact our customers, and ensure safe rollout should similar changes occur in the future. We recognize this was a significant outage to our users that rely on GitHub’s workloads and apologize for the impact this had.
Feb 3, 00:56 UTC
Update - Actions is operating normally.
Feb 3, 00:56 UTC
Update - Based on our telemetry, most customers should see full recovery from failing GitHub Actions jobs on hosted runners.
We are monitoring closely to confirm complete recovery.
Other GitHub features that rely on GitHub Actions (for example, Copilot Coding Agent and Dependabot) should also see recovery.
Feb 2, 23:50 UTC
Update - Actions is experiencing degraded performance. We are continuing to investigate.
Feb 2, 23:43 UTC
Update - Copilot is operating normally.
Feb 2, 23:42 UTC
Update - Pages is operating normally.
Feb 2, 23:31 UTC
Update - Our upstream provider has applied a mitigation to address queuing and job failures on hosted runners.
Telemetry shows improvement, and we are monitoring closely for full recovery.
Feb 2, 22:53 UTC
Update - We continue to investigate failures impacting GitHub Actions hosted-runner jobs.
We're waiting on our upstream provider to apply the identified mitigations, and we're preparing to resume job processing as safely as possible.
Feb 2, 22:10 UTC
Update - Copilot is experiencing degraded performance. We are continuing to investigate.
Feb 2, 21:27 UTC
Update - We continue to investigate failures impacting GitHub Actions hosted-runner jobs.
We have identified the root cause and are working with our upstream provider to mitigate.
This is also impacting GitHub features that rely on GitHub Actions (for example, Copilot Coding Agent and Dependabot).
Feb 2, 21:13 UTC
Update - The team continues to investigate issues causing GitHub Actions jobs on hosted runners to remain queued for extended periods, with a percentage of jobs failing. We will continue to provide updates as we make progress toward mitigation.
Feb 2, 20:27 UTC
Update - Pages is experiencing degraded performance. We are continuing to investigate.
Feb 2, 19:48 UTC
Update - The team continues to investigate issues causing GitHub Actions jobs on hosted runners to remain queued for extended periods, with a percentage of jobs failing. We will continue to provide updates as we make progress toward mitigation.
Feb 2, 19:44 UTC
Update - Actions is experiencing degraded availability. We are continuing to investigate.
Feb 2, 19:43 UTC
Update - GitHub Actions hosted runners are experiencing high wait times across all labels. Self-hosted runners are not impacted.
Feb 2, 19:07 UTC
Investigating - We are investigating reports of degraded performance for Actions
Feb 2, 19:03 UTC
Their status page seems to think everything's A-OK.