Skip to content

Commit a5629fb

Browse files
committed
Merge branch 'release-local-2506' of https://github.com/MicrosoftDocs/azure-stack-docs-pr into rb-2506-upd
2 parents ffc1c61 + 293e7ea commit a5629fb

13 files changed

+503
-992
lines changed

AKS-Arc/TOC.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -193,6 +193,8 @@
193193
href: entra-prompts.md
194194
- name: BGP with FRR not working
195195
href: connectivity-troubleshoot.md
196+
- name: Cluster status stuck during upgrade
197+
href: cluster-upgrade-status.md
196198
- name: Reference
197199
items:
198200
- name: Azure CLI

AKS-Arc/aks-troubleshoot.md

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ title: Troubleshoot common issues in AKS enabled by Azure Arc
33
description: Learn about common issues and workarounds in AKS enabled by Arc.
44
ms.topic: how-to
55
author: sethmanheim
6-
ms.date: 06/18/2025
6+
ms.date: 06/27/2025
77
ms.author: sethm
88
ms.lastreviewed: 04/01/2025
99
ms.reviewer: abha
@@ -16,7 +16,7 @@ This section describes how to find solutions for issues you encounter when using
1616

1717
## Open a support request
1818

19-
To open a support request, see the [Get support](/azure/aks/hybrid/help-support) article for information about how to use the Azure portal to get support or open a support request for AKS Arc.
19+
To open a support request, see the [Get support](help-support.md) article for information about how to use the Azure portal to get support or open a support request for AKS Arc.
2020

2121
## Known issues
2222

@@ -28,6 +28,7 @@ The following sections describe known issues for AKS enabled by Azure Arc:
2828
| AKS steady state | [AKS Arc telemetry pod consumes too much memory and CPU](telemetry-pod-resources.md) | Active |
2929
| AKS steady state | [Disk space exhaustion on control plane VMs due to accumulation of kube-apiserver audit logs](kube-apiserver-log-overflow.md) | Active |
3030
| AKS cluster delete | [Deleted AKS Arc cluster still visible on Azure portal](deleted-cluster-visible.md) | Active |
31+
| AKS cluster upgrade | [AKS Arc cluster stuck in "Upgrading" state](cluster-upgrade-status.md) | Fixed in 2505 release |
3132
| AKS cluster delete | [Can't fully delete AKS Arc cluster with PodDisruptionBudget (PDB) resources](delete-cluster-pdb.md) | Fixed in 2503 release |
3233
| Azure portal | [Can't see VM SKUs on Azure portal](check-vm-sku.md) | Fixed in 2411 release |
3334
| MetalLB Arc extension | [Connectivity issues with MetalLB](load-balancer-issues.md) | Fixed in 2411 release |
@@ -42,9 +43,10 @@ The following sections describe known issues for AKS enabled by Azure Arc:
4243
| Create validation | [KubeAPIServer unreachable error](kube-api-server-unreachable.md) |
4344
| Network configuration issues | [Use diagnostic checker](aks-arc-diagnostic-checker.md) |
4445
| Kubernetes steady state | [Resolve issues due to out-of-band deletion of storage volumes](delete-storage-volume.md) |
46+
| Kubernetes steady state | [Repeated Entra authentication prompts when running kubectl with Kubernetes RBAC](entra-prompts.md) |
4547
| Release validation | [Azure Advisor upgrade recommendation message](azure-advisor-upgrade.md) |
4648
| Network validation | [Network validation error due to .local domain](network-validation-error-local.md) |
47-
| BGP with FRR not working | [Troubleshoot BGP with FRR in AKS Arc environments](connectivity-troubleshoot.md) |
49+
| Network validation | [Troubleshoot BGP with FRR in AKS Arc environments](connectivity-troubleshoot.md) |
4850

4951
## Next steps
5052

AKS-Arc/cluster-upgrade-status.md

Lines changed: 147 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,147 @@
1+
---
2+
title: Troubleshoot issue in which the cluster is stuck in Upgrading state
3+
description: Learn how to troubleshoot and mitigate the issue when an AKS enabled by Arc cluster is stuck in 'Upgrading' state.
4+
ms.topic: troubleshooting
5+
author: rcheeran
6+
ms.author: rcheeran
7+
ms.date: 06/27/2025
8+
ms.reviewer: abha
9+
10+
---
11+
12+
# Troubleshoot AKS Arc cluster stuck in "Upgrading" state
13+
14+
This article describes how to fix an issue in which your Azure Kubernetes Service enabled by Arc (AKS Arc) cluster is stuck in the **Upgrading** state. This issue typically occurs after you update Azure Local to version 2503 or 2504, and you then try to upgrade the Kubernetes version on your cluster.
15+
16+
## Symptoms
17+
18+
When you try to upgrade an AKS Arc cluster, you notice that the `currentState` property of the cluster remains in the **Upgrading** state.
19+
20+
```azurecli
21+
az aksarc upgrade --name "cluster-name" --resource-group "rg-name"
22+
```
23+
24+
```output
25+
===> Kubernetes might be unavailable during cluster upgrades.
26+
Are you sure you want to perform this operation? (y/N): y
27+
The cluster is on version 1.28.9 and is not in a failed state.
28+
29+
===> This will upgrade the control plane AND all nodepools to version 1.30.4. Continue? (y/N): y
30+
Upgrading the AKSArc cluster. This operation might take a while...
31+
{
32+
"extendedLocation": {
33+
"name": "/subscriptions/resourceGroups/Bellevue/providers/Microsoft.ExtendedLocation/customLocations/bel-CL",
34+
"type": "CustomLocation"
35+
},
36+
"id": "/subscriptions/fbaf508b-cb61-4383-9cda-a42bfa0c7bc9/resourceGroups/Bellevue/providers/Microsoft.Kubernetes/ConnectedClusters/Bel-cluster/providers/Microsoft.HybridContainerService/ProvisionedClusterInstances/default",
37+
"name": "default",
38+
"properties": {
39+
"kubernetesVersion": "1.30.4",
40+
"provisioningState": "Succeeded",
41+
"currentState": "Upgrading",
42+
"errorMessage": null,
43+
"operationStatus": null
44+
"agentPoolProfiles": [
45+
{
46+
...
47+
```
48+
49+
## Cause
50+
51+
- The issue is caused by a recent change introduced in Azure Local version 2503. Under certain conditions, if there are transient or intermittent failures during the Kubernetes upgrade process, they're not correctly detected or recovered from. This can cause the cluster state to remain in the **Upgrading** state.
52+
- You see this issue if the AKS Arc custom location extension `hybridaksextension` version is 2.1.211 or 2.1.223. You can run the following command to check the extension version on your cluster:
53+
54+
```azurecli
55+
az login --use-device-code --tenant <Azure tenant ID>
56+
az account set -s <subscription ID>
57+
$res=get-archcimgmt
58+
az k8s-extension show -g $res.HybridaksExtension.resourceGroup -c $res.ResourceBridge.name --cluster-type appliances --name hybridaksextension
59+
```
60+
61+
```output
62+
{
63+
"aksAssignedIdentity": null,
64+
"autoUpgradeMinorVersion": false,
65+
"configurationProtectedSettings": {},
66+
"currentVersion": "2.1.211",
67+
"customLocationSettings": null,
68+
"errorInfo": null,
69+
"extensionType": "microsoft.hybridaksoperator",
70+
...
71+
}
72+
```
73+
74+
## Mitigation
75+
76+
This issue was fixed in AKS on [Azure Local, version 2505](/azure/azure-local/whats-new?view=azloc-2505&preserve-view=true#features-and-improvements-in-2505). Upgrade your Azure Local deployment to the 2505 build. After you update, [verify that the Kubernetes version was upgraded](#verification) and the `currentState` property of the cluster shows as **Succeeded**.
77+
78+
### Workaround for Azure Local versions 2503 or 2504
79+
80+
This issue only affects clusters in Azure Local version 2503 or 2504, and on AKS Arc extension versions 2.1.211 or 2.1.223. The mitigation described here is applicable only when you are unable to upgrade to 2505.
81+
82+
You can resolve the issue by running the AKS Arc `update` command. The `update` command restarts the upgrade flow. You can run the `aksarc update` command with placeholder parameters, which do not impact the state of the cluster. So in this case, you can run the `update` command to enable NFS or SMB drivers if those features aren't already enabled. First, check if any of the storage drivers are already enabled:
83+
84+
```azurecli
85+
az login --use-device-code --tenant <Azure tenant ID>
86+
az account set -s <subscription ID>
87+
az aksarc show -g <resource_group_name> -n <cluster_name>
88+
```
89+
90+
Check the storage profile section:
91+
92+
```json
93+
"storageProfile": {
94+
"nfsCsiDriver": {
95+
"enabled": false
96+
},
97+
"smbCsiDriver": {
98+
99+
"enabled": true
100+
}
101+
}
102+
```
103+
104+
If one of the drivers is disabled, you can enable it using one of the following commands:
105+
106+
```azurecli
107+
az aksarc update --enable-smb-driver -g <resource_group_name> -n <cluster_name>
108+
az aksarc update --enable-nfs-driver -g <resource_group_name> -n <cluster_name>
109+
```
110+
111+
Running the `aksarc update` command should resolve the issue and the `currentState` property of the cluster should now show as **Succeeded**. Once the status is updated, if you don't want to keep the drivers enabled, you can reverse this action by running one of the following commands:
112+
113+
```azurecli
114+
az aksarc update --disable-smb-driver -g <resource_group_name> -n <cluster_name>
115+
az aksarc update --disable-nfs-driver -g <resource_group_name> -n <cluster_name>
116+
```
117+
118+
If both drivers are already enabled on your cluster, you can disable the one that's not in use. If you require both drivers to remain enabled, contact Microsoft Support for further assistance.
119+
120+
## Verification
121+
122+
To confirm the K8s version upgrade is complete, run the following command and check that the `currentState` property in the JSON output is set to **Succeeded**.
123+
124+
```azurecli
125+
az aksarc show -g <resource_group> -n <cluster_name>
126+
```
127+
128+
```output
129+
...
130+
...
131+
"provisioningState": "Succeeded",
132+
"status": {
133+
"currentState": "Succeeded",
134+
"errorMessage": null,
135+
"operationStatus": null
136+
"controlPlaneStatus": { ...
137+
...
138+
```
139+
140+
## Contact Microsoft Support
141+
142+
If the problem persists, collect the [AKS cluster logs](get-on-demand-logs.md) before you [create a support request](aks-troubleshoot.md#open-a-support-request).
143+
144+
## Next steps
145+
146+
- [Use the diagnostic checker tool to identify common environment issues](aks-arc-diagnostic-checker.md)
147+
- [Review AKS on Azure Local architecture](cluster-architecture.md)

AKS-Arc/entra-prompts.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ This issue is caused by [a GitHub bug](https://github.com/Azure/kubelogin/issues
2626

2727
To mitigate this issue, you can use one of the following two methods:
2828

29-
- Downgrade **kubelogin** to version 1.9.0. This stable version does not have the bug that causes repeated authentication prompts. You can [download this version from the GitHub repository](https://github.com/int128/kubelogin/releases/tag/v1.9.0). Select the appropriate asset for your OS or architecture, extract it, and replace your existing **kubelogin** binary.
29+
- Downgrade **kubelogin** to version 0.1.9. This stable version does not have the bug that causes repeated authentication prompts. You can [download this version from the GitHub repository](https://github.com/Azure/kubelogin/releases/tag/v0.1.9). Select the appropriate asset for your OS or architecture, extract it, and replace your existing **kubelogin** binary.
3030
- Alternatively, if you have administrator permissions, you can use the `--admin` flag with the `az aksarc get-credentials` command. This method bypasses **kubelogin** authentication by retrieving admin credentials directly:
3131

3232
```azurecli

azure-local/faq.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -44,7 +44,7 @@ sections:
4444
4545
- question: Are there bandwidth or latency requirements between Azure Local and the cloud?
4646
answer: |
47-
No. Limited-bandwidth connections like rural T1 lines or satellite/cellular connections are adequate for Azure Local to sync. The minimum required connectivity is several kilobytes per day. More services might require extra bandwidth, especially to replicate or back up whole VMs, download large software updates, or upload verbose logs for analysis and monitoring in the cloud.
47+
Yes. Limited-bandwidth connections like rural T1 lines or satellite/cellular connections are adequate for Azure Local to sync. The minimum required connectivity is 10 Mb per day. More services might require extra bandwidth, especially to replicate or back up whole VMs, download large software updates, or upload verbose logs for analysis and monitoring in the cloud.
4848
4949
- question: Does Azure Local require continuous connectivity to the cloud?
5050
answer: |

azure-local/manage/add-server.md

Lines changed: 3 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Learn how to manage capacity on your Azure Local, version 23H2 syst
44
ms.topic: how-to
55
author: alkohli
66
ms.author: alkohli
7-
ms.date: 05/28/2025
7+
ms.date: 06/26/2025
88
---
99

1010
# Add a node on Azure Local
@@ -110,25 +110,13 @@ On the new node that you plan to add, follow these steps.
110110
111111
3. Assign the following permissions to the newly added nodes:
112112

113-
- Azure Local Device Management Role
113+
- Azure Stack HCI Device Management Role
114114
- Key Vault Secrets User
115115
For more information, see [Assign permissions to the node](../deploy/deployment-arc-register-server-permissions.md).
116116

117117
On a node that already exists on your system, follow these steps:
118118

119-
1. Sign in with the domain user credentials (AzureStackLCMUser or another user with equivalent permissions) that you provided during the deployment of the system.
120-
121-
1. (Optional) Before you add the node, make sure to get an updated authentication token. Run the following command:
122-
123-
```powershell
124-
Update-AuthenticationToken
125-
```
126-
127-
1. If you are running a version prior to 2405.3, you must run the following command on the new node to clean up conflicting files:
128-
129-
```powershell
130-
Get-ChildItem -Path "$env:SystemDrive\NugetStore" -Exclude Microsoft.AzureStack.Solution.LCMControllerWinService*,Microsoft.AzureStack.Role.Deployment.Service* | Remove-Item -Recurse -Force
131-
```
119+
1. Sign in with the domain user credentials (AzureStackLCMUser or another user with equivalent permissions) that you provided during the deployment of the system.
132120

133121
1. Run the following command to add the new incoming node using a local administrator credential for the new node:
134122

azure-local/manage/repair-server.md

Lines changed: 2 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Learn how to repair a node on your Azure Local, version 23H2 system
44
ms.topic: how-to
55
author: alkohli
66
ms.author: alkohli
7-
ms.date: 05/08/2025
7+
ms.date: 06/26/2025
88
---
99

1010
# Repair a node on Azure Local
@@ -134,19 +134,12 @@ Follow these steps on the node you're trying to repair.
134134
135135
1. Assign the following permissions to the repaired node:
136136

137-
- Azure Local Device Management Role
137+
- Azure Stack HCI Device Management Role
138138
- Key Vault Secrets User
139139
For more information, see [Assign permissions to the node](../deploy/deployment-arc-register-server-permissions.md).
140140

141141
Follow these steps on another node that is a member of the same Azure Local instance.
142142

143-
144-
1. If you are running a version prior to 2405.3, you must run the following command to clean up conflicting files:
145-
146-
```powershell
147-
Get-ChildItem -Path "$env:SystemDrive\NugetStore" -Exclude Microsoft.AzureStack.Solution.LCMControllerWinService*,Microsoft.AzureStack.Role.Deployment.Service* | Remove-Item -Recurse -Force
148-
```
149-
150143
1. Sign into the node that is already a member of the system, with the domain user credentials that you provided during the deployment of the system. Run the following command to repair the incoming node:
151144

152145
```powershell

azure-local/migrate/migrate-vmware-requirements.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ title: Review requirements for VMware VM migration to Azure Local using Azure Mi
33
description: Learn the system requirements for VMware migration to Azure Local using Azure Migrate (preview).
44
author: alkohli
55
ms.topic: how-to
6-
ms.date: 05/16/2025
6+
ms.date: 06/19/2025
77
ms.author: alkohli
88
ms.custom: references_regions
99
---
@@ -67,8 +67,6 @@ For any subscriptions hosting resources used in migration, such as Azure Migrate
6767

6868
- Ensure that vCenter Server permission requirements are met. For more information, see [VMware vSphere requirements (agentless)](/azure/migrate/migrate-support-matrix-vmware-migration#agentless-migration).
6969

70-
- In this release, you can only migrate VMs that have disks attached to the VMFS Datastores. If the VM disks aren't attached to the VMFS Datastore, the disks can’t be migrated.
71-
7270
- Before you begin, for all VMware VMs, bring all the disks online and persist the drive letter. For more information, see how to [configure a SAN policy](/azure/migrate/prepare-for-migration#configure-san-policy) to bring the disks online.
7371

7472
- The VMware source environment must be able to initiate a network connection with the target Azure Local instance, either by being on the same on-premises network or by using a VPN.

azure-stack/includes/app-service-hub-update-banner.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ author: apwestgarth
33
ms.author: anwestg
44
ms.service: azure-stack
55
ms.topic: include
6-
ms.date: 01/31/2025
6+
ms.date: 06/25/2025
77
ms.reviewer: sethm
88
ms.lastreviewed: 10/28/2020
99
---
@@ -13,4 +13,4 @@ ms.lastreviewed: 10/28/2020
1313
>
1414
> | Supported minimum Azure Stack Hub version | App Service RP version |
1515
> |-----|---|
16-
> | 2311 and later | 24R1 [Installer](https://aka.ms/appsvcupdate24R1installer) ([release notes](../operator/app-service-release-notes-2024R1.md)) |
16+
> | 2311 and later | 25R1 [Installer](https://aka.ms/appsvcupdate25R1installer) 25R1 [Offline Package](https://aka.ms/appsvcupdate25R1offline) ([release notes](../operator/app-service-release-notes-2025R1.md)) |

azure-stack/operator/TOC.yml

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -386,16 +386,10 @@
386386
href: app-service-migrate-sql-server.md
387387
- name: Release notes
388388
items:
389+
- name: 25R1 release notes [June 2025]
390+
href: app-service-release-notes-2025R1.md
389391
- name: 24R1 release notes [December 2024]
390392
href: app-service-release-notes-2024R1.md
391-
- name: 2302 release notes [February 2023]
392-
href: app-service-release-notes-2302.md
393-
- name: 2022 H1 release notes [October 2022]
394-
href: app-service-release-notes-2022-h1.md
395-
- name: 2021 Q3 release notes [December 2021]
396-
href: app-service-release-notes-2021-Q3.md
397-
- name: 2021 Q1 release notes [June 2021]
398-
href: app-service-release-notes-2021-Q1.md
399393
- name: Azure Container Registry
400394
items:
401395
- name: Overview

0 commit comments

Comments
 (0)