Skip to content

Commit fc20e2e

Browse files
committed
Acrolinx improvements
1 parent 76a9c9d commit fc20e2e

File tree

1 file changed

+7
-4
lines changed

1 file changed

+7
-4
lines changed

articles/aks/node-auto-repair.md

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -31,13 +31,13 @@ If AKS identifies an unhealthy node that remains unhealthy for *five* minutes, A
3131
2. If the node restart is unsuccessful, AKS reimages the node.
3232
3. If the reimage is unsuccessful and it's a Linux node, AKS redeploys the node.
3333

34-
Alternative remediations are investigated by AKS engineers if auto-repair is unsuccessful.
34+
AKS engineers investigate alternative remediations if auto-repair is unsuccessful.
3535

3636
If you want the remediator to reimage the node, you can add the `nodeCondition "customerMarkedAsUnhealthy": true`.
3737

3838
## Node auto-drain
3939

40-
[Scheduled events][scheduled-events] can occur on the underlying VMs in any of your node pools. For [spot node pools][spot-node-pools], scheduled events may cause a *preempt* node event for the node. Certain node events, such as *preempt*, cause AKS node auto-drain to attempt a cordon and drain of the affected node. This process allows for a reschedule of any affected workloads on that node. When this happens, you might notice the node receives a taint with `"remediator.aks.microsoft.com/unschedulable"`, because of `"kubernetes.azure.com/scalesetpriority: spot"`.
40+
[Scheduled events][scheduled-events] can occur on the underlying VMs in any of your node pools. For [spot node pools][spot-node-pools], scheduled events may cause a *preempt* node event for the node. Certain node events, such as *preempt*, cause AKS node auto-drain to attempt a cordon and drain of the affected node. This process enables rescheduling for any affected workloads on that node. You might notice the node receives a taint with `"remediator.aks.microsoft.com/unschedulable"`, because of `"kubernetes.azure.com/scalesetpriority: spot"`.
4141

4242
The following table shows the node events and actions they cause for AKS node auto-drain:
4343

@@ -47,11 +47,14 @@ The following table shows the node events and actions they cause for AKS node au
4747
| Reboot | The VM is scheduled for reboot. The VM's non-persistent memory is lost. | No action. |
4848
| Redeploy | The VM is scheduled to move to another node. The VM's ephemeral disks are lost. | Cordon and drain. |
4949
| Preempt | The spot VM is being deleted. The VM's ephemeral disks are lost. | Cordon and drain |
50-
| Terminate | The VM is scheduled to be deleted.| Cordon and drain. |
50+
| Terminate | The VM is scheduled for deletion.| Cordon and drain. |
5151

5252
## Limitations
5353

54-
In many cases, AKS can determine if a node is unhealthy and attempt to repair the issue. However, there are cases where AKS either can't repair the issue or detect that an issue exists. For example, AKS can't detect issues if a node status isn't being reported due to error in network configuration or has failed to initially register as a healthy node.
54+
In many cases, AKS can determine if a node is unhealthy and attempt to repair the issue. However, there are cases where AKS either can't repair the issue or detect that an issue exists. For example, AKS can't detect issues in the following example scenarios:
55+
56+
* A node status isn't being reported due to error in network configuration.
57+
* A node failed to initially register as a healthy node.
5558

5659
## Next steps
5760

0 commit comments

Comments
 (0)