We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Describe scenario We are running multiple AKS and we have weekly automatic patching enabled.
Some of our critical pod have a PDB configured.
In some rate condition, the automatic update failed. The reason is that there is a timeout. A configured PDB, deny to drain a node.
Question
To recover from that situation, i have to manually restart the pod. Thats fine.
However. The node pool remains in Failed state, including the extra nodes remains as well.
How I can recover from that state? How I can re-trigger the automatic update?
One solution is manually delete the old VM from the VMSS. But thats kinda tricky on Node Pools which large amount of nodes.
The text was updated successfully, but these errors were encountered:
az aks update -n $aks -g $rG
No other arguments/parameters.
Sorry, something went wrong.
So is no Portal Experience, right? I can't trigger update at portal, if its on the latest version.
From document, yes. Maybe someone else knows it is hidden somewhere.
https://learn.microsoft.com/troubleshoot/azure/azure-kubernetes/availability-performance/cluster-node-virtual-machine-failed-state
Thanks, I will try that on next incident!
No branches or pull requests
Describe scenario
We are running multiple AKS and we have weekly automatic patching enabled.
Some of our critical pod have a PDB configured.
In some rate condition, the automatic update failed. The reason is that there is a timeout. A configured PDB, deny to drain a node.
Question
To recover from that situation, i have to manually restart the pod. Thats fine.
However. The node pool remains in Failed state, including the extra nodes remains as well.
How I can recover from that state? How I can re-trigger the automatic update?
One solution is manually delete the old VM from the VMSS. But thats kinda tricky on Node Pools which large amount of nodes.
The text was updated successfully, but these errors were encountered: