Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] The number of unavailable instances may exceed the maxUnavailable limit during upgrades #8078

Open
YTGhost opened this issue Sep 3, 2024 · 1 comment
Assignees
Labels
kind/bug Something isn't working ks Stale

Comments

@YTGhost
Copy link
Contributor

YTGhost commented Sep 3, 2024

Describe the bug
When performing an upgrade, if the user upgrades to a faulty image causing the Container to crash shortly thereafter, during this brief period (before the Container crashes), since the Pod has already been upgraded, the next Pod upgrade might be permitted (with currentUnavailable being 0 and the previous Pod being checked as healthy). This can lead to the final number of unavailable instances exceeding the maxUnavailable limit.

Expected behavior
In addition to the isHealthy check, it may also be necessary to rely on MinReadySeconds for the isRunningAndAvailable check. At the same time, we also need to ensure that the call to isRunningAndAvailable happens after the kubelet has completed updating the status.

@YTGhost YTGhost added the kind/bug Something isn't working label Sep 3, 2024
@free6om free6om added the ks label Sep 3, 2024
Copy link

github-actions bot commented Oct 7, 2024

This issue has been marked as stale because it has been open for 30 days with no activity

@github-actions github-actions bot added the Stale label Oct 7, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/bug Something isn't working ks Stale
Projects
None yet
Development

No branches or pull requests

3 participants