-
Notifications
You must be signed in to change notification settings - Fork 195
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remove backends from backend services before updating #643
Comments
This issue is currently awaiting triage. If the repository mantainers determine this is a relevant issue, they will accept it by applying the The Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
hi @dippynark , could you expand on this motivation to understand better what you are trying to achieve so we can offer you the best solution? there may be a better solution than having to exclude the nodes with labels :) |
I am occasionally seeing timeouts to a Service of type LoadBalancer on GKE (internal network load balancer with unmanaged instance groups for backends) when I label a Node with
node.kubernetes.io/exclude-from-external-load-balancers: true
and then delete the underlying instance after its health status disappears from the backend service (I am doing this to periodically force the cluster autoscaler to kick in in the hope that it'll replace on-demand VMs with spot VMs).This only seems to happen when I remove the last instance in a zone and the corresponding unmanaged instance group is removed as a backend from the backend service. I would instead expect connection draining to apply (0 seconds by default) and close or reset any in-flight connections, but I think this is happening because the backend service is being updated rather than the instance group being explicitly removed; according to the connection draining documentation this is not one of the events that triggers connection draining.
A previous version of the backend service documentation said
it can take several minutes for your changes to propagate throughout the network
and there does not seem to be a way to monitor this propagation through the API so maybe I am not waiting long enough before deleting an instance?I have found it hard to debug this scenario properly so I'm not confident that this is the cause, but does this sound like a possible bug? If so, would explicitly removing instance groups that are no longer specified as backends be an improvement to ensure connection draining applies in this situation?
The text was updated successfully, but these errors were encountered: