-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
sunset and remove https://github.com/kubernetes/enhancements/blob/master/keps/sig-cluster-lifecycle/kubeadm/0004-bootstrap-checkpointing.md #1103
Comments
/sig cluster-lifecycle |
/assign |
And in the future we might want checkpointing to survive DC outage, but the current design was somewhat weak (since it was limited). This allows us to take more time to come back and do checkpointing in a broader sense without having the urgency to make it work for kubeadm (even though it may be valuable). |
@timothysc @smarterclayton @yastij I'm collecting enhancements to be tracked for the 1.16 release cycle. Can someone shed some light if this is an actual enhancement that needs to be tracked? KEP? etc etc. Thanks! |
@kacole2 |
This would be a step back IMO. I've wanted to use it to checkpoint a few things like haproxy or keepalived for providing the load balancing service kubeadm is currently lacking. But unfortunately the issue #72202 is still blocking that. Its a simple change to get things working. Then it can be iterated on. Please don't rip it out because you don't want to use it. Others do though. |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
/remove-lifecycle stale |
Just putting it out there: Shouldn't we instead drop support for single-node control planes? At least for non-test/toy scenarios. ie: just don't attempt to provide automatic recovery from a full DC outage for single-node clusters. Ditto for any other feature that falls outside a simple "kind" test usecase, like upgrading existing clusters. It seems peculiar to make HA meet the requirements of non-HA, rather than the other way around. |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
/remove-lifecycle stale |
going back to this topic. overall self-hosting is mostly out of scope for SIG Cluster Lifecycle at this point. one exception is boot-kube, but they have not shown interest in this proposal. given the feature was not successful we should revert the changes in the kubelet introduced by kubernetes/kubernetes#50984 and let another party take over this topic if they want to. the discussion here was suggesting that: i have logged this to get help from contributors with the cleanup:
statistically, from what i've seen from recent user reports, there are still single-node production clusters out there, due to a variety of factors including budget. i have not seen any requests for checkpointing or self-hosting recently, though.
kubeadm has no plans to provide a load balancer by default. we do have some updated documentation though: it still feels right to remove the existing partial feature, as is and let another SIG / different contributors take over by writing a new KEP. |
actually....looks like someone sent the PR already as part of the kubelet flag cleanup: we have too many tickets for this at this point, so i'm going to close this one. |
During earlier phases of kubeadm's life we wanted to pursue the option of full self hosted control plane. This idea breaks down during certain corner conditions, namely single control plane node cluster on a full DC outage. The idea was explored but eventually abandoned due to security constraints involved in checkpointing, but also b/c other ideas obviated the need to manage k8s in this manner.
/cc @yastij
The text was updated successfully, but these errors were encountered: