Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

S3 snapshots missing from etcdsnapshotfile list / k3s-etcd-snapshots configmap when multiple nodes save at the same time #10989

Open
brandond opened this issue Oct 3, 2024 · 0 comments
Assignees

Comments

@brandond
Copy link
Member

brandond commented Oct 3, 2024

K3s tracking issue for:

This is extremely difficult to trigger outside of Rancher. When the user triggers an etcd snapshot via Rancher, all etcd nodes simultaneously save snapshots and attempt to reconcile their local snapshots, and the snapshots they see on S3, with ETCDSnapshotFile resources. Since there is a slight time gap in between when objects are listed from S3, and when ETCDSnapshotFiles resources are listed, not all cluster members will have a consistent view of objects in both stores, and may end up deleting ETCDSnapshotFile resources for snapshots that other nodes just finished uploading. The resources will be recreated next time S3 is reconciled (the next time snapshots are saved/deleted/pruned) but this does cause a temporary mismatch between what shows up in kubectl get etcdsnapshotfile and kubectl get configmap -n kube-system k3s-etcd-snapshots and what shows up in k3s etcd-snapshot ls (and what's actually in S3).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: To Test
Development

No branches or pull requests

2 participants