This message was deleted.
# general
a
This message was deleted.
c
I am assuming you’re talking about a single-node cluster, and not snapshotting random workers in a larger cluster?
h
yeah that’s correct
c
well the node is Ready when you snapshot it, right?
When you bring it back up, the node is still in the datastore as Ready
h
right
c
It won’t get moved to NotReady until the controller manager comes up and notices that the kubelet isn’t reporting in
so it’ll go to NotReady as the kubelet initializes, and then back to Ready
h
I’m not surprised that it goes into a NotReady state intermittently, but once it does it takes like 5-15 minutes to go back into a Ready state
c
You could probably check the kubelet logs or just describe it to see what’s blocking readiness
I would probably recommend against snapshotting the whole node though, in favor of using etcd and taking etcd snapshots
unless for some reason you have other stuff on there that you’re trying to capture in the snapshot, and you’re OK with a potentially inconsistent datastore state
h
makes sense; and yeah I totally understand etcd snapshots are preferable just not 100% feasible in our slightly-odd use case (effectively using k3s as a replacement for a single-node docker compose type thing)