This message was deleted.
# general
This message was deleted.
If this is RKE1, and in Rancher afaics, you have to trigger cluster provisioning to let RKE check cluster state and fix any missing containers. You can easily do this by editing the cluster as YAML and changing
Copy code
as that does not actually change anything in the cluster but does trigger provisioning. If the cluster is broken beyond repair, you can try to restore a backup. As its just a
docker run
of a container, you can also figure it out yourself and run it manually but it needs a bit more context and preparation to be able to do that (what versions, what config etc)
I tried to find out, how these containers started using below command and output is there. Can I use these details to start "kube-controller-manager" container or not? docker ps --no-trunc | grep kube-controller-manager 864328ddb7fc4c6c08baf2c98e8aaad5f68251ce934c345a6c0cac431dbba05a rancher/hyperkube:v1.22.9-rancher1 "/opt/rke-tools/ kube-controller-manager --service-account-private-key-file=/etc/kubernetes/ssl/kube-service-account-token-key.pem --allocate-node-cidrs=true --leader-elect=true --authorization-kubeconfig=/etc/kubernetes/ssl/kubecfg-kube-controller-manager.yaml --service-cluster-ip-range= --node-monitor-grace-period=40s --root-ca-file=/etc/kubernetes/ssl/kube-ca.pem --profiling=false --address= --allow-untagged-cloud=true --authentication-kubeconfig=/etc/kubernetes/ssl/kubecfg-kube-controller-manager.yaml --configure-cloud-routes=false --v=2 --cloud-provider= --cluster-cidr= --kubeconfig=/etc/kubernetes/ssl/kubecfg-kube-controller-manager.yaml --enable-hostpath-provisioner=false --pod-eviction-timeout=5m0s --terminated-pod-gc-threshold=1000 --use-service-account-credentials=true"
I dont want to clean up the Node and Join again as of now, need to know other options to start "kube-controller-manager" container .