adamant-kite-43734
08/09/2023, 9:00 PMcreamy-pencil-82913
08/10/2023, 9:46 PMcreamy-pencil-82913
08/10/2023, 9:47 PMcreamy-pencil-82913
08/10/2023, 9:47 PMmany-nightfall-61858
08/10/2023, 9:48 PMmany-nightfall-61858
08/10/2023, 9:55 PMmany-nightfall-61858
08/10/2023, 9:56 PMrke2 server --help | grep enable-servicelb
many-nightfall-61858
08/29/2023, 12:56 AMcloud-provider-name: aws
cloud-provider-config: "/etc/rancher/rke2/cloud.conf" # only EC2 autoscale is setup
enable-servicelb: true
Then cloud-controller-manager
goes in a restart loop:
NAME READY STATUS RESTARTS AGE
cloud-controller-manager-ip-x-x-x-x.region.compute.internal 0/1 Running 3 (64s ago) 13m
cloud-controller-manager-ip-y-y-y-y.region.compute.internal 0/1 Running 3 (57s ago) 12m
cloud-controller-manager-ip-z-z-z-z.region.compute.internal 0/1 Running 6 (90s ago) 25m
LoadBalancer services are still able to be created however, and I am able to connect to those services. But this pod never transitions to a Ready state. When I disable cloud-provider, it goes into a Ready state. Is that expected?
I also tried setting
disable-cloud-controller: true
but cloud-controller-manager
still runs with the restart behavior. The events and logs dont seem to indicate much except for this. I cant find much online about it. So was wondering if you might know what the problem is?
Warning Unhealthy 72s (x30 over 6m2s) kubelet Startup probe failed: Get "<https://localhost:10258/healthz>": dial tcp [::1]:10258: connect: connection refused
creamy-pencil-82913
08/29/2023, 1:09 AMcreamy-pencil-82913
08/29/2023, 1:14 AMcreamy-pencil-82913
08/29/2023, 1:15 AMmany-nightfall-61858
08/29/2023, 4:49 PMI0829 16:39:51.868208 1 controllermanager.go:152] Version: v1.26.3-k3s1
I0829 16:39:51.868715 1 leaderelection.go:248] attempting to acquire leader lease kube-system/rke2-cloud-controller-manager...
many-nightfall-61858
08/29/2023, 4:50 PMcreamy-pencil-82913
08/29/2023, 4:51 PMmany-nightfall-61858
08/29/2023, 4:59 PMI0829 16:48:55.909440 1 controllermanager.go:152] Version: v1.26.3-k3s1
I0829 16:48:55.909885 1 leaderelection.go:248] attempting to acquire leader lease kube-system/rke2-cloud-controller-manager...
I0829 16:49:11.826660 1 leaderelection.go:258] successfully acquired lease kube-system/rke2-cloud-controller-manager
I0829 16:49:11.826838 1 event.go:294] "Event occurred" object="kube-system/rke2-cloud-controller-manager" fieldPath="" kind="Lease" apiVersion="<http://coordination.k8s.io/v1|coordination.k8s.io/v1>" type="Normal" reason="LeaderElection" message="ip.compute.internal_3be410ca-ed7d-42cf-a20e-5705884b82d7 became leader"
time="2023-08-29T16:49:11Z" level=info msg="Creating service-controller event broadcaster"
time="2023-08-29T16:49:12Z" level=info msg="Starting /v1, Kind=Node controller"
time="2023-08-29T16:49:12Z" level=info msg="Starting /v1, Kind=Pod controller"
time="2023-08-29T16:49:12Z" level=info msg="Starting apps/v1, Kind=DaemonSet controller"
time="2023-08-29T16:49:12Z" level=info msg="Starting <http://discovery.k8s.io/v1|discovery.k8s.io/v1>, Kind=EndpointSlice controller"
I0829 16:49:12.446502 1 controllermanager.go:311] Started "service"
W0829 16:49:12.446517 1 controllermanager.go:288] "route" is disabled
W0829 16:49:12.446521 1 controllermanager.go:288] "cloud-node" is disabled
W0829 16:49:12.446525 1 controllermanager.go:288] "cloud-node-lifecycle" is disabled
I0829 16:49:12.446622 1 controller.go:227] Starting service controller
I0829 16:49:12.446639 1 shared_informer.go:273] Waiting for caches to sync for service
I0829 16:49:12.547574 1 shared_informer.go:280] Caches are synced for service
I0829 16:49:12.547735 1 event.go:294] "Event occurred" object="default/hello" fieldPath="" kind="Service" apiVersion="v1" type="Normal" reason="EnsuringLoadBalancer" message="Ensuring load balancer"
I0829 16:49:12.551680 1 event.go:294] "Event occurred" object="default/hello" fieldPath="" kind="Service" apiVersion="v1" type="Normal" reason="AppliedDaemonSet" message="Applied LoadBalancer DaemonSet kube-system/svclb-hello-9329e302"
they’re each in a CrashLoop now since yesterday. once the leader restarts, one of the other replicas acquires the lease. there doesnt seem to be a log indicating what caused the previous crashcreamy-pencil-82913
08/29/2023, 5:08 PMkubectl logs --previous
to look at the logs from the crashed podcreamy-pencil-82913
08/29/2023, 5:09 PMmany-nightfall-61858
08/29/2023, 5:14 PMcreamy-pencil-82913
08/29/2023, 5:30 PMcreamy-pencil-82913
08/29/2023, 5:30 PMmany-nightfall-61858
08/30/2023, 11:35 PMstatus:
conditions:
- lastProbeTime: null
lastTransitionTime: "2023-08-30T23:28:39Z"
status: "True"
type: Initialized
- lastProbeTime: null
lastTransitionTime: "2023-08-30T23:28:39Z"
message: 'containers with unready status: [cloud-controller-manager]'
reason: ContainersNotReady
status: "False"
type: Ready
- lastProbeTime: null
lastTransitionTime: "2023-08-30T23:28:39Z"
message: 'containers with unready status: [cloud-controller-manager]'
reason: ContainersNotReady
status: "False"
type: ContainersReady
- lastProbeTime: null
lastTransitionTime: "2023-08-30T23:28:39Z"
status: "True"
type: PodScheduled
many-nightfall-61858
08/30/2023, 11:39 PM$ kubectl -n kube-system logs cloud-controller-manager-a
I0830 23:33:33.630303 1 controllermanager.go:152] Version: v1.26.3-k3s1
I0830 23:33:33.630806 1 leaderelection.go:248] attempting to acquire leader lease kube-system/rke2-cloud-controller-manager...
$ kubectl -n kube-system logs cloud-controller-manager-b
I0830 23:32:41.339963 1 controllermanager.go:152] Version: v1.26.3-k3s1
I0830 23:32:41.340429 1 leaderelection.go:248] attempting to acquire leader lease kube-system/rke2-cloud-controller-manager...
I0830 23:32:58.320255 1 leaderelection.go:258] successfully acquired lease kube-system/rke2-cloud-controller-manager
I0830 23:32:58.320324 1 event.go:294] "Event occurred" object="kube-system/rke2-cloud-controller-manager" fieldPath="" kind="Lease" apiVersion="<http://coordination.k8s.io/v1|coordination.k8s.io/v1>" type="Normal" reason="LeaderElection" message="ip-_0328c001-e000-45b2-9200-b49b119cdb71 became leader"
time="2023-08-30T23:32:58Z" level=info msg="Creating service-controller event broadcaster"
time="2023-08-30T23:32:58Z" level=info msg="Starting /v1, Kind=Node controller"
time="2023-08-30T23:32:59Z" level=info msg="Starting /v1, Kind=Pod controller"
time="2023-08-30T23:32:59Z" level=info msg="Starting apps/v1, Kind=DaemonSet controller"
W0830 23:32:59.691042 1 controllermanager.go:288] "cloud-node-lifecycle" is disabled
time="2023-08-30T23:32:59Z" level=info msg="Starting <http://discovery.k8s.io/v1|discovery.k8s.io/v1>, Kind=EndpointSlice controller"
I0830 23:32:59.691346 1 controllermanager.go:311] Started "service"
W0830 23:32:59.691356 1 controllermanager.go:288] "route" is disabled
W0830 23:32:59.691360 1 controllermanager.go:288] "cloud-node" is disabled
I0830 23:32:59.691485 1 controller.go:227] Starting service controller
I0830 23:32:59.691500 1 shared_informer.go:273] Waiting for caches to sync for service
I0830 23:32:59.791739 1 shared_informer.go:280] Caches are synced for service
I0830 23:32:59.791908 1 event.go:294] "Event occurred" object="default/hello" fieldPath="" kind="Service" apiVersion="v1" type="Normal" reason="EnsuringLoadBalancer" message="Ensuring load balancer"
I0830 23:32:59.795705 1 event.go:294] "Event occurred" object="default/hello" fieldPath="" kind="Service" apiVersion="v1" type="Normal" reason="AppliedDaemonSet" message="Applied LoadBalancer DaemonSet kube-system/svclb-hello-9c549e63"
[anduril@ip-10-32-57-32 ~]$ kubectl -n kube-system logs cloud-controller-manager-3
I0830 23:33:46.427698 1 controllermanager.go:152] Version: v1.26.3-k3s1
I0830 23:33:46.428187 1 leaderelection.go:248] attempting to acquire leader lease kube-system/rke2-cloud-controller-manager...
creamy-pencil-82913
08/31/2023, 4:44 AMmany-nightfall-61858
08/31/2023, 5:09 PMnodeSelector:
<http://node-role.kubernetes.io/control-plane|node-role.kubernetes.io/control-plane>: "true"
args:
- --configure-cloud-routes=false
- --cloud-provider=aws
many-nightfall-61858
08/31/2023, 5:10 PM