abundant-hair-58573
07/09/2025, 8:54 PMcreamy-pencil-82913
07/09/2025, 9:04 PMkubectl get nodes
), not the IP. Did something happen that caused the node name to change?creamy-pencil-82913
07/09/2025, 9:05 PMabundant-hair-58573
07/09/2025, 9:06 PMcreamy-pencil-82913
07/09/2025, 9:07 PMkubectl get nodes
?abundant-hair-58573
07/09/2025, 9:10 PM/etc/rancher/rke2/config.yaml.d/99-aws-id.yaml
with
kubelet-arg+:
- --hostname-override=ip-xxx-xxx-xxx-xxx.us-iso-east-1.compute.internal
kube-proxy-arg+:
- --hostname-override=ip-xxx-xxx-xxx-xxx.us-iso-east-1.compute.internal
node-name: ip-xxx-xxx-xxx-xxx.us-iso-east-1.compute.internal
node-label+
- node-type=controlplane
creamy-pencil-82913
07/09/2025, 9:10 PMcreamy-pencil-82913
07/09/2025, 9:10 PMnode-name: xxx
option is for. If you just go poking at the hostname override in individual component args, rke2 itself will not be aware of that.creamy-pencil-82913
07/09/2025, 9:11 PMabundant-hair-58573
07/09/2025, 9:12 PMabundant-hair-58573
07/09/2025, 9:12 PMcreamy-pencil-82913
07/09/2025, 9:13 PMabundant-hair-58573
07/09/2025, 9:15 PMabundant-hair-58573
07/09/2025, 9:20 PMcreamy-pencil-82913
07/09/2025, 9:41 PMcreamy-pencil-82913
07/09/2025, 9:42 PMip-xxx-xxx-xxx-xxx.us-iso-east-1.compute.internal
but the log says it is looking for ip-xxx-xxx-xxx-xxx
without the fqdn?abundant-hair-58573
07/10/2025, 2:09 PM/etc/rancher/rke2/config.yaml.d/99-aws-id.yaml
kubelet-arg+:
- --hostname-override=ip-xxx-xxx-xxx-xxx.ec2.internal
kube-proxy-arg+:
- --hostname-override=ip-xxx-xxx-xxx-xxx.ec2.internal
node-name: ip-xxx-xxx-xxx-xxx.ec2.internal
This is from the rke2-server log on the control plane that's trying to upgrade
rke2[27169]: time="2025-07-10T14:05:26Z" level=info msg="Waiting for control-plane node ip-xxx-xxx-xxx-xxx.domain.org startup: nodes \"ip-xxx-xxx-xxx-xxx.domain.org\" not found"
This is going form 1.28.15 to 1.29.15. Our cloud controller manager was still at 1.27.x from our initial install, I didn't upgrade that way back when I upgraded from 1.27.x to 1.28.15. First thing I did here was upgrade the cloud controller manager to 1.28.11, then I just added the plan to the SUC
# Server plan
apiVersion: upgrade.cattle.io/v1
kind: Plan
metadata:
name: controlplane-plan-v1-29-15
namespace: cattle-system
labels:
rke2-upgrade: controlplane
spec:
concurrency: 1
nodeSelector:
matchExpressions:
- {key: node-role.kubernetes.io/control-plane, operator: In, values: ["true"]}
tolerations:
- key: "node-role.kubernetes.io/control-plane"
operator: "Equal"
effect: "NoSchedule"
- key: "CriticalAddonsOnly"
operator: "Equal"
value: "true"
effect: "NoExecute"
serviceAccountName: system-upgrade-controller
cordon: true
upgrade:
image: rancher/rke2-upgrade
version: v1.29.15+rke2r1
abundant-hair-58573
07/10/2025, 2:21 PMI0710 14:18:01.299173 27479 status_manager.go:877] "Failed to update status for pod" pod="kube-system/kube-proxy-ip-xxx-xxx-xxx-xxx.ec2.internal" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1c6fb9d-34b3-45a5-9adf-d50451828562\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-07-10T14:04:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-07-10T14:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-07-10T14:04:49Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"<containerd://ca9ff807b8758ff432cb1d5b355dc79259311198edad8f4de046885f376b46d>5\\\",\\\"image\\\":\\\"<http://docker-remote.artifactory.domain.org/rancher/hardened-kubernetes:v1.29.15-rke2r1-build20250312\\\|docker-remote.artifactory.domain.org/rancher/hardened-kubernetes:v1.29.15-rke2r1-build20250312\\\>",\\\"imageID\\\":\\\"<http://docker-remote.artifactory.domain.org.org/rancher/hardened-kubernetes@sha256:34aaaf8700ef979929c3b1dbfb2d8de2b25c00a68a6a6b540293d6f576cb89fd\\\|docker-remote.artifactory.domain.org.org/rancher/hardened-kubernetes@sha256:34aaaf8700ef979929c3b1dbfb2d8de2b25c00a68a6a6b540293d6f576cb89fd\\\>",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-07-10T14:04:30Z\\\"}}}],\\\"hostIPs\\\":[{\\\"ip\\\":\\\"10.114.49.20\\\"}]}}\" for pod \"kube-system\"/\"kube-proxy-ip-xxx-xxx-xxx-xxx.ec2.internal\": pods \"kube-proxy-ip-xxx-xxx-xxx-xxx.ec2.internal\" is forbidden: node \"<http://ip-xxx-xxx-xxx-xxx.domain.org|ip-xxx-xxx-xxx-xxx.domain.org>\" can only update pod status for pods with spec.nodeName set to itself"
abundant-hair-58573
07/10/2025, 4:04 PMabundant-hair-58573
07/10/2025, 5:15 PMcreamy-pencil-82913
07/10/2025, 5:18 PMabundant-hair-58573
07/10/2025, 5:43 PM