handsome-monitor-6885702/15/2023, 8:43 AM
loud-eve-7345702/16/2023, 5:32 AM
cool-monkey-7177402/16/2023, 10:09 AM
param in haproxy, and I used the
config for my rke2-nginx-ingress-controller. Unfortunately when I do this my services under nodeport are no longer working, and I also want to keep the ip for the nodeport in order to apply network policies based on client ip, any idea how to fix maybe the proxy or my nodeport so it works ?
shy-megabyte-7549202/16/2023, 5:44 PM
loud-eve-7345702/17/2023, 7:20 AM
boundless-waitress-8874802/17/2023, 1:57 PM
boundless-waitress-8874802/17/2023, 1:58 PM
cni: - multus - canal
boundless-waitress-8874802/17/2023, 2:03 PM
boundless-waitress-8874802/17/2023, 2:06 PM
steep-london-5309302/19/2023, 10:19 AM
Events: Type Reason Age From Message ---- ------ ---- ---- ------- Warning Unhealthy 22m (x1150 over 19h) kubelet Liveness probe failed: Get "<http://10.42.0.160:8080/health>": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Warning BackOff 7m27s (x3234 over 19h) kubelet Back-off restarting failed container Warning Unhealthy 2m27s (x3001 over 19h) kubelet Readiness probe failed: Get "<http://10.42.0.160:8181/ready>": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
write-kubeconfig-mode: "0644" advertise-address: - 192.168.0.12 cni: - cilium disable-kube-proxy: true disable: - rke2-canal - rke2-kube-proxy # Make a etcd snapshot every 6 hours etcd-snapshot-schedule-cron: " */6 * * *" # Keep 56 etcd snapshorts (equals to 2 weeks with 6 a day) etcd-snapshot-retention: 56 node-taint: - "CriticalAddonsOnly=true:NoExecute" tls-san: - 192.168.0.12
little-ram-1768302/20/2023, 4:50 AM
But looks like there is no config map for kube-proxy. When I try to edit
kubeproxy: extra_args: ipvs-strict-arp: 'true' proxy-mode: ipvs
to add custom config to kubeproxy I'm not sure where I should add this cfg. Could you help me, please? 🙂
cool-monkey-7177402/20/2023, 10:27 AM
to Local in the Nodeport definition, unfortunately when I do this I can't access my service even without NetworkPolicy. Does someone have any idea on how to fix that ? Thank you !
worried-ram-8108402/20/2023, 1:14 PM
wonderful-pizza-3091902/21/2023, 2:05 AM
acceptable-air-4960002/21/2023, 11:56 AM
wonderful-pizza-3091902/22/2023, 3:21 PM
thankful-train-278102/22/2023, 5:06 PM
loud-eve-7345702/23/2023, 2:38 AM
and try to put some things like below to make the kubernetes registry point to own harbor proxy cache, but it doesn’t work and the containerd still tries to pull image from the original one. What’s wrong with my configuration? thanks.
ambitious-butcher-8370302/23/2023, 4:53 PM
loud-eve-7345702/24/2023, 6:27 AM
witty-honey-1805202/24/2023, 8:48 PM
hundreds-airport-6619602/24/2023, 10:41 PM
little-ram-1768302/25/2023, 4:32 AM
kube-proxy-arg: - proxy-mode=ipvs - ipvs-strict-arp=true
. I mean in which section of yaml
strong-sunset-3002/25/2023, 2:50 PM
worried-ram-8108402/26/2023, 11:55 AM
and the master didn't join the cluster, what could be the issue?
rke2 server --server <https://172.16.1.11:6443>
shy-zebra-5307402/28/2023, 4:20 PM
user, etc…. I’m deploying in a 3-master HA configuration. What I am seeing is sometimes (not all) one of the master nodes (could be A, B or C) will hang. I am unable to SSH in, I’m unable to run any commands against the node. The node never recovers. I have to login to AWS console and manually stop and start the node, at which point it joins and I can add workers. I am trying to ensure that I am compliant with the supported runtime and requirements exactly as they should be as this will be a large production cluster and need to ensure it is exactly as needed. Anyone else had this issue? Anyone else seen this before? My process is: 1) I start Master A, wait 5m 2) Add Join Token to B/C 3) Start Master B, wait 5m 4) Start Master C, wait 5m… The entire process is automated with ansible. As I stated, this doesn’t happen every time. One thing I was doing is SSH into Master A and run
so I’m not sure if querying the API server while things are getting going would have this effect. I also provided all of the logs to AWS support for them to review OOM errors, but nothing looked out of the ordinary. RKE2 Version:
watch -n1 "kubectl get nodes"
Should I maybe add more time (more than 5m) between the start of each Master node?
RHEL 8.5 - minimal install
bright-fireman-4214403/01/2023, 12:22 AM
bright-fireman-4214403/01/2023, 12:23 AM
bright-fireman-4214403/01/2023, 12:24 AM
creamy-pencil-8291303/01/2023, 12:24 AM