handsome-monitor-68857
02/15/2023, 8:43 AMloud-eve-73457
02/16/2023, 5:32 AMcool-monkey-71774
02/16/2023, 10:09 AMsend-proxy
param in haproxy, and I used the use-proxy-protocol
config for my rke2-nginx-ingress-controller. Unfortunately when I do this my services under nodeport are no longer working, and I also want to keep the ip for the nodeport in order to apply network policies based on client ip, any idea how to fix maybe the proxy or my nodeport so it works ?shy-megabyte-75492
02/16/2023, 5:44 PMloud-eve-73457
02/17/2023, 7:20 AMboundless-waitress-88748
02/17/2023, 1:57 PMboundless-waitress-88748
02/17/2023, 1:58 PMcni:
- multus
- canal
boundless-waitress-88748
02/17/2023, 2:03 PMboundless-waitress-88748
02/17/2023, 2:06 PMsteep-london-53093
02/19/2023, 10:19 AMEvents:
Type Reason Age From Message
---- ------ ---- ---- -------
Warning Unhealthy 22m (x1150 over 19h) kubelet Liveness probe failed: Get "<http://10.42.0.160:8080/health>": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
Warning BackOff 7m27s (x3234 over 19h) kubelet Back-off restarting failed container
Warning Unhealthy 2m27s (x3001 over 19h) kubelet Readiness probe failed: Get "<http://10.42.0.160:8181/ready>": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
Configuration:
write-kubeconfig-mode: "0644"
advertise-address:
- 192.168.0.12
cni:
- cilium
disable-kube-proxy: true
disable:
- rke2-canal
- rke2-kube-proxy
# Make a etcd snapshot every 6 hours
etcd-snapshot-schedule-cron: " */6 * * *"
# Keep 56 etcd snapshorts (equals to 2 weeks with 6 a day)
etcd-snapshot-retention: 56
node-taint:
- "CriticalAddonsOnly=true:NoExecute"
tls-san:
- 192.168.0.12
little-ram-17683
02/20/2023, 4:50 AMkube-proxy
:
kubeproxy:
extra_args:
ipvs-strict-arp: 'true'
proxy-mode: ipvs
But looks like there is no config map for kube-proxy. When I try to edit cluster.yaml
to add custom config to kubeproxy I'm not sure where I should add this cfg. Could you help me, please? 🙂cool-monkey-71774
02/20/2023, 10:27 AMexternalTrafficPolicy
to Local in the Nodeport definition, unfortunately when I do this I can't access my service even without NetworkPolicy. Does someone have any idea on how to fix that ? Thank you !worried-ram-81084
02/20/2023, 1:14 PMwonderful-pizza-30919
02/21/2023, 2:05 AMacceptable-air-49600
02/21/2023, 11:56 AMwonderful-pizza-30919
02/22/2023, 3:21 PMthankful-train-2781
02/22/2023, 5:06 PMloud-eve-73457
02/23/2023, 2:38 AMregistries.yaml
and try to put some things like below to make the kubernetes registry point to own harbor proxy cache, but it doesn’t work and the containerd still tries to pull image from the original one. What’s wrong with my configuration? thanks.ambitious-butcher-83703
02/23/2023, 4:53 PMloud-eve-73457
02/24/2023, 6:27 AMwitty-honey-18052
02/24/2023, 8:48 PMhundreds-airport-66196
02/24/2023, 10:41 PMlittle-ram-17683
02/25/2023, 4:32 AMkube-proxy-arg:
- proxy-mode=ipvs
- ipvs-strict-arp=true
in cluster.yaml
. I mean in which section of yamlstrong-sunset-30
02/25/2023, 2:50 PMworried-ram-81084
02/26/2023, 11:55 AMrke2 server --server <https://172.16.1.11:6443>
and the master didn't join the cluster, what could be the issue?shy-zebra-53074
02/28/2023, 4:20 PMetcd
user, etc….
I’m deploying in a 3-master HA configuration. What I am seeing is sometimes (not all) one of the master nodes (could be A, B or C) will hang. I am unable to SSH in, I’m unable to run any commands against the node. The node never recovers. I have to login to AWS console and manually stop and start the node, at which point it joins and I can add workers. I am trying to ensure that I am compliant with the supported runtime and requirements exactly as they should be as this will be a large production cluster and need to ensure it is exactly as needed.
Anyone else had this issue? Anyone else seen this before? My process is: 1) I start Master A, wait 5m 2) Add Join Token to B/C 3) Start Master B, wait 5m 4) Start Master C, wait 5m… The entire process is automated with ansible. As I stated, this doesn’t happen every time.
One thing I was doing is SSH into Master A and run watch -n1 "kubectl get nodes"
so I’m not sure if querying the API server while things are getting going would have this effect. I also provided all of the logs to AWS support for them to review OOM errors, but nothing looked out of the ordinary.
RKE2 Version: 1.26.0+rke2r2
OS: RHEL 8.5 - minimal install
Should I maybe add more time (more than 5m) between the start of each Master node?bright-fireman-42144
03/01/2023, 12:22 AMbright-fireman-42144
03/01/2023, 12:23 AMbright-fireman-42144
03/01/2023, 12:24 AMcreamy-pencil-82913
03/01/2023, 12:24 AM