crooked-cat-21365
01/10/2023, 2:43 PMhundreds-airport-66196
01/10/2023, 2:43 PMhundreds-evening-84071
01/10/2023, 3:06 PM413 Request Entity Too Large
Has anyone else experienced this? How to resolve this?hundreds-airport-66196
01/10/2023, 4:55 PMboundless-eye-27124
01/11/2023, 4:41 AMcrooked-cat-21365
01/11/2023, 11:53 AMroot@srvl034a:~# grep cgroup /proc/self/mounts
cgroup2 /sys/fs/cgroup cgroup2 rw,nosuid,nodev,noexec,relatime,nsdelegate,memory_recursiveprot 0 0
bitter-beard-2626
01/11/2023, 11:56 AMkubectl rollout restart daemonset -n kube-system cilium
caused a cluster outage. Output from journalctl -u rke2-server
Jan 10 11:31:21 k8s-build001-master001.dc1 rke2[980]: time="2023-01-10T11:31:21Z" level=error msg="Remotedialer proxy error" error="websocket: close 1006 (abnormal closure): unexpected EOF"
Jan 10 11:31:21 k8s-build001-master001.dc1 rke2[980]: time="2023-01-10T11:31:21Z" level=info msg="error in remotedialer server [400]: websocket: close 1006 (abnormal closure): unexpected EOF"
Jan 10 11:31:26 k8s-build001-master001.dc1 rke2[980]: time="2023-01-10T11:31:26Z" level=info msg="Connecting to proxy" url="<wss://10.81.23.12:9345/v1-rke2/connect>"
Jan 10 11:31:26 k8s-build001-master001.dc1 rke2[980]: time="2023-01-10T11:31:26Z" level=error msg="Failed to connect to proxy. Empty dialer response" error="dial tcp 10.81.23.12:9345: connect: connection refused"
Jan 10 11:31:26 k8s-build001-master001.dc1 rke2[980]: time="2023-01-10T11:31:26Z" level=error msg="Remotedialer proxy error" error="dial tcp 10.81.23.12:9345: connect: connection refused"
Jan 10 11:31:28 k8s-build001-master001.dc1 rke2[980]: {"level":"warn","ts":"2023-01-10T11:31:28.069Z","logger":"etcd-client","caller":"v3@v3.5.4-k3s1/retry_interceptor.go:62","msg":"retrying of unary invoker failed","target":"<etcd-endpoints://0xc000aace00/127.0.0.1:2379>","attempt":0,"error":"rpc error: code = DeadlineExceeded desc = context deadline exceeded"}
Jan 10 11:31:28 k8s-build001-master001.dc1 rke2[980]: time="2023-01-10T11:31:28Z" level=error msg="Failed to get recorded learner progress from etcd: context deadline exceeded"
Jan 10 11:31:31 k8s-build001-master001.dc1 rke2[980]: time="2023-01-10T11:31:31Z" level=info msg="Connecting to proxy" url="<wss://10.81.23.12:9345/v1-rke2/connect>"
Jan 10 11:32:10 k8s-build001-master001.dc1 rke2[980]: E0110 11:32:10.988495 980 leaderelection.go:330] error retrieving resource lock kube-system/rke2: Get "<https://127.0.0.1:6443/api/v1/namespaces/kube-system/configmaps/rke2>": stream error: stream ID 2425069; INTERNAL_ERROR; received from peer
k8s recovered after about 5 minutes.
In the hubble monitor I notice return traffic is denied from etcd port
Jan 10 11:51:29.529: 10.81.23.12:2380 (kube-apiserver) <> 10.81.23.11:32768 (host) policy-verdict:none DENIED (TCP Flags: SYN, ACK)
Jan 10 11:51:29.529: 10.81.23.12:2380 (kube-apiserver) <> 10.81.23.11:32768 (host) Policy denied DROPPED (TCP Flags: SYN, ACK)
Logs from cilium agent
Policy verdict log: flow 0xc18ce21f local EP ID 2998, remote ID kube-apiserver, proto 6, ingress, action deny, match none, 10.81.23.12:2380 -> 10.81.23.11:32770 tcp SYN, ACK
xx drop (Policy denied) flow 0xc18ce21f to endpoint 0, file bpf_host.c line 2147, , identity kube-apiserver->unknown: 10.81.23.12:2380 -> 10.81.23.11:32770 tcp SYN, ACK
Policy verdict log: flow 0xe145a85 local EP ID 2998, remote ID kube-apiserver, proto 6, ingress, action deny, match none, 10.81.23.12:2380 -> 10.81.23.11:32768 tcp SYN, ACK
xx drop (Policy denied) flow 0xe145a85 to endpoint 0, file bpf_host.c line 2147, , identity kube-apiserver->unknown: 10.81.23.12:2380 -> 10.81.23.11:32768 tcp SYN, ACK
Anyone that has encountered similar issues?narrow-noon-75604
01/12/2023, 7:58 AMsquare-policeman-85866
01/12/2023, 10:12 AMnarrow-noon-75604
01/12/2023, 11:50 AMloud-receptionist-98355
01/12/2023, 2:17 PMloud-receptionist-98355
01/12/2023, 2:19 PMloud-receptionist-98355
01/12/2023, 2:20 PMdefault
namespace also worksrough-ocean-41843
01/12/2023, 4:53 PMloud-jackal-15304
01/12/2023, 6:24 PMearly-engineer-43393
01/16/2023, 5:23 PMvsphere CSI controller
, we kept getting an i/o timeout when connecting to vsphere and the only way around it was to add our name server to the k8s netpol, obviously we dont want to do that, we deployed another test pod using the exact same netpol however we excluded the name server and it worked fine.
We noticed that dnsPolicy on the vsphere CSI controller is Default
, we updated that too ClusterFirst
and could see the connection working as expected, however now the pods keep going into. CrashLoopBackoff
with the following error:
Still connecting to unix:///csi/csi.sock
Any ideas?adorable-hairdresser-9976
01/18/2023, 11:21 PMrefined-eye-25557
01/20/2023, 6:41 AMconfig.yaml
file before first starting RKE2 server. How can I generate that K10 token without Kubernetes already up an running?narrow-noon-75604
01/20/2023, 7:38 PMfuture-monitor-61871
01/20/2023, 9:23 PMagreeable-sugar-99934
01/21/2023, 8:53 AMtls-san
Cluster setup works fine, all 3 nodes are in the k8s cluster. kubectl get node
does work when using the loaadbalancer IP in my kubeconfig.
But when I want to execute a kubectl logs <pod
I get an error
Error from server: Get "<https://10.0.0.4:10250/containerLogs/kube-system/cilium-9cg2l/cilium-agent?follow=true>": x509: certificate is valid for 127.0.0.1, <publicipv4>, <publicipv6>, not 10.0.0.4
the node has a public and internal interface and it seems the internal ip address is nod added to the kubelet certificate. How can I change this?bright-farmer-78407
01/23/2023, 10:07 PMcreamy-autumn-87994
01/24/2023, 10:12 PMorange-airplane-98016
01/24/2023, 10:35 PMred-magician-75203
01/25/2023, 8:18 PMblue-controller-9088
01/26/2023, 11:45 PMsparse-fireman-14239
01/27/2023, 7:33 AMcool-monkey-71774
01/27/2023, 7:33 AMFATA[0006] starting kubernetes: preparing server: failed to get CA certs: Get "<https://xxxx>:xxxx:xxxx:xxxx:xxxx:xxxx:9345/cacerts": dial tcp [xxxx:xxxx:xxxx:xxxx:xxxx:xxxx]:9345: connect: no route to host
I manage to get it to work by editing the /etc/rancher/rke2/config.yaml.d/50-rancher.yaml file and changing the IPv6 to IPv4.
But I got another issue for the worker nodes even after updating the url, the node gets a 401 unauthorized response, when I try to add other master nodes it works fine
FATA[0000] starting kubernetes: preparing server: <https://xxx.xxx.xxx.xxx:9345/v1-rke2/server-bootstrap>: 401 Unauthorized
Thanks for your timebest-microphone-20624
01/27/2023, 12:15 PMgentle-petabyte-40055
01/29/2023, 9:29 PM