This message was deleted.
# general
a
This message was deleted.
p
Any log from the docker container or on eks side?
g
Copy code
024/05/12 10:09:04 [ERROR] Failed to install system chart rancher-provisioning-capi: pod cattle-system/helm-operation-xqsmw failed, watch closed
2024/05/12 10:09:37 [ERROR] error syncing 'c-9wr8x': handler eks-operator-controller: waiting eks crd to be initialized, cluster: c-9wr8x, requeuing
2024/05/12 10:10:06 [ERROR] Failed to install system chart fleet-crd: pod cattle-system/helm-operation-blzxs failed, watch closed
2024/05/12 10:11:08 [ERROR] Failed to install system chart rancher-eks-operator: pod cattle-system/helm-operation-2v5hr failed, watch closed
2024/05/12 10:11:37 [ERROR] error syncing 'c-9wr8x': handler eks-operator-controller: waiting eks crd to be initialized, cluster: c-9wr8x, requeuing
2024/05/12 10:12:11 [ERROR] Failed to install system chart rancher-eks-operator-crd: pod cattle-system/helm-operation-wc8b9 failed, watch closed
W0512 10:12:53.067919      38 warnings.go:80] <http://cluster.x-k8s.io/v1alpha3|cluster.x-k8s.io/v1alpha3> MachineDeployment is deprecated; use <http://cluster.x-k8s.io/v1beta1|cluster.x-k8s.io/v1beta1> MachineDeployment
2024/05/12 10:13:13 [ERROR] Failed to install system chart rancher-eks-operator: pod cattle-system/helm-operation-9lpzd failed, watch closed
2024/05/12 10:13:37 [ERROR] error syncing 'c-9wr8x': handler eks-operator-controller: waiting eks crd to be initialized, cluster: c-9wr8x, requeuing
2024/05/12 10:14:15 [ERROR] Failed to install system chart rancher-eks-operator-crd: pod cattle-system/helm-operation-plqsc failed, watch closed
2024/05/12 10:15:17 [ERROR] Failed to install system chart rancher-eks-operator: pod cattle-system/helm-operation-xwtx4 failed, watch closed
W0512 10:15:25.088161      38 warnings.go:80] <http://cluster.x-k8s.io/v1alpha3|cluster.x-k8s.io/v1alpha3> Cluster is deprecated; use <http://cluster.x-k8s.io/v1beta1|cluster.x-k8s.io/v1beta1> Cluster
W0512 10:15:33.020579      38 warnings.go:80] <http://cluster.x-k8s.io/v1alpha3|cluster.x-k8s.io/v1alpha3> MachineHealthCheck is deprecated; use <http://cluster.x-k8s.io/v1beta1|cluster.x-k8s.io/v1beta1> MachineHealthCheck
2024/05/12 10:15:37 [ERROR] error syncing 'c-9wr8x': handler eks-operator-controller: waiting eks crd to be initialized, cluster: c-9wr8x, requeuing
W0512 10:15:50.088007      38 warnings.go:80] <http://cluster.x-k8s.io/v1alpha3|cluster.x-k8s.io/v1alpha3> Machine is deprecated; use <http://cluster.x-k8s.io/v1beta1|cluster.x-k8s.io/v1beta1> Machine
2024/05/12 10:16:19 [ERROR] Failed to install system chart rancher-eks-operator-crd: pod cattle-system/helm-operation-n6p69 failed, watch closed
W0512 10:16:54.024091      38 warnings.go:80] <http://cluster.x-k8s.io/v1alpha3|cluster.x-k8s.io/v1alpha3> MachineSet is deprecated; use <http://cluster.x-k8s.io/v1beta1|cluster.x-k8s.io/v1beta1> MachineSet
@powerful-librarian-10572
p
I think you can find the failed pod in the local cluster, or at last i hope you can
You need to get the job logs at least, but if it's in the failed cluster, you're in for some pain
g
i am new in rancher, so didn't get you.
p
Go to your "local" cluster (rancher cluster) and check in the system workloads if there is not a failed pod that may contains some insights
g
i don't have local cluster, i have installed it by docker over ec2.
p
That is the local cluster.
g
found this.
p
yep, now go in workloads and show system namespaces
g
local
p
Yep, theres your issue, your local node (i think) is on disk pressure so kubes cant schedule on it
you're out of disk space, basically
g
what is the minimum requirement to run rancher.
in terms of cpu, mem and disk
p
You should have less than 80% disk usage on your rancher node
My current rancher install requires 4gigs of storage
g
got it, let me increase it and then try.
2024/05/12 10:58:59 [ERROR] error syncing 'c-hnmh2': handler eks-operator-controller: error for cluster [c-hnmh2]: health error for node group [beta-ng] in cluster [c-hnmh2]: Could not launch Spot Instances. UnfulfillableCapacity - Unable to fulfill capacity due to your request configuration. Please adjust your request and try again. Launching EC2 instance failed., requeuing
@powerful-librarian-10572
p
No effin clue about how eks work doe
Re-create the cluster from scratch? delete it from rancher and try to re-inmport it
g
ok
Clsuetr is active now Thanks a lot @powerful-librarian-10572 for this. One more doubt i have, i am getting this continuously in logs.
Copy code
netpolMgr: handleHostNetwork: couldn't parse PodCIDR() for node ip-11-5-6-145.ap-south-1.compute.internal err=invalid CIDR address:
One more doubt, How i can make this persistance, for example if my container goes down and new will come up then all my data and setup is gone, how i can prevent this. @powerful-librarian-10572
g
Thanks
164 Views