```I get an error while deploying RKE with 1 maste...
# rke
s
Copy code
I get an error while deploying RKE with 1 master(role:controlplane,etcd) and 2 workers(role: worker). Following is my cluster.yml
nodes: - address: <<VM-1>> port: "22" internal_address: "" role: - controlplane - etcd hostname_override: "" user: vudemaster docker_socket: /var/run/docker.sock ssh_key: "" ssh_key_path: ~/.ssh/id_rsa ssh_cert: "" ssh_cert_path: "" labels: {} taints: [] - address: <<VM-2>> port: "22" internal_address: "" role: - worker hostname_override: "" user: vudeadmin docker_socket: /var/run/docker.sock ssh_key: "" ssh_key_path: ~/.ssh/id_rsa ssh_cert: "" ssh_cert_path: "" labels: {} taints: [] - address: <<VM-3>> port: "22" internal_address: "" role: - worker hostname_override: "" user: vudeadmin docker_socket: /var/run/docker.sock ssh_key: "" ssh_key_path: ~/.ssh/id_rsa ssh_cert: "" ssh_cert_path: "" labels: {} taints: [] Error: FATA[0115] [workerPlane] Failed to bring up Worker Plane: [Failed to verify healthcheck: Failed to check http://localhost:10256/healthz for service [kube-proxy] on host [cdedops-vude4.elements.local]: Get "http://localhost:10256/healthz": Unable to access the service on localhost:10256. The service might be still starting up. Error: ssh: rejected: connect failed (Connection refused), log: E0919 093712.439028 2897 node.go:161] Failed to retrieve node info: Get "https://127.0.0.1:6443/api/v1/nodes/cdedops-vude4.elements.local": read tcp 127.0.0.148478 &gt;127.0.0.16443: read: connection reset by peer] . But If I would specify roles as [controlplane,etcd] in all nodes, then it works fine. The error comes when role is given as worker alone. Refer configs above. Please suggest.