https://rancher.com/ logo
Title
b

bumpy-portugal-40754

06/28/2022, 11:57 PM
Recently after a powerdown of the complete harvester cluster I got 2 volumes with a stopped replica because... I don't know. Longhorn GUI complained about the fact that 2 manager instances are running for both volumes. They did, surprisingly not on the node with the stopped replica. I killed one manager pod and the volume got rebuild. Any opinion about this "strategy" to solve this issue? How to debug? Are there remains of the old replica left?
1
a

aloof-hair-13897

06/29/2022, 3:30 AM
Hi, @bumpy-portugal-40754 How did you do power down the harvester cluster and that including the nodes with volumes’ replica? It is normal that you would get stopped replicas if the node containing the replica is down.
Longhorn GUI complained about the fact that 2 manager instances are running for both volumes.
Did you mean instance-manager-e-xxx and instance-manager-r-xxx? Could you provide the support bundle then we could find out what happened.
b

bumpy-portugal-40754

06/29/2022, 11:11 AM
I didn't do the powerdown myself, but I'd suspect it to be graceful. Instance-manager-e... was there twice. The one which is shown in the longhorn gui. I already created a case with suse. Are you able to access it?
c

cuddly-vase-67379

06/29/2022, 12:30 PM
Could you show me where it is?
b

bumpy-portugal-40754

06/29/2022, 1:18 PM
Looks like I was wrong about the way the machines were powered down. It seems that the servers were simply switched off.