This is full information. Do you think this problem caused by bindnetaddr is listening on IP of hv117? and when this node reboot, all cluster dump.
And this is full log before whole cluster rebooted:
Hi,
I have proxmox server with 11 nodes.
And this is our corosync config:
when we reboot server hv117 which has IP 10.10.30.167: cluster is still up. But when this server booted into OS, all cluster rebooted.
May we know how to debug this problem?
Thanks,
Hi,
We have this error, this usually come after running backup on our PVE.
root@hv101:~# service pveproxy status
● pveproxy.service - PVE API Proxy Server
Loaded: loaded (/lib/systemd/system/pveproxy.service; enabled; vendor preset: enabled)
Active: active (running) since Wed 2018-08-22...
Yes,
1 for public network
1 for private network LAN - cross VM connection
1 for CEPH - 10Gbps would be better
1 for cluster network - this must be dedicated for ring0 network and you should add ring1 network. You may want to take a look here...
So this will require downtime to making change.
You can create nfs share of /var/lib/vz on Proxmox A then Proxmox B will connect to this folder. Then stop VM on Proxmox A and start on Proxmox B -->move disk -->done.
Note: performances will be degrade but your VM can online with minimal downtime.
The easy way is attach both storage to Proxmox Host and use “move disk” function to move your qcow2 vdisk to new storage. This is same to live migration.
Thanks Udo, we solved a lot of case from your reply on this community.
Correct. But we have 8 x 10Gbps per node:
bond0: vmbr0 for VM public network connection. This bond run as mode active-backup with dual 10Gbps SFP+
bond1: this bond works as cluster network, migration network, 2 differences...
Hi,
I have 10 nodes in cluster, is this reliable quorum?
We are usually facing reboot on the weekend (during backup VM). Currently we have stopped backup all nodes and create backup cron per node to avoid this random reboot. It has been up more than 2 weeks.
May we know how to debug this?
This is strange. We did not use any LXC on this cluster. We have 2 incidents last weekend. Proxmox random reboot on Saturday and Sunday after upgrade to Proxmox latest version. Is this the root cause of issue?
proxmox-ve: 5.2-2 (running kernel: 4.15.17-3-pve)
pve-manager: 5.2-3 (running...
Hello,
We receive 1 cluster reboot random (HP Blade server) and our WHMCS connect to this cluster is really slow. We check pveproxy status and it showing this output:
pveproxy.service - PVE API Proxy Server
Loaded: loaded (/lib/systemd/system/pveproxy.service; enabled; vendor preset...
There is something stranges, i have delete old pool and created a new pool with size=2 and min_size=1. health is ok and 12 OSDs in.
This morning 1 OSD out again:
Degraded data redundancy: 2480/235442 objects degraded (1.053%), 7 pgs degraded, 7 pgs undersized
pg 3.46 is stuck undersized for...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.