Adamb
It sounds correct. However when my physical network card on the node dies it affects VMs using it as well. I may be using completely different network card for cluster network and in that case VMs bridging to it will be affected too.
Wouldn't be migrating / relocating VMs to be a...
Is support for RRP not implemented at all?
It seems so. I cannot migrate vm:101 via second ring.
Issuing migrate to alternative node name fails with (node virt2n1-la-int is not online)
The only method forked for me is to login into the test node add static route to a primary ring network...
I am testing on a real cluster so I decided to open a new thread to avoid the confusion.
I have RRP (two different networks) configured on corosync. After testing HA in case of network failure I wonder now if it makes sense at all.
When I stop one of the interfaces on node, corosync declares...
Dietmar,
Thank you for your reply.
So there is 120 seconds for locking and 60 seconds for fencing, correct?
Are there any plans to make those numbers configurable?
Finally I tested 2 out of three nodes down situation
info 07:34:47 hardware: crm on node 'node2' killed by poweroff...
I also tried a node power off simulation.
I killed node1 with vm:101 running on it
It took approximately 3 minutes to fence the node 1
Then vm:101 started on node2 almost immediately.
Again something doesn't compute here, why wait 3 minutes to fence the dead node?
info 07:01:24...
For better understanding of new HA mechanism, I have decided to try you pve-ha-simulator.
I started all nodes and enabled one vm:101.
Then I migrated vm:101 to node 2 so far so good.
Finally I disabled network on node2.
Simulator fenced the node2 and started vm:101 on node1, however it took 3...
I really had to check first :(. "FIXED BY UPDATING TO THE LATEST PACKAGES" was relevant at least for MasterTH. The new ha related debs were released last night. I just installed them and will report if I can reproduce the HA stuck issue. Interesting enough, these packages don't trigger any...
Spirit,
Forgot to mention. I stopped / started pvedaemon first and then restarted pve-ha-lrm on all nodes.
So no-resume issue doesn't happen anymore. But stuck after HA migrate starts with ha-manager showing vm on source node is still happening. I will run another batch of migrations to see...
MasterTH,
It seems to be related to problem in a different thread. Unless something was released overnight, your last statement is irrelevant. Please don't hijack the thread.
Sent from my SM-G900V using Tapatalk
I restarted pve-ha-lrm on all nodes. This problem where ha starts but nothing happens seems to be harder to reproduce but it is not fixed with new patched deb
Sent from my SM-G900V using Tapatalk
I now recall it happened one time with me before today's update. Fixing it this way may be acceptable during the testing like now, but it definitely a problem in production. Do you know what is happening there?
I installed it and ran but now I have the problem which looks even worse. HA Migrate task executed but, nothing else happens.
Daemon log on the source node continuously printing
Oct 15 07:08:59 virt2n1-la pve-ha-lrm[119275]: service 'vm:101' not on this node at...
On source node I see:
Oct 14 12:11:58 virt2n2-la pve-ha-lrm[158659]: Task still active, waitingOct 14 12:11:59 virt2n2-la pve-ha-lrm[158659]: Task still active, waiting
Oct 14 12:12:00 virt2n2-la pve-ha-lrm[158659]: Task still active, waiting
Oct 14 12:12:01 virt2n2-la pve-ha-lrm[158659]: Task...
Changed printout as requested
task started by HA resource agent
Oct 13 21:39:41 starting migration of VM 101 to node 'virt2n2-la' (38.102.250.229)
Oct 13 21:39:41 copying disk images
Oct 13 21:39:41 starting VM 101 on remote node 'virt2n2-la'
Oct 13 21:39:42 starting ssh migration tunnel
Oct...
Responding to a time check first:
Source Node
root@virt2n2-la:~# date && mv /etc/pve/nodes/virt2n2-la/qemu-server/101.conf /etc/pve/nodes/virt2n1-la/qemu-server/
Tue Oct 13 21:29:23 PDT 2015
Target node
root@virt2n1-la:~# ls -l --time-style=full-iso...
Look at the file time stamp on the remote node.
ls -l --time-style=full-iso /etc/pve/nodes/virt2n2-la/qemu-server/101.conf
-rw-r----- 1 root www-data 392 2015-10-13 20:50:31.000000000 -0700 /etc/pve/nodes/virt2n2-la/qemu-server/101.conf
Now look at the time stamp on the task log errors...
Here is the task log
Oct 13 20:50:19 starting migration of VM 101 to node 'virt2n2-la' (38.102.250.229)
Oct 13 20:50:19 copying disk images
Oct 13 20:50:19 starting VM 101 on remote node 'virt2n2-la'
Oct 13 20:50:21 starting ssh migration tunnel
Oct 13 20:50:21 starting online/live migration...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.