Problems with migration of VM using rrp_mode

e.helix

Member
Jan 15, 2019
9
0
6
35
Hi,

I use corosync rrp_mode in my cluster, 6 nodes in total. RRP_mode is passive (used this https://pve.proxmox.com/wiki/Separate_Cluster_Network manual while setting it up). When the ring0 interface is disabled on any cluster node it becomes impossible to migrate the VM using ring1 to another node even if the migration network is specified in datacenter.cfg file or "qm migrate" command is used.
I can't figure out why migration doesn't work with the ring1 interface running and the ring0 interface turned off.

Any ideas?

My cluster configuration is given below.

Corosync.conf
logging {
debug: off
to_syslog: yes
}

nodelist {
node {
name: adm02-tstpve02
nodeid: 2
quorum_votes: 1
ring0_addr: 192.168.210.2
ring1_addr: 192.168.220.2
}
node {
name: adm03-tstpve03
nodeid: 3
quorum_votes: 1
ring0_addr: 192.168.210.3
ring1_addr: 192.168.220.3
}
node {
name: adm04-tstpve04
nodeid: 4
quorum_votes: 1
ring0_addr: 192.168.210.4
ring1_addr: 192.168.220.4
}
node {
name: adm05-tstpve05
nodeid: 5
quorum_votes: 1
ring0_addr: 192.168.210.5
ring1_addr: 192.168.220.5
}
node {
name: adm06-tstpve06
nodeid: 6
quorum_votes: 1
ring0_addr: 192.168.210.6
ring1_addr: 192.168.220.6
}
node {
name: adm07-tstpve07
nodeid: 7
quorum_votes: 1
ring0_addr: 192.168.210.7
ring1_addr: 192.168.220.7
}
}

quorum {
provider: corosync_votequorum
}

totem {
cluster_name: Proxtst
config_version: 13
interface {
bindnetaddr: 192.168.210.0
ringnumber: 0
}
interface {
bindnetaddr: 192.168.220.0
ringnumber: 1
}
ip_version: ipv4
rrp_mode: passive
secauth: on
version: 2
}

Cluster version
proxmox-ve: 5.1-25 (running kernel: 4.13.4-1-pve)
pve-manager: 5.2-10 (running version: 5.2-10/6f892b40)
pve-kernel-4.13.4-1-pve: 4.13.4-26
ceph: 12.0.0-pve2
corosync: 2.4.2-pve5
criu: 2.11.1-1~bpo90
glusterfs-client: 3.8.8-1
ksm-control-daemon: not correctly installed libjs-extjs: 6.0.1-2
libpve-access-control: 5.0-8
libpve-apiclient-perl: 2.0-5
libpve-common-perl: 5.0-41
libpve-guest-common-perl: 2.0-18
libpve-http-server-perl: 2.0-11
libpve-storage-perl: 5.0-30
libqb0: 1.0.1-1
lvm2: 2.02.168-pve6
lxc-pve: 3.0.2+pve1-3
lxcfs: 3.0.2-2
novnc-pve: 1.0.0-2
openvswitch-switch: 2.7.0-3
proxmox-widget-toolkit: 1.0-20
pve-cluster: 5.0-30
pve-container: 2.0-29
pve-docs: 5.2-9
pve-firewall: 3.0-14
pve-firmware: 2.0-6
pve-ha-manager: 2.0-5
pve-i18n: 1.0-6
pve-libspice-server1: 0.14.1-1
pve-qemu-kvm: 2.12.1-1
pve-xtermjs: 1.0-5
qemu-server: 5.0-38
smartmontools: 6.5+svn4324-1
spiceterm: 3.0-5
vncterm: 1.5-3

Hosts configuration
127.0.0.1 localhost
10.150.30.22 adm03-tstpve03.spb.helix.ru pvelocalhost
192.168.210.3 adm03-tstpve03.spb.helix.ru adm03-tstpve03

# corosync ring0 network hosts
192.168.210.2 adm02-tstpve02.spb.helix.ru adm02-tstpve02
192.168.210.4 adm04-tstpve04.spb.helix.ru adm04-tstpve04
192.168.210.5 adm05-tstpve05.spb.helix.ru adm05-tstpve05
192.168.210.6 adm06-tstpve06.spb.helix.ru adm06-tstpve06
192.168.210.7 adm07-tstpve07.spb.helix.ru adm07-tstpve07

# corosync ring1 network hosts
192.168.220.2 adm02-tstpve02.spb.helix.ru adm02-tstpve02
192.168.220.3 adm03-tstpve03.spb.helix.ru adm03-tstpve03
192.168.220.4 adm04-tstpve04.spb.helix.ru adm04-tstpve04
192.168.220.5 adm05-tstpve05.spb.helix.ru adm05-tstpve05
192.168.220.6 adm06-tstpve06.spb.helix.ru adm06-tstpve06
192.168.220.7 adm07-tstpve07.spb.helix.ru adm07-tstpve07


# The following lines are desirable for IPv6 capable hosts
::1 localhost ip6-localhost ip6-loopback
ff02::1 ip6-allnodes
ff02::2 ip6-allrouters
 
I use corosync rrp_mode in my cluster, 6 nodes in total. RRP_mode is passive (used this https://pve.proxmox.com/wiki/Separate_Cluster_Network manual while setting it up). When the ring0 interface is disabled on any cluster node it becomes impossible to migrate the VM using ring1 to another node even if the migration network is specified in datacenter.cfg file or "qm migrate" command is used.
I can't figure out why migration doesn't work with the ring1 interface running and the ring0 interface turned off.


The network to be used is not necessarily the corosync network but the ntwoek specified in /etc/pve/datacenter.cfg (see https://pve.proxmox.com/wiki/Manual:_datacenter.cfg ); detault is the network defined in /etc/hosts which is also default for ring0. So - in case of default configuration - when ring0 is down migration will not work either.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!