Hi everybody,
I want to migrate a container from one server (node-1) to another server (node-2) (the replication for this CT is active only to a 3rd server (node-3)). I first created a group for the HA where the 3rd server (node-3) has priority. I then set the HA for the CT to this group.
When I shut down the node-1 server, the CT was stopped and then moved to node-2. I'm now trying to move the CT back to node-1 and there... problem !
Here is the Proxmox version of my nodes:
...In order to solve the problem I had to do this on node-1:
...then from the GUI, I was able to make the migration which went well.
Has anyone ever encountered this problem ?
Thanks
I want to migrate a container from one server (node-1) to another server (node-2) (the replication for this CT is active only to a 3rd server (node-3)). I first created a group for the HA where the 3rd server (node-3) has priority. I then set the HA for the CT to this group.
When I shut down the node-1 server, the CT was stopped and then moved to node-2. I'm now trying to move the CT back to node-1 and there... problem !
Code:
task started by HA resource agent
2018-01-30 12:26:35 starting migration of CT 101 to node 'monserveur' (10.12.1.5)
2018-01-30 12:26:35 found local volume 'zfs-cadstorage-CT:subvol-101-disk-1' (in current VM config)
full send of cadzfs/CT/subvol-101-disk-1@__replicate_101-0_1517310001__ estimated size is 3.57G
send from @__replicate_101-0_1517310001__ to cadzfs/CT/subvol-101-disk-1@__migration__ estimated size is 24.8M
total estimated size is 3.60G
TIME SENT SNAPSHOT
cadzfs/CT/subvol-101-disk-1 name cadzfs/CT/subvol-101-disk-1 -
volume 'cadzfs/CT/subvol-101-disk-1' already exists
command 'zfs send -Rpv -- cadzfs/CT/subvol-101-disk-1@__migration__' failed: got signal 13
send/receive failed, cleaning up snapshot(s)..
2018-01-30 12:26:36 ERROR: command 'set -o pipefail && pvesm export zfs-cadstorage-CT:subvol-101-disk-1 zfs - -with-snapshots 0 -snapshot __migration__ | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=monserveur' root@10.12.1.5 -- pvesm import zfs-cadstorage-CT:subvol-101-disk-1 zfs - -with-snapshots 0 -delete-snapshot __migration__' failed: exit code 255
2018-01-30 12:26:36 aborting phase 1 - cleanup resources
2018-01-30 12:26:36 ERROR: found stale volume copy 'zfs-cadstorage-CT:subvol-101-disk-1' on node 'monserveur'
2018-01-30 12:26:36 start final cleanup
2018-01-30 12:26:36 ERROR: migration aborted (duration 00:00:01): command 'set -o pipefail && pvesm export zfs-cadstorage-CT:subvol-101-disk-1 zfs - -with-snapshots 0 -snapshot __migration__ | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=monserveur' root@10.12.1.5 -- pvesm import zfs-cadstorage-CT:subvol-101-disk-1 zfs - -with-snapshots 0 -delete-snapshot __migration__' failed: exit code 255
TASK ERROR: migration aborted
Here is the Proxmox version of my nodes:
Code:
root@monserveur:~# pveversion -v
proxmox-ve: 5.1-38 (running kernel: 4.13.13-5-pve)
pve-manager: 5.1-43 (running version: 5.1-43/bdb08029)
pve-kernel-4.13.4-1-pve: 4.13.4-26
pve-kernel-4.13.13-5-pve: 4.13.13-38
libpve-http-server-perl: 2.0-8
lvm2: 2.02.168-pve6
corosync: 2.4.2-pve3
libqb0: 1.0.1-1
pve-cluster: 5.0-19
qemu-server: 5.0-20
pve-firmware: 2.0-3
libpve-common-perl: 5.0-25
libpve-guest-common-perl: 2.0-14
libpve-access-control: 5.0-7
libpve-storage-perl: 5.0-17
pve-libspice-server1: 0.12.8-3
vncterm: 1.5-3
pve-docs: 5.1-16
pve-qemu-kvm: 2.9.1-6
pve-container: 2.0-18
pve-firewall: 3.0-5
pve-ha-manager: 2.0-4
ksm-control-daemon: 1.2-2
glusterfs-client: 3.8.8-1
lxc-pve: 2.1.1-2
lxcfs: 2.0.8-1
criu: 2.11.1-1~bpo90
novnc-pve: 0.6-4
smartmontools: 6.5+svn4324-1
zfsutils-linux: 0.7.4-pve2~bpo9
ceph: 12.2.2-pve1
root@monserveur:~#
...In order to solve the problem I had to do this on node-1:
Code:
root@monserveur:~# zfs destroy cadzfs/CT/subvol-101-disk-1
cannot destroy 'cadzfs/CT/subvol-101-disk-1': filesystem has children
use '-r' to destroy the following datasets:
cadzfs/CT/subvol-101-disk-1@__replicate_101-0_1517310001__
root@monserveur:~# zfs destroy -r cadzfs/CT/subvol-101-disk-1
...then from the GUI, I was able to make the migration which went well.
Has anyone ever encountered this problem ?
Thanks