Hi Folks!
I'm trying out Proxmox VE ( 6.0-9 ) plus Linstor Plugin as a Datacenter Virtualization Solution for the company I work for. So far, pretty good. I've found a glitch though, and as I dont really know if it's about proxmox or about Linstor + ZFS, I'll post this in both communities. My scenario:
- 2 clustered nodes for testing
- Both feature a ZFS pool ( tank - to be original
- Installed and configured Linstor + DRBD9 + Controller in both ( node1 as Controller )
- Created DRBD storage pool based on ZFS ( tried both thick & thin )
So, to sum up, I'm using a ZFS pool on both nodes as VM image storage and as a storage pool from which to create my DRBD replicated volumes. I then created a couple VMs, one of them backed by a ZFS volume, and the other one backed by a DRBD replicated resource.
Here's the glitch: The ZFS backed VM I can migrate, no problem, but the DRBD backed VM, it wont migrate. When trying to, it looks like proxmox incorrectly detects it's backing storage as a ZFS volume, instead of the DRBD resource. VM's local disk is: "drbdstorage:vm-101-disk-1". I'm wondering if using the same ZFS pool for both VMs and Linstor is not supported for this setup . . .
This is my storage.cfg:
drbd: drbdstorage
content rootdir,images
controller 10.17.0.71
redundancy 2
zfspool: zp1
pool tank
content images,rootdir
sparse 1
Here's my problematic VM's config:
agent: 1
bootdisk: scsi0
cores: 2
ide2: none,media=cdrom
memory: 2048
name: vm-xenial-server
net0: virtio=06:04:B3:E1:75:AF,bridge=vmbr0,tag=1
numa: 0
ostype: l26
scsi0: drbdstorage:vm-101-disk-1,cache=unsafe,size=8G
scsihw: virtio-scsi-pci
smbios1: uuid=9a8e8301-a865-4544-a367-5111cfcbff16
sockets: 2
vmgenid: 2b7a93b0-f808-44a5-818a-cdfb4a70586e
Here's the migration log extract:
2019-10-20 08:52:25 use dedicated network address for sending migration traffic (10.17.0.72)
2019-10-20 08:52:25 starting migration of VM 101 to node 'ich2' (10.17.0.72)
-------> 2019-10-20 08:52:25 found local disk 'zp1:vm-101-disk-1_00000' (via storage) <-------
2019-10-20 08:52:25 copying disk images
full send of tank/vm-101-disk-1_00000@__migration__ estimated size is 8.20G
total estimated size is 8.20G
TIME SENT SNAPSHOT tank/vm-101-disk-1_00000@__migration__
tank/vm-101-disk-1_00000 name tank/vm-101-disk-1_00000 -
volume 'tank/vm-101-disk-1_00000' already exists
command 'zfs send -Rpv -- tank/vm-101-disk-1_00000@__migration__' failed: got signal 13
send/receive failed, cleaning up snapshot(s)..
2019-10-20 08:52:27 ERROR: Failed to sync data - command 'set -o pipefail && pvesm export zp1:vm-101-disk-1_00000 zfs - -with-snapshots 0 -snapshot __migration__ | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=ich2' root@10.17.0.72 -- pvesm import zp1:vm-101-disk-1_00000 zfs - -with-snapshots 0 -delete-snapshot __migration__' failed: exit code 255
2019-10-20 08:52:27 aborting phase 1 - cleanup resources
2019-10-20 08:52:27 ERROR: found stale volume copy 'zp1:vm-101-disk-1_00000' on node 'ich2' <------- This is no stale volume copy, it's the DRBD replica
2019-10-20 08:52:27 ERROR: migration aborted (duration 00:00:02): Failed to sync data - command 'set -o pipefail && pvesm export zp1:vm-101-disk-1_00000 zfs - -with-snapshots 0 -snapshot __migration__ | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=ich2' root@10.17.0.72 -- pvesm import zp1:vm-101-disk-1_00000 zfs - -with-snapshots 0 -delete-snapshot __migration__' failed: exit code 255
TASK ERROR: migration aborted
Anyone tried something of this sort ? Any pointers ?
Thank you!
marcelo
I'm trying out Proxmox VE ( 6.0-9 ) plus Linstor Plugin as a Datacenter Virtualization Solution for the company I work for. So far, pretty good. I've found a glitch though, and as I dont really know if it's about proxmox or about Linstor + ZFS, I'll post this in both communities. My scenario:
- 2 clustered nodes for testing
- Both feature a ZFS pool ( tank - to be original
- Installed and configured Linstor + DRBD9 + Controller in both ( node1 as Controller )
- Created DRBD storage pool based on ZFS ( tried both thick & thin )
So, to sum up, I'm using a ZFS pool on both nodes as VM image storage and as a storage pool from which to create my DRBD replicated volumes. I then created a couple VMs, one of them backed by a ZFS volume, and the other one backed by a DRBD replicated resource.
Here's the glitch: The ZFS backed VM I can migrate, no problem, but the DRBD backed VM, it wont migrate. When trying to, it looks like proxmox incorrectly detects it's backing storage as a ZFS volume, instead of the DRBD resource. VM's local disk is: "drbdstorage:vm-101-disk-1". I'm wondering if using the same ZFS pool for both VMs and Linstor is not supported for this setup . . .
This is my storage.cfg:
drbd: drbdstorage
content rootdir,images
controller 10.17.0.71
redundancy 2
zfspool: zp1
pool tank
content images,rootdir
sparse 1
Here's my problematic VM's config:
agent: 1
bootdisk: scsi0
cores: 2
ide2: none,media=cdrom
memory: 2048
name: vm-xenial-server
net0: virtio=06:04:B3:E1:75:AF,bridge=vmbr0,tag=1
numa: 0
ostype: l26
scsi0: drbdstorage:vm-101-disk-1,cache=unsafe,size=8G
scsihw: virtio-scsi-pci
smbios1: uuid=9a8e8301-a865-4544-a367-5111cfcbff16
sockets: 2
vmgenid: 2b7a93b0-f808-44a5-818a-cdfb4a70586e
Here's the migration log extract:
2019-10-20 08:52:25 use dedicated network address for sending migration traffic (10.17.0.72)
2019-10-20 08:52:25 starting migration of VM 101 to node 'ich2' (10.17.0.72)
-------> 2019-10-20 08:52:25 found local disk 'zp1:vm-101-disk-1_00000' (via storage) <-------
2019-10-20 08:52:25 copying disk images
full send of tank/vm-101-disk-1_00000@__migration__ estimated size is 8.20G
total estimated size is 8.20G
TIME SENT SNAPSHOT tank/vm-101-disk-1_00000@__migration__
tank/vm-101-disk-1_00000 name tank/vm-101-disk-1_00000 -
volume 'tank/vm-101-disk-1_00000' already exists
command 'zfs send -Rpv -- tank/vm-101-disk-1_00000@__migration__' failed: got signal 13
send/receive failed, cleaning up snapshot(s)..
2019-10-20 08:52:27 ERROR: Failed to sync data - command 'set -o pipefail && pvesm export zp1:vm-101-disk-1_00000 zfs - -with-snapshots 0 -snapshot __migration__ | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=ich2' root@10.17.0.72 -- pvesm import zp1:vm-101-disk-1_00000 zfs - -with-snapshots 0 -delete-snapshot __migration__' failed: exit code 255
2019-10-20 08:52:27 aborting phase 1 - cleanup resources
2019-10-20 08:52:27 ERROR: found stale volume copy 'zp1:vm-101-disk-1_00000' on node 'ich2' <------- This is no stale volume copy, it's the DRBD replica
2019-10-20 08:52:27 ERROR: migration aborted (duration 00:00:02): Failed to sync data - command 'set -o pipefail && pvesm export zp1:vm-101-disk-1_00000 zfs - -with-snapshots 0 -snapshot __migration__ | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=ich2' root@10.17.0.72 -- pvesm import zp1:vm-101-disk-1_00000 zfs - -with-snapshots 0 -delete-snapshot __migration__' failed: exit code 255
TASK ERROR: migration aborted
Anyone tried something of this sort ? Any pointers ?
Thank you!
marcelo