Hi,
I recently reinstalled to PVE 5.1 and have been using a desktop as backup when doing changes to the server.
Migrating between the two has worked flawlessly (2-cluster with zfs), but for some reason the replication decided to send a full image while it already exists on the other side:
This unfortunately prevents me from migrating back to the server. I can remove the volume for this machine, but I have others in 3-500GB range which I would like not to "reset" in this way. Can I do anything to get the replication back to normal?
Thanks in advance,
Bo
Output from failed migration (but I suspect the failing replication above as root cause)
desktop pveversion:
Server pveversion (subscribed):
I recently reinstalled to PVE 5.1 and have been using a desktop as backup when doing changes to the server.
Migrating between the two has worked flawlessly (2-cluster with zfs), but for some reason the replication decided to send a full image while it already exists on the other side:
2017-11-17 07:06:59 204-1: start replication job
2017-11-17 07:06:59 204-1: guest => CT 204, running => 0
2017-11-17 07:06:59 204-1: volumes => local-zfs:subvol-204-disk-1
2017-11-17 07:07:00 204-1: create snapshot '__replicate_204-1_1510898819__' on local-zfs:subvol-204-disk-1
2017-11-17 07:07:00 204-1: full sync 'local-zfs:subvol-204-disk-1' (__replicate_204-1_1510898819__)
2017-11-17 07:07:00 204-1: full send of rpool/data/subvol-204-disk-1@__replicate_204-0_1510573211__ estimated size is 537M
2017-11-17 07:07:00 204-1: send from @__replicate_204-0_1510573211__ to rpool/data/subvol-204-disk-1@__replicate_204-1_1510898819__ estimated size is 1.60M
2017-11-17 07:07:00 204-1: total estimated size is 539M
2017-11-17 07:07:00 204-1: TIME SENT SNAPSHOT
2017-11-17 07:07:01 204-1: rpool/data/subvol-204-disk-1 name rpool/data/subvol-204-disk-1 -
2017-11-17 07:07:01 204-1: volume 'rpool/data/subvol-204-disk-1' already exists
2017-11-17 07:07:01 204-1: command 'zfs send -Rpv -- rpool/data/subvol-204-disk-1@__replicate_204-1_1510898819__' failed: got signal 13
2017-11-17 07:07:01 204-1: delete previous replication snapshot '__replicate_204-1_1510898819__' on local-zfs:subvol-204-disk-1
2017-11-17 07:07:01 204-1: end replication job with error: command 'set -o pipefail && pvesm export local-zfs:subvol-204-disk-1 zfs - -with-snapshots 1 -snapshot __replicate_204-1_1510898819__ | /usr/bin/cstream -t 10000000 | /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=pve' root@192.168.110.2 -- pvesm import local-zfs:subvol-204-disk-1 zfs - -with-snapshots 1' failed: exit code 255
This unfortunately prevents me from migrating back to the server. I can remove the volume for this machine, but I have others in 3-500GB range which I would like not to "reset" in this way. Can I do anything to get the replication back to normal?
Thanks in advance,
Bo
Output from failed migration (but I suspect the failing replication above as root cause)
2017-11-17 07:06:53 shutdown CT 204
2017-11-17 07:06:59 starting migration of CT 204 to node 'pve' (192.168.110.2)
2017-11-17 07:06:59 found local volume 'local-zfs:subvol-204-disk-1' (in current VM config)
2017-11-17 07:06:59 start replication job
2017-11-17 07:06:59 guest => CT 204, running => 0
2017-11-17 07:06:59 volumes => local-zfs:subvol-204-disk-1
2017-11-17 07:07:00 create snapshot '__replicate_204-1_1510898819__' on local-zfs:subvol-204-disk-1
2017-11-17 07:07:00 full sync 'local-zfs:subvol-204-disk-1' (__replicate_204-1_1510898819__)
2017-11-17 07:07:00 full send of rpool/data/subvol-204-disk-1@__replicate_204-0_1510573211__ estimated size is 537M
2017-11-17 07:07:00 send from @__replicate_204-0_1510573211__ to rpool/data/subvol-204-disk-1@__replicate_204-1_1510898819__ estimated size is 1.60M
2017-11-17 07:07:00 total estimated size is 539M
2017-11-17 07:07:00 TIME SENT SNAPSHOT
2017-11-17 07:07:01 rpool/data/subvol-204-disk-1 name rpool/data/subvol-204-disk-1 -
2017-11-17 07:07:01 volume 'rpool/data/subvol-204-disk-1' already exists
2017-11-17 07:07:01 command 'zfs send -Rpv -- rpool/data/subvol-204-disk-1@__replicate_204-1_1510898819__' failed: got signal 13
send/receive failed, cleaning up snapshot(s)..
2017-11-17 07:07:01 delete previous replication snapshot '__replicate_204-1_1510898819__' on local-zfs:subvol-204-disk-1
2017-11-17 07:07:01 end replication job with error: command 'set -o pipefail && pvesm export local-zfs:subvol-204-disk-1 zfs - -with-snapshots 1 -snapshot __replicate_204-1_1510898819__ | /usr/bin/cstream -t 10000000 | /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=pve' root@192.168.110.2 -- pvesm import local-zfs:subvol-204-disk-1 zfs - -with-snapshots 1' failed: exit code 255
2017-11-17 07:07:01 ERROR: command 'set -o pipefail && pvesm export local-zfs:subvol-204-disk-1 zfs - -with-snapshots 1 -snapshot __replicate_204-1_1510898819__ | /usr/bin/cstream -t 10000000 | /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=pve' root@192.168.110.2 -- pvesm import local-zfs:subvol-204-disk-1 zfs - -with-snapshots 1' failed: exit code 255
2017-11-17 07:07:01 aborting phase 1 - cleanup resources
2017-11-17 07:07:01 start final cleanup
2017-11-17 07:07:01 ERROR: migration aborted (duration 00:00:08): command 'set -o pipefail && pvesm export local-zfs:subvol-204-disk-1 zfs - -with-snapshots 1 -snapshot __replicate_204-1_1510898819__ | /usr/bin/cstream -t 10000000 | /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=pve' root@192.168.110.2 -- pvesm import local-zfs:subvol-204-disk-1 zfs - -with-snapshots 1' failed: exit code 255
TASK ERROR: migration aborted
desktop pveversion:
proxmox-ve: 5.1-26 (running kernel: 4.13.4-1-pve)
pve-manager: 5.1-36 (running version: 5.1-36/131401db)
pve-kernel-4.13.4-1-pve: 4.13.4-26
pve-kernel-4.10.15-1-pve: 4.10.15-15
libpve-http-server-perl: 2.0-6
lvm2: 2.02.168-pve6
corosync: 2.4.2-pve3
libqb0: 1.0.1-1
pve-cluster: 5.0-15
qemu-server: 5.0-17
pve-firmware: 2.0-3
libpve-common-perl: 5.0-20
libpve-guest-common-perl: 2.0-13
libpve-access-control: 5.0-7
libpve-storage-perl: 5.0-16
pve-libspice-server1: 0.12.8-3
vncterm: 1.5-2
pve-docs: 5.1-12
pve-qemu-kvm: 2.9.1-2
pve-container: 2.0-17
pve-firewall: 3.0-3
pve-ha-manager: 2.0-3
ksm-control-daemon: 1.2-2
glusterfs-client: 3.8.8-1
lxc-pve: 2.1.0-2
lxcfs: 2.0.7-pve4
criu: 2.11.1-1~bpo90
novnc-pve: 0.6-4
smartmontools: 6.5+svn4324-1
zfsutils-linux: 0.7.3-pve1~bpo9
Server pveversion (subscribed):
proxmox-ve: 5.1-26 (running kernel: 4.13.4-1-pve)
pve-manager: 5.1-36 (running version: 5.1-36/131401db)
pve-kernel-4.13.4-1-pve: 4.13.4-26
libpve-http-server-perl: 2.0-6
lvm2: 2.02.168-pve6
corosync: 2.4.2-pve3
libqb0: 1.0.1-1
pve-cluster: 5.0-15
qemu-server: 5.0-17
pve-firmware: 2.0-3
libpve-common-perl: 5.0-20
libpve-guest-common-perl: 2.0-13
libpve-access-control: 5.0-7
libpve-storage-perl: 5.0-16
pve-libspice-server1: 0.12.8-3
vncterm: 1.5-2
pve-docs: 5.1-12
pve-qemu-kvm: 2.9.1-2
pve-container: 2.0-17
pve-firewall: 3.0-3
pve-ha-manager: 2.0-3
ksm-control-daemon: 1.2-2
glusterfs-client: 3.8.8-1
lxc-pve: 2.1.0-2
lxcfs: 2.0.7-pve4
criu: 2.11.1-1~bpo90
novnc-pve: 0.6-4
smartmontools: 6.5+svn4324-1
zfsutils-linux: 0.7.3-pve1~bpo9
Last edited: