Hi all. I've searched this forum but no solution helped me. Here the point:
I have 2 nodes. PVE-01 and PVE-04. and virtual machine 103 on pve01 with 3 storages/3 disc's
I have replication from pve01 to pve004 that did not work.
my pve:
My Steps:
remove replication task.
zfs list -t snapshot
zfs destroy -d dpool2/vm-103-disk-0@__replicate_103-0_1691751612__
remove from PVE04 - dpool2,spool2,rpool2 from datacenter.
destroy on pve04 all zfs.
create net zfs with same name's dpool2/spool2/rpool2 (because such names on PVE01)
start new replication job from scratch. wait about 10 hours (about 3 TB data)
came at morning and:
why this volume 'dpool2/vm-103-disk-0' already exists happend? sure it need to be exists because of replication..
I need help please
I have 2 nodes. PVE-01 and PVE-04. and virtual machine 103 on pve01 with 3 storages/3 disc's
I have replication from pve01 to pve004 that did not work.
Code:
2024-02-14 09:59:00 103-0: start replication job
2024-02-14 09:59:00 103-0: guest => VM 103, running => 1900724
2024-02-14 09:59:00 103-0: volumes => VM-Data-2:vm-103-disk-0,VM-RabbitMQ-2:vm-103-disk-0,VM-System-2:vm-103-disk-0
2024-02-14 09:59:01 103-0: freeze guest filesystem
2024-02-14 09:59:01 103-0: create snapshot '__replicate_103-0_1707893940__' on VM-Data-2:vm-103-disk-0
2024-02-14 09:59:01 103-0: create snapshot '__replicate_103-0_1707893940__' on VM-RabbitMQ-2:vm-103-disk-0
2024-02-14 09:59:01 103-0: create snapshot '__replicate_103-0_1707893940__' on VM-System-2:vm-103-disk-0
2024-02-14 09:59:02 103-0: thaw guest filesystem
2024-02-14 09:59:02 103-0: using secure transmission, rate limit: 500 MByte/s
2024-02-14 09:59:02 103-0: full sync 'VM-Data-2:vm-103-disk-0' (__replicate_103-0_1707893940__)
2024-02-14 09:59:02 103-0: using a bandwidth limit of 500000000 bytes per second for transferring 'VM-Data-2:vm-103-disk-0'
2024-02-14 09:59:03 103-0: full send of dpool2/vm-103-disk-0@__replicate_103-0_1707893940__ estimated size is 1.20T
2024-02-14 09:59:03 103-0: total estimated size is 1.20T
2024-02-14 09:59:03 103-0: TIME SENT SNAPSHOT dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 09:59:04 103-0: 09:59:04 1.18M dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 09:59:05 103-0: 09:59:05 64.7M dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 09:59:06 103-0: 09:59:06 128M dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
....
....
2024-02-14 13:28:18 103-0: 13:28:18 1.22T dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 13:28:19 103-0: successfully imported 'VM-Data-2:vm-103-disk-0'
2024-02-14 13:28:19 103-0: full sync 'VM-RabbitMQ-2:vm-103-disk-0' (__replicate_103-0_1707893940__)
2024-02-14 13:28:19 103-0: using a bandwidth limit of 500000000 bytes per second for transferring 'VM-RabbitMQ-2:vm-103-disk-0'
2024-02-14 13:28:20 103-0: full send of rpool2/vm-103-disk-0@__replicate_103-0_1707893940__ estimated size is 20.5G
2024-02-14 13:28:20 103-0: total estimated size is 20.5G
2024-02-14 13:28:20 103-0: volume 'rpool2/vm-103-disk-0' already exists
2024-02-14 13:28:20 103-0: TIME SENT SNAPSHOT rpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 13:28:20 103-0: 1204 B 1.2 KB 0.99 s 1211 B/s 1.18 KB/s
2024-02-14 13:28:20 103-0: write: Broken pipe
2024-02-14 13:28:20 103-0: warning: cannot send 'rpool2/vm-103-disk-0@__replicate_103-0_1707893940__': signal received
2024-02-14 13:28:20 103-0: cannot send 'rpool2/vm-103-disk-0': I/O error
2024-02-14 13:28:20 103-0: command 'zfs send -Rpv -- rpool2/vm-103-disk-0@__replicate_103-0_1707893940__' failed: exit code 1
2024-02-14 13:28:21 103-0: delete previous replication snapshot '__replicate_103-0_1707893940__' on VM-Data-2:vm-103-disk-0
2024-02-14 13:28:22 103-0: delete previous replication snapshot '__replicate_103-0_1707893940__' on VM-RabbitMQ-2:vm-103-disk-0
2024-02-14 13:28:23 103-0: delete previous replication snapshot '__replicate_103-0_1707893940__' on VM-System-2:vm-103-disk-0
2024-02-14 13:28:24 103-0: end replication job with error: command 'set -o pipefail && pvesm export VM-RabbitMQ-2:vm-103-disk-0 zfs - -with-snapshots 1 -snapshot __replicate_103-0_1707906600__ | /usr/bin/cstream -t 500000000 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=brd-pve04' root@192.168.91.5 -- pvesm import VM-Data-2:vm-103-disk-0 zfs - -with-snapshots 1 -snapshot __replicate_103-0_1707906600__ -allow-rename 0' failed: exit code 255
my pve:
Code:
pveversion -v
proxmox-ve: 8.1.0 (running kernel: 6.2.16-6-pve)
pve-manager: 8.1.3 (running version: 8.1.3/b46aac3b42da5d15)
proxmox-kernel-helper: 8.1.0
proxmox-kernel-6.5: 6.5.11-7
proxmox-kernel-6.5.11-7-pve-signed: 6.5.11-7
proxmox-kernel-6.5.11-6-pve-signed: 6.5.11-6
pve-kernel-5.4: 6.4-20
proxmox-kernel-6.2.16-20-pve: 6.2.16-20
proxmox-kernel-6.2: 6.2.16-20
proxmox-kernel-6.2.16-6-pve: 6.2.16-7
pve-kernel-5.4.203-1-pve: 5.4.203-1
pve-kernel-5.4.73-1-pve: 5.4.73-1
ceph-fuse: 17.2.7-pve1
corosync: 3.1.7-pve3
criu: 3.17.1-2
glusterfs-client: 10.3-5
ifupdown: residual config
ifupdown2: 3.2.0-1+pmx7
ksm-control-daemon: 1.4-1
libjs-extjs: 7.0.0-4
libknet1: 1.28-pve1
libproxmox-acme-perl: 1.5.0
libproxmox-backup-qemu0: 1.4.1
libproxmox-rs-perl: 0.3.3
libpve-access-control: 8.0.7
libpve-apiclient-perl: 3.3.1
libpve-common-perl: 8.1.0
libpve-guest-common-perl: 5.0.6
libpve-http-server-perl: 5.0.5
libpve-network-perl: 0.9.5
libpve-rs-perl: 0.8.7
libpve-storage-perl: 8.0.5
libqb0: 1.0.5-1
libspice-server1: 0.15.1-1
lvm2: 2.03.16-2
lxc-pve: 5.0.2-4
lxcfs: 5.0.3-pve4
novnc-pve: 1.4.0-3
proxmox-backup-client: 3.1.2-1
proxmox-backup-file-restore: 3.1.2-1
proxmox-kernel-helper: 8.1.0
proxmox-mail-forward: 0.2.2
proxmox-mini-journalreader: 1.4.0
proxmox-offline-mirror-helper: 0.6.3
proxmox-widget-toolkit: 4.1.3
pve-cluster: 8.0.5
pve-container: 5.0.8
pve-docs: 8.1.3
pve-edk2-firmware: 4.2023.08-2
pve-firewall: 5.0.3
pve-firmware: 3.9-1
pve-ha-manager: 4.0.3
pve-i18n: 3.1.5
pve-qemu-kvm: 8.1.2-6
pve-xtermjs: 5.3.0-3
qemu-server: 8.0.10
smartmontools: 7.3-pve1
spiceterm: 3.3.0
swtpm: 0.8.0+pve1
vncterm: 1.8.0
zfsutils-linux: 2.2.2-pve1
My Steps:
remove replication task.
zfs list -t snapshot
zfs destroy -d dpool2/vm-103-disk-0@__replicate_103-0_1691751612__
remove from PVE04 - dpool2,spool2,rpool2 from datacenter.
destroy on pve04 all zfs.
create net zfs with same name's dpool2/spool2/rpool2 (because such names on PVE01)
start new replication job from scratch. wait about 10 hours (about 3 TB data)
came at morning and:
Code:
24-02-14 13:40:00 103-0: start replication job
2024-02-14 13:40:00 103-0: guest => VM 103, running => 1900724
2024-02-14 13:40:00 103-0: volumes => VM-Data-2:vm-103-disk-0,VM-RabbitMQ-2:vm-103-disk-0,VM-System-2:vm-103-disk-0
2024-02-14 13:40:01 103-0: freeze guest filesystem
2024-02-14 13:40:01 103-0: create snapshot '__replicate_103-0_1707907200__' on VM-Data-2:vm-103-disk-0
2024-02-14 13:40:01 103-0: create snapshot '__replicate_103-0_1707907200__' on VM-RabbitMQ-2:vm-103-disk-0
2024-02-14 13:40:01 103-0: create snapshot '__replicate_103-0_1707907200__' on VM-System-2:vm-103-disk-0
2024-02-14 13:40:02 103-0: thaw guest filesystem
2024-02-14 13:40:02 103-0: using secure transmission, rate limit: 500 MByte/s
2024-02-14 13:40:02 103-0: full sync 'VM-Data-2:vm-103-disk-0' (__replicate_103-0_1707907200__)
2024-02-14 13:40:03 103-0: using a bandwidth limit of 500000000 bytes per second for transferring 'VM-Data-2:vm-103-disk-0'
2024-02-14 13:40:03 103-0: full send of dpool2/vm-103-disk-0@__replicate_103-0_1707907200__ estimated size is 1.20T
2024-02-14 13:40:03 103-0: total estimated size is 1.20T
2024-02-14 13:40:03 103-0: TIME SENT SNAPSHOT dpool2/vm-103-disk-0@__replicate_103-0_1707907200__
2024-02-14 13:40:03 103-0: volume 'dpool2/vm-103-disk-0' already exists
2024-02-14 13:40:03 103-0: 27652 B 27.0 KB 0.67 s 41427 B/s 40.46 KB/s
2024-02-14 13:40:03 103-0: write: Broken pipe
2024-02-14 13:40:04 103-0: warning: cannot send 'dpool2/vm-103-disk-0@__replicate_103-0_1707907200__': signal received
2024-02-14 13:40:04 103-0: cannot send 'dpool2/vm-103-disk-0': I/O error
2024-02-14 13:40:04 103-0: command 'zfs send -Rpv -- dpool2/vm-103-disk-0@__replicate_103-0_1707907200__' failed: exit code 1
2024-02-14 13:40:04 103-0: delete previous replication snapshot '__replicate_103-0_1707907200__' on VM-Data-2:vm-103-disk-0
2024-02-14 13:40:04 103-0: delete previous replication snapshot '__replicate_103-0_1707907200__' on VM-RabbitMQ-2:vm-103-disk-0
2024-02-14 13:40:04 103-0: delete previous replication snapshot '__replicate_103-0_1707907200__' on VM-System-2:vm-103-disk-0
2024-02-14 13:40:05 103-0: end replication job with error: command 'set -o pipefail && pvesm export VM-Data-2:vm-103-disk-0 zfs - -with-snapshots 1 -sna
why this volume 'dpool2/vm-103-disk-0' already exists happend? sure it need to be exists because of replication..
I need help please
Last edited: