another volume 'dpool2' already exists

jice

New Member
Feb 14, 2024
4
1
3
Hi all. I've searched this forum but no solution helped me. Here the point:
I have 2 nodes. PVE-01 and PVE-04. and virtual machine 103 on pve01 with 3 storages/3 disc's
I have replication from pve01 to pve004 that did not work.

Code:
2024-02-14 09:59:00 103-0: start replication job
2024-02-14 09:59:00 103-0: guest => VM 103, running => 1900724
2024-02-14 09:59:00 103-0: volumes => VM-Data-2:vm-103-disk-0,VM-RabbitMQ-2:vm-103-disk-0,VM-System-2:vm-103-disk-0
2024-02-14 09:59:01 103-0: freeze guest filesystem
2024-02-14 09:59:01 103-0: create snapshot '__replicate_103-0_1707893940__' on VM-Data-2:vm-103-disk-0
2024-02-14 09:59:01 103-0: create snapshot '__replicate_103-0_1707893940__' on VM-RabbitMQ-2:vm-103-disk-0
2024-02-14 09:59:01 103-0: create snapshot '__replicate_103-0_1707893940__' on VM-System-2:vm-103-disk-0
2024-02-14 09:59:02 103-0: thaw guest filesystem
2024-02-14 09:59:02 103-0: using secure transmission, rate limit: 500 MByte/s
2024-02-14 09:59:02 103-0: full sync 'VM-Data-2:vm-103-disk-0' (__replicate_103-0_1707893940__)
2024-02-14 09:59:02 103-0: using a bandwidth limit of 500000000 bytes per second for transferring 'VM-Data-2:vm-103-disk-0'
2024-02-14 09:59:03 103-0: full send of dpool2/vm-103-disk-0@__replicate_103-0_1707893940__ estimated size is 1.20T
2024-02-14 09:59:03 103-0: total estimated size is 1.20T
2024-02-14 09:59:03 103-0: TIME        SENT   SNAPSHOT dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 09:59:04 103-0: 09:59:04   1.18M   dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 09:59:05 103-0: 09:59:05   64.7M   dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 09:59:06 103-0: 09:59:06    128M   dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
....
....
2024-02-14 13:28:18 103-0: 13:28:18   1.22T   dpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 13:28:19 103-0: successfully imported 'VM-Data-2:vm-103-disk-0'
2024-02-14 13:28:19 103-0: full sync 'VM-RabbitMQ-2:vm-103-disk-0' (__replicate_103-0_1707893940__)
2024-02-14 13:28:19 103-0: using a bandwidth limit of 500000000 bytes per second for transferring 'VM-RabbitMQ-2:vm-103-disk-0'
2024-02-14 13:28:20 103-0: full send of rpool2/vm-103-disk-0@__replicate_103-0_1707893940__ estimated size is 20.5G
2024-02-14 13:28:20 103-0: total estimated size is 20.5G
2024-02-14 13:28:20 103-0: volume 'rpool2/vm-103-disk-0' already exists
2024-02-14 13:28:20 103-0: TIME        SENT   SNAPSHOT rpool2/vm-103-disk-0@__replicate_103-0_1707893940__
2024-02-14 13:28:20 103-0: 1204 B 1.2 KB 0.99 s 1211 B/s 1.18 KB/s
2024-02-14 13:28:20 103-0: write: Broken pipe
2024-02-14 13:28:20 103-0: warning: cannot send 'rpool2/vm-103-disk-0@__replicate_103-0_1707893940__': signal received
2024-02-14 13:28:20 103-0: cannot send 'rpool2/vm-103-disk-0': I/O error
2024-02-14 13:28:20 103-0: command 'zfs send -Rpv -- rpool2/vm-103-disk-0@__replicate_103-0_1707893940__' failed: exit code 1
2024-02-14 13:28:21 103-0: delete previous replication snapshot '__replicate_103-0_1707893940__' on VM-Data-2:vm-103-disk-0
2024-02-14 13:28:22 103-0: delete previous replication snapshot '__replicate_103-0_1707893940__' on VM-RabbitMQ-2:vm-103-disk-0
2024-02-14 13:28:23 103-0: delete previous replication snapshot '__replicate_103-0_1707893940__' on VM-System-2:vm-103-disk-0
2024-02-14 13:28:24 103-0: end replication job with error: command 'set -o pipefail && pvesm export VM-RabbitMQ-2:vm-103-disk-0 zfs - -with-snapshots 1 -snapshot __replicate_103-0_1707906600__ | /usr/bin/cstream -t 500000000 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=brd-pve04' root@192.168.91.5 -- pvesm import VM-Data-2:vm-103-disk-0 zfs - -with-snapshots 1 -snapshot __replicate_103-0_1707906600__ -allow-rename 0' failed: exit code 255

my pve:

Code:
pveversion -v
proxmox-ve: 8.1.0 (running kernel: 6.2.16-6-pve)
pve-manager: 8.1.3 (running version: 8.1.3/b46aac3b42da5d15)
proxmox-kernel-helper: 8.1.0
proxmox-kernel-6.5: 6.5.11-7
proxmox-kernel-6.5.11-7-pve-signed: 6.5.11-7
proxmox-kernel-6.5.11-6-pve-signed: 6.5.11-6
pve-kernel-5.4: 6.4-20
proxmox-kernel-6.2.16-20-pve: 6.2.16-20
proxmox-kernel-6.2: 6.2.16-20
proxmox-kernel-6.2.16-6-pve: 6.2.16-7
pve-kernel-5.4.203-1-pve: 5.4.203-1
pve-kernel-5.4.73-1-pve: 5.4.73-1
ceph-fuse: 17.2.7-pve1
corosync: 3.1.7-pve3
criu: 3.17.1-2
glusterfs-client: 10.3-5
ifupdown: residual config
ifupdown2: 3.2.0-1+pmx7
ksm-control-daemon: 1.4-1
libjs-extjs: 7.0.0-4
libknet1: 1.28-pve1
libproxmox-acme-perl: 1.5.0
libproxmox-backup-qemu0: 1.4.1
libproxmox-rs-perl: 0.3.3
libpve-access-control: 8.0.7
libpve-apiclient-perl: 3.3.1
libpve-common-perl: 8.1.0
libpve-guest-common-perl: 5.0.6
libpve-http-server-perl: 5.0.5
libpve-network-perl: 0.9.5
libpve-rs-perl: 0.8.7
libpve-storage-perl: 8.0.5
libqb0: 1.0.5-1
libspice-server1: 0.15.1-1
lvm2: 2.03.16-2
lxc-pve: 5.0.2-4
lxcfs: 5.0.3-pve4
novnc-pve: 1.4.0-3
proxmox-backup-client: 3.1.2-1
proxmox-backup-file-restore: 3.1.2-1
proxmox-kernel-helper: 8.1.0
proxmox-mail-forward: 0.2.2
proxmox-mini-journalreader: 1.4.0
proxmox-offline-mirror-helper: 0.6.3
proxmox-widget-toolkit: 4.1.3
pve-cluster: 8.0.5
pve-container: 5.0.8
pve-docs: 8.1.3
pve-edk2-firmware: 4.2023.08-2
pve-firewall: 5.0.3
pve-firmware: 3.9-1
pve-ha-manager: 4.0.3
pve-i18n: 3.1.5
pve-qemu-kvm: 8.1.2-6
pve-xtermjs: 5.3.0-3
qemu-server: 8.0.10
smartmontools: 7.3-pve1
spiceterm: 3.3.0
swtpm: 0.8.0+pve1
vncterm: 1.8.0
zfsutils-linux: 2.2.2-pve1

My Steps:
remove replication task.
zfs list -t snapshot
zfs destroy -d dpool2/vm-103-disk-0@__replicate_103-0_1691751612__
remove from PVE04 - dpool2,spool2,rpool2 from datacenter.
destroy on pve04 all zfs.
create net zfs with same name's dpool2/spool2/rpool2 (because such names on PVE01)

start new replication job from scratch. wait about 10 hours (about 3 TB data)
came at morning and:

Code:
24-02-14 13:40:00 103-0: start replication job
2024-02-14 13:40:00 103-0: guest => VM 103, running => 1900724
2024-02-14 13:40:00 103-0: volumes => VM-Data-2:vm-103-disk-0,VM-RabbitMQ-2:vm-103-disk-0,VM-System-2:vm-103-disk-0
2024-02-14 13:40:01 103-0: freeze guest filesystem
2024-02-14 13:40:01 103-0: create snapshot '__replicate_103-0_1707907200__' on VM-Data-2:vm-103-disk-0
2024-02-14 13:40:01 103-0: create snapshot '__replicate_103-0_1707907200__' on VM-RabbitMQ-2:vm-103-disk-0
2024-02-14 13:40:01 103-0: create snapshot '__replicate_103-0_1707907200__' on VM-System-2:vm-103-disk-0
2024-02-14 13:40:02 103-0: thaw guest filesystem
2024-02-14 13:40:02 103-0: using secure transmission, rate limit: 500 MByte/s
2024-02-14 13:40:02 103-0: full sync 'VM-Data-2:vm-103-disk-0' (__replicate_103-0_1707907200__)
2024-02-14 13:40:03 103-0: using a bandwidth limit of 500000000 bytes per second for transferring 'VM-Data-2:vm-103-disk-0'
2024-02-14 13:40:03 103-0: full send of dpool2/vm-103-disk-0@__replicate_103-0_1707907200__ estimated size is 1.20T
2024-02-14 13:40:03 103-0: total estimated size is 1.20T
2024-02-14 13:40:03 103-0: TIME        SENT   SNAPSHOT dpool2/vm-103-disk-0@__replicate_103-0_1707907200__
2024-02-14 13:40:03 103-0: volume 'dpool2/vm-103-disk-0' already exists
2024-02-14 13:40:03 103-0: 27652 B 27.0 KB 0.67 s 41427 B/s 40.46 KB/s
2024-02-14 13:40:03 103-0: write: Broken pipe
2024-02-14 13:40:04 103-0: warning: cannot send 'dpool2/vm-103-disk-0@__replicate_103-0_1707907200__': signal received
2024-02-14 13:40:04 103-0: cannot send 'dpool2/vm-103-disk-0': I/O error
2024-02-14 13:40:04 103-0: command 'zfs send -Rpv -- dpool2/vm-103-disk-0@__replicate_103-0_1707907200__' failed: exit code 1
2024-02-14 13:40:04 103-0: delete previous replication snapshot '__replicate_103-0_1707907200__' on VM-Data-2:vm-103-disk-0
2024-02-14 13:40:04 103-0: delete previous replication snapshot '__replicate_103-0_1707907200__' on VM-RabbitMQ-2:vm-103-disk-0
2024-02-14 13:40:04 103-0: delete previous replication snapshot '__replicate_103-0_1707907200__' on VM-System-2:vm-103-disk-0
2024-02-14 13:40:05 103-0: end replication job with error: command 'set -o pipefail && pvesm export VM-Data-2:vm-103-disk-0 zfs - -with-snapshots 1 -sna


why this volume 'dpool2/vm-103-disk-0' already exists happend? sure it need to be exists because of replication..
I need help please
 
Last edited:
Are you using pve zfs replication? So you want to archieve a 2node zfs cluster with a external quorum vote to get a cluster?
You need to have the same pool name on both nodes. Im not sure if this will fix your error, but that is some thing that is definetly wrong.

You dont need to go to cli for zfs replication based setups.
 
Are you using pve zfs replication? So you want to archieve a 2node zfs cluster with a external quorum vote to get a cluster?
You need to have the same pool name on both nodes. Im not sure if this will fix your error, but that is some thing that is definetly wrong.

You dont need to go to cli for zfs replication based setups.
I have same names
- PVE01 dpool2
- PVE04 dpool2
did not help :(
 
Last edited:
So whats the spool and rpool about, they are still in your first post and seem to be important?
See also your log:
2024-02-14 13:28:20 103-0: TIME SENT SNAPSHOT rpool2/vm-103-disk-0@__replicate_103-0_1707893940__

Can you try it it with a new ressource, a blank container for example and see if the replication works? As I said, you dont need to do anything on cli. So dont mess around with zfs commands on cli if you dont need to.
 
Last edited:
So whats the spool and rpool about, they are still in your first post and seem to be important?
See also your log:


Can you try it it with a new ressource, a blank container for example and see if the replication works? As I said, you dont need to do anything on cli. So dont mess around with zfs commands on cli if you dont need to.
"blank container" you mean create another VM with empty disks?
 
  • Like
Reactions: jsterr
"blank container" you mean create another VM with empty disks?
No a new container with a template, I mean a new ressource (vm or container) not one that was used for replication before.
After creation please try replication only via UI and post the results here.
 
I found solution.

The trouble was that when I recteated new dpool I made only half of work. Because I need also delete all created containers. This is normal for ZFS but in webUI I cant do that (may be feature request?)
so I need use key -f
The solution was to:
zfs list
- check all drives
and tnen on target "PVE 04" -

zfs destroy dpool2/vm-103-disk-0 - delete containers. then rpool/spool same
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!