I have used the instructions (https://pve.proxmox.com/wiki/Cluster_Manager) to create a cluster:
hp1# pvecm create YOUR-CLUSTER-NAME
hp2# pvecm add IP-ADDRESS-CLUSTER
The cluster seems to be working as I see the other nodes:
However, when I setup the replication for a CT, it fails (see below). Anybody have an idea what I might be doing wrong?
2017-11-29 22:52:00 19101-0: start replication job
2017-11-29 22:52:00 19101-0: guest => CT 19101, running => 0
2017-11-29 22:52:00 19101-0: volumes => clusterpool:subvol-19101-disk-1
2017-11-29 22:52:01 19101-0: create snapshot '__replicate_19101-0_1512013920__' on clusterpool:subvol-19101-disk-1
2017-11-29 22:52:02 19101-0: full sync 'clusterpool:subvol-19101-disk-1' (__replicate_19101-0_1512013920__)
2017-11-29 22:52:02 19101-0: internal error: Invalid argument
2017-11-29 22:52:02 19101-0: command 'zfs send -Rpv -- rpool/subvol-19101-disk-1@__replicate_19101-0_1512013920__' failed: got signal 6
2017-11-29 22:52:02 19101-0: cannot receive: failed to read from stream
2017-11-29 22:52:02 19101-0: cannot open 'rpool/subvol-19101-disk-1': dataset does not exist
2017-11-29 22:52:02 19101-0: command 'zfs recv -F -- rpool/subvol-19101-disk-1' failed: exit code 1
2017-11-29 22:52:02 19101-0: delete previous replication snapshot '__replicate_19101-0_1512013920__' on clusterpool:subvol-19101-disk-1
2017-11-29 22:52:02 19101-0: end replication job with error: command 'set -o pipefail && pvesm export clusterpool:subvol-19101-disk-1 zfs - -with-snapshots 1 -snapshot __replicate_19101-0_1512013920__ | /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=vm192' root@10.99.0.192 -- pvesm import clusterpool:subvol-19101-disk-1 zfs - -with-snapshots 1' failed: exit code 1
pvecm status:
Quorum information
------------------
Date: Wed Nov 29 22:47:45 2017
Quorum provider: corosync_votequorum
Nodes: 3
Node ID: 0x00000001
Ring ID: 1/12
Quorate: Yes
Votequorum information
----------------------
Expected votes: 3
Highest expected: 3
Total votes: 3
Quorum: 2
Flags: Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000001 1 10.99.0.191 (local)
0x00000002 1 10.99.0.192
0x00000003 1 10.99.0.193
cat /etc/pve/storage.cfg:
dir: local
path /var/lib/vz
content backup,iso,vztmpl
zfspool: local-zfs
pool rpool/data
content rootdir,images
sparse 1
zfspool: clusterpool
pool rpool
content rootdir,images
nodes vm192,vm193,vm191
sparse 1
pveversion:
pve-manager/5.1-36/131401db (running kernel: 4.10.15-1-pve)
hp1# pvecm create YOUR-CLUSTER-NAME
hp2# pvecm add IP-ADDRESS-CLUSTER
The cluster seems to be working as I see the other nodes:
However, when I setup the replication for a CT, it fails (see below). Anybody have an idea what I might be doing wrong?
2017-11-29 22:52:00 19101-0: start replication job
2017-11-29 22:52:00 19101-0: guest => CT 19101, running => 0
2017-11-29 22:52:00 19101-0: volumes => clusterpool:subvol-19101-disk-1
2017-11-29 22:52:01 19101-0: create snapshot '__replicate_19101-0_1512013920__' on clusterpool:subvol-19101-disk-1
2017-11-29 22:52:02 19101-0: full sync 'clusterpool:subvol-19101-disk-1' (__replicate_19101-0_1512013920__)
2017-11-29 22:52:02 19101-0: internal error: Invalid argument
2017-11-29 22:52:02 19101-0: command 'zfs send -Rpv -- rpool/subvol-19101-disk-1@__replicate_19101-0_1512013920__' failed: got signal 6
2017-11-29 22:52:02 19101-0: cannot receive: failed to read from stream
2017-11-29 22:52:02 19101-0: cannot open 'rpool/subvol-19101-disk-1': dataset does not exist
2017-11-29 22:52:02 19101-0: command 'zfs recv -F -- rpool/subvol-19101-disk-1' failed: exit code 1
2017-11-29 22:52:02 19101-0: delete previous replication snapshot '__replicate_19101-0_1512013920__' on clusterpool:subvol-19101-disk-1
2017-11-29 22:52:02 19101-0: end replication job with error: command 'set -o pipefail && pvesm export clusterpool:subvol-19101-disk-1 zfs - -with-snapshots 1 -snapshot __replicate_19101-0_1512013920__ | /usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=vm192' root@10.99.0.192 -- pvesm import clusterpool:subvol-19101-disk-1 zfs - -with-snapshots 1' failed: exit code 1
pvecm status:
Quorum information
------------------
Date: Wed Nov 29 22:47:45 2017
Quorum provider: corosync_votequorum
Nodes: 3
Node ID: 0x00000001
Ring ID: 1/12
Quorate: Yes
Votequorum information
----------------------
Expected votes: 3
Highest expected: 3
Total votes: 3
Quorum: 2
Flags: Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000001 1 10.99.0.191 (local)
0x00000002 1 10.99.0.192
0x00000003 1 10.99.0.193
cat /etc/pve/storage.cfg:
dir: local
path /var/lib/vz
content backup,iso,vztmpl
zfspool: local-zfs
pool rpool/data
content rootdir,images
sparse 1
zfspool: clusterpool
pool rpool
content rootdir,images
nodes vm192,vm193,vm191
sparse 1
pveversion:
pve-manager/5.1-36/131401db (running kernel: 4.10.15-1-pve)