Someday there will be instructions for Storage Replication?

ldsjfasl'dfls

New Member
Sep 15, 2017
6
0
1
35
Already many times mentioned in various reviews that there is a new feature of Storage Replication.
https://pve.proxmox.com/wiki/Storage_Replication
But instructions how to organize it or where not. As if he did not try to do it himself, nothing happens. Many different variants of errors. Without a normal instruction, this function is completely meaningless.
 
If your Cluster configured correct you have no errors.

Nobody can help you without any hint or error statement.
 
Yes, I'm just talking about this opportunity and talking. Storage Replication.
The description of the fact that there was such an opportunity is, but how to configure it or where the instructions are not!
 
"If your Cluster configured correct you have no errors."
I'm just talking about this and there is no instruction on how to properly configure a cluster for Storage Replication.
 
This feature should work out of the box no special configuration are required.
Add a job as explained in the doku.

If you have a problems the are many possibilities why it will not work.
You have to analyse your problem or ask for help.
 
Now there are two mistakes:

unable to open file - No such file or directory


2017-09-14 13:10:23 108-0: start replication job
2017-09-14 13:10:23 108-0: guest => VM 108, running => 0
2017-09-14 13:10:23 108-0: volumes => zfs209nw:vm-108-disk-3
2017-09-14 13:10:24 108-0: (remote_prepare_local_job) could not activate storage 'zfs209nw', zfs error: Insecure $ENV{ENV} while running with -T switch at /usr/share/perl5/IPC/Open3.pm line 281.
2017-09-14 13:10:24 108-0: (remote_prepare_local_job)
2017-09-14 13:10:24 108-0: end replication job with error: command '/usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=srv-209NJ' root@172.15.11.202 -- pvesr prepare-local-job 108-0 zfs209nw:vm-108-disk-3 --last_sync 0' failed: exit code
 
This should not happend and looks like a bug.

Can you please post me the output of the following commands?
I need it form both nodes.

Code:
zpool status
 
[root@srv-209NW ~]# zpool status
pool: zfs209
state: ONLINE
scan: none requested
config:

NAME STATE READ WRITE CKSUM
zfs209 ONLINE 0 0 0
sdd ONLINE 0 0 0

errors: No known data errors
[root@srv-209NW ~]#


[root@srv-209NJ ~]# zpool status
pool: zfs209
state: ONLINE
scan: none requested
config:

NAME STATE READ WRITE CKSUM
zfs209 ONLINE 0 0 0
sdd ONLINE 0 0 0

errors: No known data errors
[root@srv-209NJ ~]#
 
Thanks can you also send me the versions of you nodes?

Code:
pveversion -v
 
[root@srv-209NW ~]# pveversion -v
pve-manager: 5.0.24-alt4.M80P.1 (running version: 5.0-24/)
lvm2: 2.02.165-alt0.M80P.1
corosync: 2.4.1-alt0.M80P.1
pve-cluster: 5.0.12-alt1.M80P.1
pve-docs: 5.0.9-alt1
pve-container: 2.0.15-alt4.M80P.1
pve-firewall: 3.0.2-alt4.M80P.1
pve-ha-manager: 2.0.2-alt4.M80P.1
lxcfs: 2.0.7-alt0.M80P.1
smartmontools: 6.5-alt1
ceph: 10.2.8-alt0.M80P.1
[root@srv-209NW ~]#


[root@srv-209NJ ~]# pveversion -v
pve-manager: 5.0.24-alt4.M80P.1 (running version: 5.0-24/)
lvm2: 2.02.165-alt0.M80P.1
corosync: 2.4.1-alt0.M80P.1
pve-cluster: 5.0.12-alt1.M80P.1
pve-docs: 5.0.9-alt1
pve-container: 2.0.15-alt4.M80P.1
pve-firewall: 3.0.2-alt4.M80P.1
pve-ha-manager: 2.0.2-alt4.M80P.1
lxcfs: 2.0.7-alt0.M80P.1
smartmontools: 6.5-alt1
ceph: 10.2.8-alt0.M80P.1
[root@srv-209NJ ~]#
 
You do not run Promox VE, you are using Alt Linux please contact them.
This is a bug and can't fixed by us.
 
If your Cluster configured correct you have no errors.

Nobody can help you without any hint or error statement.

Let me bite on this after I've tried it:
Code:
2017-09-17 22:02:01 405-0: volumes => TracsSSD01zle:vm-405-disk-1,TracsZPool01:base-904-disk-3/vm-405-disk-3,TracsZPool01zle:vm-405-disk-1
2017-09-17 22:02:02 405-0: (remote_prepare_local_job) storage 'TracsSSD01zle' is not available on node 'proxtracs02'
2017-09-17 22:02:02 405-0: end replication job with error: command '/usr/bin/ssh -o 'BatchMode=yes' -o 'HostKeyAlias=proxtracs02' root@193.70.9.90 -- pvesr prepare-local-job 405-0 TracsSSD01zle:vm-405-disk-1 TracsZPool01:base-904-disk-3/vm-405-disk-3 TracsZPool01zle:vm-405-disk-1 --last_sync 0 --parent_snapname P20170917_2147_Rob_afterRuns' failed: exit code 255

So, the first problem in the documentation: No assumptions on the setup of the ZPOOLs, nor the StorageConfigurations.
I configured my storage in the cluster specifically name based on the local node's name, ie. Tracs01 & Tracs02 the two nodes with their local ZFS pools named differently, so when I create the scripts that will create the VMs (using `qm` ) it helps to catch errors when you try to create/use the wrong storage on the wrong hypervisor and when I compare reports of storage usage, it's easier to differentiate.

So, let's try next to create a "duplicate" named storage pool on Tracs02, with the same name as on Tracs01:
Code:
create storage failed: storage ID 'TracsSSD01zle' already defined (500)

Right, at present I'm stumped, as I haven't seen any documentation that explains the assumptions and how it'll impact my setup, or what I should've set up in my setup to make it work?
 
  • Like
Reactions: Tommmii