Hello
Today I noticed a strange behaviour in pvesr, it started to run a job id every minute even if the job is set to run every 30 minutes. I had to disable the replication to stop it.
pveversion is the latest
My pvesr list (task is actually disabled or it will start replicaton every minute again):
replication.cfg
These are the latest pvesr status before I disabled the task
The replication end with OK but NextSync is always pending
These are the latest logs of the job
Is there something wrong in configuration or is pvesr bugging?
Today I noticed a strange behaviour in pvesr, it started to run a job id every minute even if the job is set to run every 30 minutes. I had to disable the replication to stop it.
pveversion is the latest
Code:
$ pveversion -v
proxmox-ve: 5.1-25 (running kernel: 4.13.4-1-pve)
pve-manager: 5.1-36 (running version: 5.1-36/131401db)
pve-kernel-4.13.4-1-pve: 4.13.4-25
pve-kernel-4.10.15-1-pve: 4.10.15-15
pve-kernel-4.10.17-3-pve: 4.10.17-23
libpve-http-server-perl: 2.0-6
lvm2: 2.02.168-pve6
corosync: 2.4.2-pve3
libqb0: 1.0.1-1
pve-cluster: 5.0-15
qemu-server: 5.0-17
pve-firmware: 2.0-3
libpve-common-perl: 5.0-20
libpve-guest-common-perl: 2.0-13
libpve-access-control: 5.0-7
libpve-storage-perl: 5.0-16
pve-libspice-server1: 0.12.8-3
vncterm: 1.5-2
pve-docs: 5.1-12
pve-qemu-kvm: 2.9.1-2
pve-container: 2.0-17
pve-firewall: 3.0-3
pve-ha-manager: 2.0-3
ksm-control-daemon: 1.2-2
glusterfs-client: 3.8.8-1
lxc-pve: 2.1.0-2
lxcfs: 2.0.7-pve4
criu: 2.11.1-1~bpo90
novnc-pve: 0.6-4
smartmontools: 6.5+svn4324-1
zfsutils-linux: 0.7.2-pve1~bpo90
ceph: 12.2.1-pve3
My pvesr list (task is actually disabled or it will start replicaton every minute again):
Code:
$ pvesr list
JobID Target Schedule Rate Enabled
105-0 local/pve-hs-3 */30 5 no
replication.cfg
Code:
$ cat /etc/pve/replication.cfg
local: 105-0
comment WIN2012
disable
target pve-hs-3
rate 5
schedule */30
These are the latest pvesr status before I disabled the task
Code:
$ pvesr status
JobID Enabled Target LastSync NextSync Duration FailCount State
105-0 Yes local/pve-hs-3 2017-10-29_06:22:00 pending 8.277227 0 OK
$ pvesr status
JobID Enabled Target LastSync NextSync Duration FailCount State
105-0 Yes local/pve-hs-3 2017-10-29_06:23:00 pending 7.40767 0 OK
$ pvesr status
JobID Enabled Target LastSync NextSync Duration FailCount State
105-0 Yes local/pve-hs-3 2017-10-29_06:24:00 pending 7.896934 0 OK
The replication end with OK but NextSync is always pending
These are the latest logs of the job
Code:
$ tail -f /var/log/pve/replicate/105-0 -n +1
2017-10-29 06:22:00 105-0: start replication job
2017-10-29 06:22:00 105-0: guest => VM 105, running => 27107
2017-10-29 06:22:00 105-0: volumes => local-zfs:vm-105-disk-1,local-zfs:vm-105-disk-2
2017-10-29 06:22:01 105-0: freeze guest filesystem
2017-10-29 06:22:02 105-0: create snapshot '__replicate_105-0_1509254520__' on local-zfs:vm-105-disk-1
2017-10-29 06:22:02 105-0: create snapshot '__replicate_105-0_1509254520__' on local-zfs:vm-105-disk-2
2017-10-29 06:22:03 105-0: thaw guest filesystem
2017-10-29 06:22:04 105-0: incremental sync 'local-zfs:vm-105-disk-1' (__replicate_105-0_1509254340__ => __replicate_105-0_1509254520__)
2017-10-29 06:22:04 105-0: send from @__replicate_105-0_1509254340__ to rpool/data/vm-105-disk-1@__replicate_105-0_1509254520__ estimated size is 4.43M
2017-10-29 06:22:04 105-0: total estimated size is 4.43M
2017-10-29 06:22:05 105-0: TIME SENT SNAPSHOT
2017-10-29 06:22:05 105-0: rpool/data/vm-105-disk-1@__replicate_105-0_1509254340__ name rpool/data/vm-105-disk-1@__replicate_105-0_1509254340__ -
2017-10-29 06:22:06 105-0: incremental sync 'local-zfs:vm-105-disk-2' (__replicate_105-0_1509254340__ => __replicate_105-0_1509254520__)
2017-10-29 06:22:06 105-0: send from @__replicate_105-0_1509254340__ to rpool/data/vm-105-disk-2@__replicate_105-0_1509254520__ estimated size is 174K
2017-10-29 06:22:06 105-0: total estimated size is 174K
2017-10-29 06:22:06 105-0: rpool/data/vm-105-disk-2@__replicate_105-0_1509254340__ name rpool/data/vm-105-disk-2@__replicate_105-0_1509254340__ -
2017-10-29 06:22:06 105-0: TIME SENT SNAPSHOT
2017-10-29 06:22:07 105-0: delete previous replication snapshot '__replicate_105-0_1509254340__' on local-zfs:vm-105-disk-1
2017-10-29 06:22:07 105-0: delete previous replication snapshot '__replicate_105-0_1509254340__' on local-zfs:vm-105-disk-2
2017-10-29 06:22:08 105-0: (remote_finalize_local_job) delete stale replication snapshot '__replicate_105-0_1509254340__' on local-zfs:vm-105-disk-1
2017-10-29 06:22:08 105-0: (remote_finalize_local_job) delete stale replication snapshot '__replicate_105-0_1509254340__' on local-zfs:vm-105-disk-2
2017-10-29 06:22:08 105-0: end replication job
tail: /var/log/pve/replicate/105-0: file truncated
2017-10-29 06:23:00 105-0: start replication job
2017-10-29 06:23:00 105-0: guest => VM 105, running => 27107
2017-10-29 06:23:00 105-0: volumes => local-zfs:vm-105-disk-1,local-zfs:vm-105-disk-2
2017-10-29 06:23:01 105-0: freeze guest filesystem
2017-10-29 06:23:02 105-0: create snapshot '__replicate_105-0_1509254580__' on local-zfs:vm-105-disk-1
2017-10-29 06:23:02 105-0: create snapshot '__replicate_105-0_1509254580__' on local-zfs:vm-105-disk-2
2017-10-29 06:23:02 105-0: thaw guest filesystem
2017-10-29 06:23:04 105-0: incremental sync 'local-zfs:vm-105-disk-1' (__replicate_105-0_1509254520__ => __replicate_105-0_1509254580__)
2017-10-29 06:23:04 105-0: send from @__replicate_105-0_1509254520__ to rpool/data/vm-105-disk-1@__replicate_105-0_1509254580__ estimated size is 2.48M
2017-10-29 06:23:04 105-0: total estimated size is 2.48M
2017-10-29 06:23:04 105-0: rpool/data/vm-105-disk-1@__replicate_105-0_1509254520__ name rpool/data/vm-105-disk-1@__replicate_105-0_1509254520__ -
2017-10-29 06:23:05 105-0: TIME SENT SNAPSHOT
2017-10-29 06:23:05 105-0: incremental sync 'local-zfs:vm-105-disk-2' (__replicate_105-0_1509254520__ => __replicate_105-0_1509254580__)
2017-10-29 06:23:06 105-0: send from @__replicate_105-0_1509254520__ to rpool/data/vm-105-disk-2@__replicate_105-0_1509254580__ estimated size is 133K
2017-10-29 06:23:06 105-0: total estimated size is 133K
2017-10-29 06:23:06 105-0: TIME SENT SNAPSHOT
2017-10-29 06:23:06 105-0: rpool/data/vm-105-disk-2@__replicate_105-0_1509254520__ name rpool/data/vm-105-disk-2@__replicate_105-0_1509254520__ -
2017-10-29 06:23:06 105-0: delete previous replication snapshot '__replicate_105-0_1509254520__' on local-zfs:vm-105-disk-1
2017-10-29 06:23:07 105-0: delete previous replication snapshot '__replicate_105-0_1509254520__' on local-zfs:vm-105-disk-2
2017-10-29 06:23:07 105-0: (remote_finalize_local_job) delete stale replication snapshot '__replicate_105-0_1509254520__' on local-zfs:vm-105-disk-1
2017-10-29 06:23:07 105-0: (remote_finalize_local_job) delete stale replication snapshot '__replicate_105-0_1509254520__' on local-zfs:vm-105-disk-2
2017-10-29 06:23:08 105-0: end replication job
tail: /var/log/pve/replicate/105-0: file truncated
2017-10-29 06:24:00 105-0: start replication job
2017-10-29 06:24:00 105-0: guest => VM 105, running => 27107
2017-10-29 06:24:00 105-0: volumes => local-zfs:vm-105-disk-1,local-zfs:vm-105-disk-2
2017-10-29 06:24:01 105-0: freeze guest filesystem
2017-10-29 06:24:02 105-0: create snapshot '__replicate_105-0_1509254640__' on local-zfs:vm-105-disk-1
2017-10-29 06:24:02 105-0: create snapshot '__replicate_105-0_1509254640__' on local-zfs:vm-105-disk-2
2017-10-29 06:24:03 105-0: thaw guest filesystem
2017-10-29 06:24:04 105-0: incremental sync 'local-zfs:vm-105-disk-1' (__replicate_105-0_1509254580__ => __replicate_105-0_1509254640__)
2017-10-29 06:24:04 105-0: send from @__replicate_105-0_1509254580__ to rpool/data/vm-105-disk-1@__replicate_105-0_1509254640__ estimated size is 2.48M
2017-10-29 06:24:04 105-0: total estimated size is 2.48M
2017-10-29 06:24:05 105-0: rpool/data/vm-105-disk-1@__replicate_105-0_1509254580__ name rpool/data/vm-105-disk-1@__replicate_105-0_1509254580__ -
2017-10-29 06:24:05 105-0: TIME SENT SNAPSHOT
2017-10-29 06:24:06 105-0: incremental sync 'local-zfs:vm-105-disk-2' (__replicate_105-0_1509254580__ => __replicate_105-0_1509254640__)
2017-10-29 06:24:06 105-0: send from @__replicate_105-0_1509254580__ to rpool/data/vm-105-disk-2@__replicate_105-0_1509254640__ estimated size is 133K
2017-10-29 06:24:06 105-0: total estimated size is 133K
2017-10-29 06:24:06 105-0: rpool/data/vm-105-disk-2@__replicate_105-0_1509254580__ name rpool/data/vm-105-disk-2@__replicate_105-0_1509254580__ -
2017-10-29 06:24:06 105-0: TIME SENT SNAPSHOT
2017-10-29 06:24:07 105-0: delete previous replication snapshot '__replicate_105-0_1509254580__' on local-zfs:vm-105-disk-1
2017-10-29 06:24:07 105-0: delete previous replication snapshot '__replicate_105-0_1509254580__' on local-zfs:vm-105-disk-2
2017-10-29 06:24:08 105-0: (remote_finalize_local_job) delete stale replication snapshot '__replicate_105-0_1509254580__' on local-zfs:vm-105-disk-1
2017-10-29 06:24:08 105-0: (remote_finalize_local_job) delete stale replication snapshot '__replicate_105-0_1509254580__' on local-zfs:vm-105-disk-2
2017-10-29 06:24:08 105-0: end replication job
Is there something wrong in configuration or is pvesr bugging?