Hi everyone,
my current setup consists of 3 nodes in a cluster, 2 with 160GB of ram and one with 192GB, actually HA is not enabled cause I'm waiting the disks to arrive to set up CEPH.
I'm starting to migrate some VMs to free RAM on the first node but the migration hangs..
it's stuck here.. I don't see anything in the log..
I also tried to create a VM on the second node and migrate to third (which are the same exact version) but still the same..
Versions:
Cluster status
my current setup consists of 3 nodes in a cluster, 2 with 160GB of ram and one with 192GB, actually HA is not enabled cause I'm waiting the disks to arrive to set up CEPH.
I'm starting to migrate some VMs to free RAM on the first node but the migration hangs..
Code:
2021-03-01 09:43:54 starting migration of VM 109 to node 'pve2' (192.168.1.109)
2021-03-01 09:43:55 found local, replicated disk 'local-zfs:vm-109-disk-0' (in current VM config)
2021-03-01 09:43:55 found local, replicated disk 'local-zfs:vm-109-disk-1' (in current VM config)
2021-03-01 09:43:55 replicating disk images
2021-03-01 09:43:55 start replication job
2021-03-01 09:43:55 guest => VM 109, running => 0
2021-03-01 09:43:55 volumes => local-zfs:vm-109-disk-0,local-zfs:vm-109-disk-1
2021-03-01 09:43:56 create snapshot '__replicate_109-0_1614588235__' on local-zfs:vm-109-disk-0
2021-03-01 09:43:57 create snapshot '__replicate_109-0_1614588235__' on local-zfs:vm-109-disk-1
2021-03-01 09:43:59 using secure transmission, rate limit: none
2021-03-01 09:43:59 incremental sync 'local-zfs:vm-109-disk-0' (__replicate_109-0_1614359062__ => __replicate_109-0_1614588235__)
2021-03-01 09:44:01 send from @__replicate_109-0_1614359062__ to rpool/data/vm-109-disk-0@__replicate_109-0_1614588235__ estimated size is 624B
2021-03-01 09:44:01 total estimated size is 624B
2021-03-01 09:44:02 TIME SENT SNAPSHOT rpool/data/vm-109-disk-0@__replicate_109-0_1614588235__
2021-03-01 09:44:03 successfully imported 'local-zfs:vm-109-disk-0'
2021-03-01 09:44:03 incremental sync 'local-zfs:vm-109-disk-1' (__replicate_109-0_1614359062__ => __replicate_109-0_1614588235__)
2021-03-01 09:44:05 send from @__replicate_109-0_1614359062__ to rpool/data/vm-109-disk-1@__replicate_109-0_1614588235__ estimated size is 624B
2021-03-01 09:44:05 total estimated size is 624B
2021-03-01 09:44:06 TIME SENT SNAPSHOT rpool/data/vm-109-disk-1@__replicate_109-0_1614588235__
2021-03-01 09:44:08 successfully imported 'local-zfs:vm-109-disk-1'
2021-03-01 09:44:08 delete previous replication snapshot '__replicate_109-0_1614359062__' on local-zfs:vm-109-disk-0
2021-03-01 09:44:09 delete previous replication snapshot '__replicate_109-0_1614359062__' on local-zfs:vm-109-disk-1
2021-03-01 09:44:12 (remote_finalize_local_job) delete stale replication snapshot '__replicate_109-0_1614359062__' on local-zfs:vm-109-disk-0
2021-03-01 09:44:12 (remote_finalize_local_job) delete stale replication snapshot '__replicate_109-0_1614359062__' on local-zfs:vm-109-disk-1
2021-03-01 09:44:12 end replication job
I also tried to create a VM on the second node and migrate to third (which are the same exact version) but still the same..
Versions:
Code:
root@pve:~# pveversion
pve-manager/6.3-2/22f57405 (running kernel: 5.4.73-1-pve)
root@pve2:~# pveversion
pve-manager/6.3-4/0a38c56f (running kernel: 5.4.98-1-pve)
root@pve3:~# pveversion
pve-manager/6.3-4/0a38c56f (running kernel: 5.4.98-1-pve)
Cluster status
Code:
root@pve:~# pvecm status
Cluster information
-------------------
Name: Cluster
Config Version: 7
Transport: knet
Secure auth: on
Quorum information
------------------
Date: Tue Mar 2 09:26:07 2021
Quorum provider: corosync_votequorum
Nodes: 3
Node ID: 0x00000001
Ring ID: 1.20a
Quorate: Yes
Votequorum information
----------------------
Expected votes: 3
Highest expected: 3
Total votes: 3
Quorum: 2
Flags: Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000001 1 192.168.1.32 (local)
0x00000002 1 192.168.1.109
0x00000003 1 192.168.1.124