I was running 4.2 for quite some time. My shared storage is a Dell MD3620i with LVM over iSCSI.
Everything has been running fine until I upgraded to 4.4 using the community updates. The servers are now reporting the storage is not online with an error 500 when I click on the storage device in the dashboard.
I can see the storage devices in command line using vgs and lvs however I am unable start containers.
I have the MD3620i configured with multipath which to be reporting properly.
I am unable to start any containers or VM's and am worried this may be causing some corruption.
I am also concerned that this may be a bug in the stable repository. Or is there a technical change that was not documented or overlooked.
Any help to identify the problem would be greatly appreciated.
Thank you
Joe
Here is multipath results:
# multipath -ll
md36xxi0_dg1_vd1 (36d4ae5200097a2c800000e5a5683f025) dm-3 DELL,MD36xxi
size=1.1T features='3 queue_if_no_path pg_init_retries 50' hwhandler='1 rdac' wp=rw
`-+- policy='round-robin 0' prio=11 status=active
|- 6:0:0:0 sdb 8:16 active ready running
|- 7:0:0:0 sdc 8:32 active ready running
|- 8:0:0:0 sdd 8:48 active ready running
`- 9:0:0:0 sde 8:64 active ready running
md36xxi0_dg2_storage1 (36d4ae5200097a36a000010595683abc5) dm-4 DELL,MD36xxi
size=5.5T features='3 queue_if_no_path pg_init_retries 50' hwhandler='1 rdac' wp=rw
`-+- policy='round-robin 0' prio=11 status=active
|- 6:0:0:1 sdf 8:80 active ready running
|- 7:0:0:1 sdg 8:96 active ready running
|- 8:0:0:1 sdi 8:128 active ready running
`- 9:0:0:1 sdh 8:112 active ready running
Here is vgs results:
# vgs
VG #PV #LV #SN Attr VSize VFree
pve 1 3 0 wz--n- 465.63g 0
storage0 1 19 0 wz--n- 1.09t 335.58g
storage1 1 8 0 wz--n- 5.46t 4.86t
Here are lvs results
# lvs
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
data pve -wi-ao---- 361.63g
root pve -wi-ao---- 96.00g
swap pve -wi-ao---- 8.00g
vm-100-disk-1 storage0 -wi------- 50.00g
vm-103-disk-1 storage0 -wi------- 8.00g
vm-105-disk-2 storage0 -wi------- 10.00g
vm-106-disk-1 storage0 -wi------- 12.00g
vm-107-disk-1 storage0 -wi------- 10.00g
vm-108-disk-1 storage0 -wi------- 10.00g
vm-109-disk-1 storage0 -wi------- 50.00g
vm-110-disk-1 storage0 -wi------- 60.00g
vm-111-disk-1 storage0 -wi------- 50.00g
vm-113-disk-1 storage0 -wi------- 20.00g
vm-114-disk-1 storage0 -wi------- 60.00g
vm-115-disk-1 storage0 -wi------- 15.00g
vm-117-disk-1 storage0 -wi------- 15.00g
vm-121-disk-1 storage0 -wi------- 64.00g
vm-122-disk-1 storage0 -wi------- 8.00g
vm-200-disk-1 storage0 -wi------- 10.00g
vm-217-disk-1 storage0 -wi------- 64.00g
vm-317-disk-1 storage0 -wi------- 64.00g
vm-601-disk-1 storage0 -wi------- 200.00g
vm-101-disk-1 storage1 -wi------- 15.00g
vm-102-disk-1 storage1 -wi------- 15.00g
vm-116-disk-1 storage1 -wi------- 30.00g
vm-119-disk-1 storage1 -wi------- 260.00g
vm-123-disk-1 storage1 -wi------- 30.00g
vm-124-disk-1 storage1 -wi------- 50.00g
vm-125-disk-1 storage1 -wi------- 15.00g
vm-501-disk-1 storage1 -wi------- 200.00g
Here is a tail of syslog showing error
# tail /var/log/syslog
Jan 23 00:58:08 sv7n2 pvestatd[23783]: status update time (6.062 seconds)
Jan 23 00:58:14 sv7n2 pvestatd[23783]: storage 'md3620i_dg1_vg1' is not online
Jan 23 00:58:16 sv7n2 pvestatd[23783]: storage 'md3620i_dg1_vg1' is not online
Jan 23 00:58:18 sv7n2 pvestatd[23783]: storage 'md3620i_dg1_vg1' is not online
Jan 23 00:58:18 sv7n2 pvestatd[23783]: status update time (6.059 seconds)
Jan 23 00:58:24 sv7n2 pvestatd[23783]: storage 'md3620i_dg1_vg1' is not online
Everything has been running fine until I upgraded to 4.4 using the community updates. The servers are now reporting the storage is not online with an error 500 when I click on the storage device in the dashboard.
I can see the storage devices in command line using vgs and lvs however I am unable start containers.
I have the MD3620i configured with multipath which to be reporting properly.
I am unable to start any containers or VM's and am worried this may be causing some corruption.
I am also concerned that this may be a bug in the stable repository. Or is there a technical change that was not documented or overlooked.
Any help to identify the problem would be greatly appreciated.
Thank you
Joe
Here is multipath results:
# multipath -ll
md36xxi0_dg1_vd1 (36d4ae5200097a2c800000e5a5683f025) dm-3 DELL,MD36xxi
size=1.1T features='3 queue_if_no_path pg_init_retries 50' hwhandler='1 rdac' wp=rw
`-+- policy='round-robin 0' prio=11 status=active
|- 6:0:0:0 sdb 8:16 active ready running
|- 7:0:0:0 sdc 8:32 active ready running
|- 8:0:0:0 sdd 8:48 active ready running
`- 9:0:0:0 sde 8:64 active ready running
md36xxi0_dg2_storage1 (36d4ae5200097a36a000010595683abc5) dm-4 DELL,MD36xxi
size=5.5T features='3 queue_if_no_path pg_init_retries 50' hwhandler='1 rdac' wp=rw
`-+- policy='round-robin 0' prio=11 status=active
|- 6:0:0:1 sdf 8:80 active ready running
|- 7:0:0:1 sdg 8:96 active ready running
|- 8:0:0:1 sdi 8:128 active ready running
`- 9:0:0:1 sdh 8:112 active ready running
Here is vgs results:
# vgs
VG #PV #LV #SN Attr VSize VFree
pve 1 3 0 wz--n- 465.63g 0
storage0 1 19 0 wz--n- 1.09t 335.58g
storage1 1 8 0 wz--n- 5.46t 4.86t
Here are lvs results
# lvs
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
data pve -wi-ao---- 361.63g
root pve -wi-ao---- 96.00g
swap pve -wi-ao---- 8.00g
vm-100-disk-1 storage0 -wi------- 50.00g
vm-103-disk-1 storage0 -wi------- 8.00g
vm-105-disk-2 storage0 -wi------- 10.00g
vm-106-disk-1 storage0 -wi------- 12.00g
vm-107-disk-1 storage0 -wi------- 10.00g
vm-108-disk-1 storage0 -wi------- 10.00g
vm-109-disk-1 storage0 -wi------- 50.00g
vm-110-disk-1 storage0 -wi------- 60.00g
vm-111-disk-1 storage0 -wi------- 50.00g
vm-113-disk-1 storage0 -wi------- 20.00g
vm-114-disk-1 storage0 -wi------- 60.00g
vm-115-disk-1 storage0 -wi------- 15.00g
vm-117-disk-1 storage0 -wi------- 15.00g
vm-121-disk-1 storage0 -wi------- 64.00g
vm-122-disk-1 storage0 -wi------- 8.00g
vm-200-disk-1 storage0 -wi------- 10.00g
vm-217-disk-1 storage0 -wi------- 64.00g
vm-317-disk-1 storage0 -wi------- 64.00g
vm-601-disk-1 storage0 -wi------- 200.00g
vm-101-disk-1 storage1 -wi------- 15.00g
vm-102-disk-1 storage1 -wi------- 15.00g
vm-116-disk-1 storage1 -wi------- 30.00g
vm-119-disk-1 storage1 -wi------- 260.00g
vm-123-disk-1 storage1 -wi------- 30.00g
vm-124-disk-1 storage1 -wi------- 50.00g
vm-125-disk-1 storage1 -wi------- 15.00g
vm-501-disk-1 storage1 -wi------- 200.00g
Here is a tail of syslog showing error
# tail /var/log/syslog
Jan 23 00:58:08 sv7n2 pvestatd[23783]: status update time (6.062 seconds)
Jan 23 00:58:14 sv7n2 pvestatd[23783]: storage 'md3620i_dg1_vg1' is not online
Jan 23 00:58:16 sv7n2 pvestatd[23783]: storage 'md3620i_dg1_vg1' is not online
Jan 23 00:58:18 sv7n2 pvestatd[23783]: storage 'md3620i_dg1_vg1' is not online
Jan 23 00:58:18 sv7n2 pvestatd[23783]: status update time (6.059 seconds)
Jan 23 00:58:24 sv7n2 pvestatd[23783]: storage 'md3620i_dg1_vg1' is not online