ZFS has '?' mark in GUI with EXT4 partition and HA migration errors

vivekdelhi

New Member
Sep 13, 2019
13
0
1
47
Dear Forum Members
Today I installed Proxmox on two blades. Each blade has two drives, No RAID. The Installation was done on /dev/sda, with EXT4 as the file-system.
I wanted to stay away from ZFS due to memory concerns since these are low RAM servers (only 32GB) and because the second drive was later to be added to a CEPH cluster.
In the GUI, i am seeing a question mark against local-zfs (screenshot attached)
The same error has repeated in both nodes
Another error I faced was in doing a HA migration of a VM from another node to this new node. The VMs data is in CEPH pool. But the HA migration failed with the following error - pve-ha-lrm[364044]: zfs error: cannot open 'rpool': no such pool

I suspect this is due to proxmox expecting a ZFS pool when there is none.

Can anyone suggest how to fix these
a. Incorrect ZFS pool being shown
b. struch HA migrate task

Thanks
Vivek

Code:
task started by HA resource agent
2020-04-18 22:16:47 starting migration of VM 106 to node 'dell0104blade01' (192.168.15.31)
zfs error: cannot open 'rpool': no such pool

zfs error: cannot open 'rpool': no such pool

2020-04-18 22:16:48 ERROR: Failed to sync data - could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
2020-04-18 22:16:48 aborting phase 1 - cleanup resources
2020-04-18 22:16:48 ERROR: migration aborted (duration 00:00:01): Failed to sync data - could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
TASK ERROR: migration aborted
Code:
18 22:25:05 dell0104blade04 pmxcfs[1165]: [dcdb] notice: all data is up to date
Apr 18 22:25:05 dell0104blade04 pmxcfs[1165]: [status] notice: received all states
Apr 18 22:25:05 dell0104blade04 pmxcfs[1165]: [status] notice: all data is up to date
Apr 18 22:25:07 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:07 dell0104blade04 pve-ha-lrm[363880]: <root@pam> starting task UPID:dell0104blade04:00058D7F:002D5B58:5E9B30EB:qmigrate:106:root@pam:
Apr 18 22:25:07 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:07 dell0104blade04 pvestatd[1316]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363903]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363903]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363903]: migration aborted
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363880]: <root@pam> end task UPID:dell0104blade04:00058D7F:002D5B58:5E9B30EB:qmigrate:106:root@pam: migration aborted
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363880]: service vm:106 not moved (migration error)
Apr 18 22:25:09 dell0104blade04 pvedaemon[201969]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:10 dell0104blade04 pvedaemon[201969]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:10 dell0104blade04 pvedaemon[201969]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:16 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:16 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:16 dell0104blade04 pvestatd[1316]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:17 dell0104blade04 pve-ha-lrm[364043]: <root@pam> starting task UPID:dell0104blade04:00058E0C:002D5F1A:5E9B30F5:qmigrate:106:root@pam:
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364044]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364044]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364044]: migration aborted
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364043]: <root@pam> end task UPID:dell0104blade04:00058E0C:002D5F1A:5E9B30F5:qmigrate:106:root@pam: migration aborted
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364043]: service vm:106 not moved (migration error)
Apr 18 22:25:18 dell0104blade04 pvedaemon[216417]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:19 dell0104blade04 pvedaemon[216417]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:19 dell0104blade04 pvedaemon[216417]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:26 dell0104blade04 pvedaemon[216809]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:26 dell0104blade04 pvedaemon[216809]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:26 dell0104blade04 pvedaemon[216809]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:26 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:27 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:27 dell0104blade04 pvestatd[1316]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:27 dell0104blade04 pve-ha-lrm[364217]: <root@pam> starting task UPID:dell0104blade04:00058EBA:002D6304:5E9B30FF:qmigrate:106:root@pam:
Apr 18 22:25:27 dell0104blade04 pve-ha-lrm[364218]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:27 dell0104blade04 pve-ha-lrm[364218]: zfs error: cannot open 'rpool': no such pool

Code:
root@dell0104blade04:~# pvs
  PV         VG                                        Fmt  Attr PSize   PFree 
  /dev/sda3  pve                                       lvm2 a--  <99.50g <12.38g
  /dev/sdb   ceph-0dbdffc1-9c32-4436-9aec-a0a53ad0c325 lvm2 a--  279.39g      0

root@dell0104blade04:~# vgs
  VG                                        #PV #LV #SN Attr   VSize   VFree 
  ceph-0dbdffc1-9c32-4436-9aec-a0a53ad0c325   1   1   0 wz--n- 279.39g      0
  pve                                         1   3   0 wz--n- <99.50g <12.38g

root@dell0104blade04:~# fdisk -l
Disk /dev/sda: 279.4 GiB, 300000000000 bytes, 585937500 sectors
Disk model: ST300MM0006     
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: gpt
Disk identifier: A0FF6AD0-5CAD-4204-80BB-72FEDB44D585

Device       Start       End   Sectors  Size Type
/dev/sda1       34      2047      2014 1007K BIOS boot
/dev/sda2     2048   1050623   1048576  512M EFI System
/dev/sda3  1050624 209715200 208664577 99.5G Linux LVM


Disk /dev/sdb: 279.4 GiB, 300000000000 bytes, 585937500 sectors
Disk model: ST300MM0006     
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


Disk /dev/mapper/pve-swap: 8 GiB, 8589934592 bytes, 16777216 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


Disk /dev/mapper/pve-root: 24.8 GiB, 26575110144 bytes, 51904512 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


Disk /dev/mapper/ceph--0dbdffc1--9c32--4436--9aec--a0a53ad0c325-osd--block--71363746--d31c--4bf6--9bf6--4375603b0d44: 279.4 GiB, 299997593600 bytes, 585932800 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes

root@dell0104blade04:~# zfs list
no datasets available
 

Attachments

  • Screenshot 2020-04-18 at 10.13.07 PM.png
    Screenshot 2020-04-18 at 10.13.07 PM.png
    214.6 KB · Views: 3
I suspect the storage.cfg needs to be amended. The current configuration is
Code:
root@dell0104blade04:~# cat /etc/pve/storage.cfg
dir: local
        path /var/lib/vz
        content backup,vztmpl,iso
        maxfiles 4
        shared 0

zfspool: local-zfs
        pool rpool/data
        content rootdir,images
        sparse 1

rbd: cephpool1
        content rootdir,images
        krbd 0
        pool cephpool1
 
Aah, Managed to Identify the issue and correct it. Its all there in the proxmox Wiki. ! Thanks If only I had looked harder earlier

the /etc/pve/stodage.cfg is shared across the cluster. Since two fo teh nodes do not have ZFS, I needed to explictly specify which all nodes had ZFS. That was easily achieved from the Cluster > Storage configuration in GUI. My cfg now reads

Code:
root@dell0104blade04:~# cat /etc/pve/storage.cfg
dir: local
        path /var/lib/vz
        content backup,vztmpl,iso
        maxfiles 4
        shared 0

zfspool: local-zfs
        pool rpool/data
        content images,rootdir
        nodes dell0104blade02,dell0104blade01,dell0104blade10,hp0105blade07duplicate
        sparse 1

rbd: cephpool1
        content rootdir,images
        krbd 0
        pool cephpool1
All errors sorted.
Thanks for the great product !