ZFS has '?' mark in GUI with EXT4 partition and HA migration errors

vivekdelhi

New Member
Sep 13, 2019
13
0
1
46
Dear Forum Members
Today I installed Proxmox on two blades. Each blade has two drives, No RAID. The Installation was done on /dev/sda, with EXT4 as the file-system.
I wanted to stay away from ZFS due to memory concerns since these are low RAM servers (only 32GB) and because the second drive was later to be added to a CEPH cluster.
In the GUI, i am seeing a question mark against local-zfs (screenshot attached)
The same error has repeated in both nodes
Another error I faced was in doing a HA migration of a VM from another node to this new node. The VMs data is in CEPH pool. But the HA migration failed with the following error - pve-ha-lrm[364044]: zfs error: cannot open 'rpool': no such pool

I suspect this is due to proxmox expecting a ZFS pool when there is none.

Can anyone suggest how to fix these
a. Incorrect ZFS pool being shown
b. struch HA migrate task

Thanks
Vivek

Code:
task started by HA resource agent
2020-04-18 22:16:47 starting migration of VM 106 to node 'dell0104blade01' (192.168.15.31)
zfs error: cannot open 'rpool': no such pool

zfs error: cannot open 'rpool': no such pool

2020-04-18 22:16:48 ERROR: Failed to sync data - could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
2020-04-18 22:16:48 aborting phase 1 - cleanup resources
2020-04-18 22:16:48 ERROR: migration aborted (duration 00:00:01): Failed to sync data - could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
TASK ERROR: migration aborted
Code:
18 22:25:05 dell0104blade04 pmxcfs[1165]: [dcdb] notice: all data is up to date
Apr 18 22:25:05 dell0104blade04 pmxcfs[1165]: [status] notice: received all states
Apr 18 22:25:05 dell0104blade04 pmxcfs[1165]: [status] notice: all data is up to date
Apr 18 22:25:07 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:07 dell0104blade04 pve-ha-lrm[363880]: <root@pam> starting task UPID:dell0104blade04:00058D7F:002D5B58:5E9B30EB:qmigrate:106:root@pam:
Apr 18 22:25:07 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:07 dell0104blade04 pvestatd[1316]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363903]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363903]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363903]: migration aborted
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363880]: <root@pam> end task UPID:dell0104blade04:00058D7F:002D5B58:5E9B30EB:qmigrate:106:root@pam: migration aborted
Apr 18 22:25:08 dell0104blade04 pve-ha-lrm[363880]: service vm:106 not moved (migration error)
Apr 18 22:25:09 dell0104blade04 pvedaemon[201969]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:10 dell0104blade04 pvedaemon[201969]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:10 dell0104blade04 pvedaemon[201969]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:16 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:16 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:16 dell0104blade04 pvestatd[1316]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:17 dell0104blade04 pve-ha-lrm[364043]: <root@pam> starting task UPID:dell0104blade04:00058E0C:002D5F1A:5E9B30F5:qmigrate:106:root@pam:
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364044]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364044]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364044]: migration aborted
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364043]: <root@pam> end task UPID:dell0104blade04:00058E0C:002D5F1A:5E9B30F5:qmigrate:106:root@pam: migration aborted
Apr 18 22:25:18 dell0104blade04 pve-ha-lrm[364043]: service vm:106 not moved (migration error)
Apr 18 22:25:18 dell0104blade04 pvedaemon[216417]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:19 dell0104blade04 pvedaemon[216417]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:19 dell0104blade04 pvedaemon[216417]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:26 dell0104blade04 pvedaemon[216809]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:26 dell0104blade04 pvedaemon[216809]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:26 dell0104blade04 pvedaemon[216809]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:26 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:27 dell0104blade04 pvestatd[1316]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:27 dell0104blade04 pvestatd[1316]: could not activate storage 'local-zfs', zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:27 dell0104blade04 pve-ha-lrm[364217]: <root@pam> starting task UPID:dell0104blade04:00058EBA:002D6304:5E9B30FF:qmigrate:106:root@pam:
Apr 18 22:25:27 dell0104blade04 pve-ha-lrm[364218]: zfs error: cannot open 'rpool': no such pool
Apr 18 22:25:27 dell0104blade04 pve-ha-lrm[364218]: zfs error: cannot open 'rpool': no such pool

Code:
root@dell0104blade04:~# pvs
  PV         VG                                        Fmt  Attr PSize   PFree 
  /dev/sda3  pve                                       lvm2 a--  <99.50g <12.38g
  /dev/sdb   ceph-0dbdffc1-9c32-4436-9aec-a0a53ad0c325 lvm2 a--  279.39g      0

root@dell0104blade04:~# vgs
  VG                                        #PV #LV #SN Attr   VSize   VFree 
  ceph-0dbdffc1-9c32-4436-9aec-a0a53ad0c325   1   1   0 wz--n- 279.39g      0
  pve                                         1   3   0 wz--n- <99.50g <12.38g

root@dell0104blade04:~# fdisk -l
Disk /dev/sda: 279.4 GiB, 300000000000 bytes, 585937500 sectors
Disk model: ST300MM0006     
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: gpt
Disk identifier: A0FF6AD0-5CAD-4204-80BB-72FEDB44D585

Device       Start       End   Sectors  Size Type
/dev/sda1       34      2047      2014 1007K BIOS boot
/dev/sda2     2048   1050623   1048576  512M EFI System
/dev/sda3  1050624 209715200 208664577 99.5G Linux LVM


Disk /dev/sdb: 279.4 GiB, 300000000000 bytes, 585937500 sectors
Disk model: ST300MM0006     
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


Disk /dev/mapper/pve-swap: 8 GiB, 8589934592 bytes, 16777216 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


Disk /dev/mapper/pve-root: 24.8 GiB, 26575110144 bytes, 51904512 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes


Disk /dev/mapper/ceph--0dbdffc1--9c32--4436--9aec--a0a53ad0c325-osd--block--71363746--d31c--4bf6--9bf6--4375603b0d44: 279.4 GiB, 299997593600 bytes, 585932800 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes

root@dell0104blade04:~# zfs list
no datasets available
 

Attachments

  • Screenshot 2020-04-18 at 10.13.07 PM.png
    Screenshot 2020-04-18 at 10.13.07 PM.png
    214.6 KB · Views: 3
I suspect the storage.cfg needs to be amended. The current configuration is
Code:
root@dell0104blade04:~# cat /etc/pve/storage.cfg
dir: local
        path /var/lib/vz
        content backup,vztmpl,iso
        maxfiles 4
        shared 0

zfspool: local-zfs
        pool rpool/data
        content rootdir,images
        sparse 1

rbd: cephpool1
        content rootdir,images
        krbd 0
        pool cephpool1
 
Aah, Managed to Identify the issue and correct it. Its all there in the proxmox Wiki. ! Thanks If only I had looked harder earlier

the /etc/pve/stodage.cfg is shared across the cluster. Since two fo teh nodes do not have ZFS, I needed to explictly specify which all nodes had ZFS. That was easily achieved from the Cluster > Storage configuration in GUI. My cfg now reads

Code:
root@dell0104blade04:~# cat /etc/pve/storage.cfg
dir: local
        path /var/lib/vz
        content backup,vztmpl,iso
        maxfiles 4
        shared 0

zfspool: local-zfs
        pool rpool/data
        content images,rootdir
        nodes dell0104blade02,dell0104blade01,dell0104blade10,hp0105blade07duplicate
        sparse 1

rbd: cephpool1
        content rootdir,images
        krbd 0
        pool cephpool1
All errors sorted.
Thanks for the great product !
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!