[SOLVED] Conflict node local-zfs

Ivan Gersi

Renowned Member
May 29, 2016
83
7
73
55
I have pve1 node with 4TB disk with fresh instalation local + local-zfs. I`ve made another node with 1TB disk with local and local-zfs storage (classic 100GB for local, rest for local-zfs).
When I made a cluster and join node2 to node1, node 2 has a problem with own local-zfs...I can see local-zfs(pve2) with question mark.
How is the best fix for thiss issue? E.g. I can resize local from 100GB to 1TB in pve2 or there is better to add new ZFS storage in cluster?
I still can see LVM-Thin in pve2 data.

root@pve1:~# df -h
Filesystem Size Used Avail Use% Mounted on
udev 3.9G 0 3.9G 0% /dev
tmpfs 794M 8.3M 786M 2% /run
rpool/ROOT/pve-1 3.5T 38G 3.5T 2% /
tmpfs 3.9G 66M 3.9G 2% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
rpool 3.5T 128K 3.5T 1% /rpool
rpool/ROOT 3.5T 128K 3.5T 1% /rpool/ROOT
rpool/data 3.5T 128K 3.5T 1% /rpool/data
//172.16.0.15/Public 3.6T 764G 2.9T 21% /mnt/pve/NAS
/dev/fuse 128M 24K 128M 1% /etc/pve
tmpfs 794M 0 794M 0% /run/user/0
pve2 has lvm structure.
root@pve2:~# vgs
VG #PV #LV #SN Attr VSize VFree
pve 1 3 0 wz--n- <930.51g <16.00g
root@pve2:~# pvs
PV VG Fmt Attr PSize PFree
/dev/sda3 pve lvm2 a-- <930.51g <16.00g
root@pve2:~# lvs
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
data pve twi-a-tz-- 794.66g 0.00 0.24
root pve -wi-ao---- 96.00g
swap pve -wi-ao---- 7.63g

What`s the best way guys?


Edit: I`ve migrate local-zfs to local and mounted to /var/lib/vz..., everything was right, but after reboot I can`t login to pve2. It has IP from dhcp-server, icmp working, but sshd not answer. It`s remote machine so I`ll have to wait from info from monitor.
 
Last edited:
There is strange issue. If I`d try to mount data via fstab Proxmox won`t boot up .
I`ve tried to set up in fstab /dev/pve/data /var/lib/vz ext4 defaults 0 2, hen without ext4 and finally only mount /dev/pve/dat /var/lib/vz.
Proxmox cras during boot because it can bre able mount this volume data.
But...when I remove this volume from fstab OS boot properly and If I want mount this volume via shell there is no problem to do it,

root@pve2:/etc# mount /dev/pve/data /var/lib/vz
root@pve2:/etc# df -h
Filesystem Size Used Avail Use% Mounted on
udev 3.8G 0 3.8G 0% /dev
tmpfs 782M 1.2M 781M 1% /run
/dev/mapper/pve-root 94G 2.7G 87G 4% /
tmpfs 3.9G 66M 3.8G 2% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
/dev/fuse 128M 24K 128M 1% /etc/pve
tmpfs 782M 0 782M 0% /run/user/0
/dev/mapper/pve-data 782G 44K 742G 1% /var/lib/vz

i`m little confused with this issue.
Lvs is classic.
root@pve2:/var/log# lvs
File descriptor 9 (pipe:[25178]) leaked on lvs invocation. Parent PID 1448: bash
File descriptor 11 (pipe:[25179]) leaked on lvs invocation. Parent PID 1448: bash
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
data pve twi-aotz-- 794.66g 0.00 0.24
root pve -wi-ao---- 96.00g
swap pve -wi-ao---- 7.63g
root@pve2:/var/log# lvdisplay
File descriptor 9 (pipe:[25178]) leaked on lvdisplay invocation. Parent PID 1448: bash
File descriptor 11 (pipe:[25179]) leaked on lvdisplay invocation. Parent PID 1448: bash
--- Logical volume ---
LV Path /dev/pve/swap
LV Name swap
VG Name pve
LV UUID rnt2fb-wXgm-lXem-aaGE-EBeL-CYEQ-p5sFEG
LV Write Access read/write
LV Creation host, time proxmox, 2023-05-02 17:04:33 +0200
LV Status available
# open 2
LV Size 7.63 GiB
Current LE 1954
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 256
Block device 253:0

--- Logical volume ---
LV Path /dev/pve/root
LV Name root
VG Name pve
LV UUID zo0W7u-9rA6-Y7vq-dbX0-GSf7-3iwD-2pBY5f
LV Write Access read/write
LV Creation host, time proxmox, 2023-05-02 17:04:33 +0200
LV Status available
# open 1
LV Size 96.00 GiB
Current LE 24576
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 256
Block device 253:1

--- Logical volume ---
LV Name data
VG Name pve
LV UUID QYlvj0-EoRn-IOkN-myG7-RnA9-jX6j-i2dPNq
LV Write Access read/write
LV Creation host, time proxmox, 2023-05-02 17:04:54 +0200
LV Pool metadata data_tmeta
LV Pool data data_tdata
LV Status available
# open 1
LV Size 794.66 GiB
Allocated pool data 0.00%
Allocated metadata 0.24%
Current LE 203433
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 256
Block device 253:4
 
AFAIU you have 2 nodes, one installed with ZFS as root FS and the other with XFS or EXT4 which means, they use LVM on the root disk.

When a node joins a cluster, quite a few config files are overwritten with the ones from the cluster. The storage.cfg is one of them. Therefore, the configuration for the "local-lvm" config for node 2 was lost. Since "local-zfs" does not have a node limitation, Proxmox VE tries to activate that storage on node 2 as well.

Ideally the nodes in the cluster are set up the same way. In your situation, you could limit "local-zfs" just to node 1. And add another storage of type LVM-thin and call it "local-lvm". It should point to the "pve/data" LV. It should be visible if you open the GUI on node 2 for that action. Don't forget to limit "local-lvm" to node 2, otherwise you will see the question mark on node 1 ;)
 
  • Like
Reactions: Ivan Gersi
AarIon I think you``re probably right but tis is my theory...I`ve made node 1 a few weeks ago so I can`t remember configuration exactly... I thought there was the same, but nod1 has 4TB zfs-local and node2 only 1TB. Cluster wanted only one zfs-local but node2 hasn`t enough capacity.and this was the reason why node2 has zfs-local with question mark and node1 has 2 volumes with the same capacity (local and zfs-local) now.
What`s the best scenario for me now? Remove zfs-local from node1 and let only local and make the same way in node2?
There is a little problem...I have to be careful because node1 is in the customer rack a there is one VM (running on local-zfs) and node2 is at my home now;o).
Som I`m going to try to fix it.

Edit: Proxmox has a special policy...in older version user has after instalation local and local-lvm (or local-thin). In newer version has user after fresh instalation only 100GB local and has make to another volume after instalation. In newest version ( think v7) user has classic local a localz-zfs volume in default...and this can be cinfusing.
I have several v7 clusters but all cluster was upgraded from v6 a they was upgraded from v5....etc. I never made cluster from v7 instantly.
 
Last edited:
aaron thanks for hint...this can be problem...
root@pve1:~# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 3.6T 0 disk
├─sda1 8:1 0 1007K 0 part
├─sda2 8:2 0 512M 0 part
└─sda3 8:3 0 3.6T 0 part
sdb 8:16 0 3.6T 0 disk
├─sdb1 8:17 0 1007K 0 part
├─sdb2 8:18 0 512M 0 part
└─sdb3 8:19 0 3.6T 0 part
sr0 11:0 1 1024M 0 rom
zd0 230:0 0 500G 0 disk
├─zd0p1 230:1 0 549M 0 part
└─zd0p2 230:2 0 499.5G 0 part
and pve2
root@pve2:/etc/pve# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 931.5G 0 disk
├─sda1 8:1 0 1007K 0 part
├─sda2 8:2 0 1G 0 part
└─sda3 8:3 0 930.5G 0 part
├─pve-swap 253:0 0 7.6G 0 lvm [SWAP]
├─pve-root 253:1 0 96G 0 lvm /
├─pve-data_tmeta 253:2 0 8.1G 0 lvm
│ └─pve-data 253:4 0 794.7G 0 lvm
└─pve-data_tdata 253:3 0 794.7G 0 lvm
└─pve-data 253:4 0 794.7G 0 lvm