Hi Proxmox team and Proxmox users,
As suggested by @fabian here is a new thread for this issue.
Congratulations to the team for the Proxmox 7 release.
I've upgraded a 4-node cluster with nvme ssd drive (nvme0n1) for filesystem and Ceph (sda, sdb) from the latest 6.x to 7.x.
Nodes are identical : same hardware, same Proxmox release, same configuration
The first 3-node upgrade were ok but for the last one (the primary node) I obtained several errors during the grub upgrade during the 6 to 7 dist-upgrade.
Hopefully the cluster works perfectly but on the primary node I always have an issue with the grub.
It seems that is a software related issue and not an hardware issue.
update-grub2 from a non-working node
nvme0n1
XXXX...XXXX is just to reduce the string message
uname -a from a non-working node
Message from @fabian
Information from a working node
Information from a non-working node
Do you know how to fix this grub issue ?
Any advice welcome,
Kind regards,
YAGA
As suggested by @fabian here is a new thread for this issue.
Congratulations to the team for the Proxmox 7 release.
I've upgraded a 4-node cluster with nvme ssd drive (nvme0n1) for filesystem and Ceph (sda, sdb) from the latest 6.x to 7.x.
Nodes are identical : same hardware, same Proxmox release, same configuration
The first 3-node upgrade were ok but for the last one (the primary node) I obtained several errors during the grub upgrade during the 6 to 7 dist-upgrade.
Hopefully the cluster works perfectly but on the primary node I always have an issue with the grub.
It seems that is a software related issue and not an hardware issue.
update-grub2 from a non-working node
nvme0n1
Code:
# update-grub2
Generating grub configuration file ...
Found linux image: /boot/vmlinuz-5.11.22-1-pve
Found initrd image: /boot/initrd.img-5.11.22-1-pve
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
Found linux image: /boot/vmlinuz-5.4.124-1-pve
Found initrd image: /boot/initrd.img-5.4.124-1-pve
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
Found linux image: /boot/vmlinuz-5.3.18-3-pve
Found initrd image: /boot/initrd.img-5.3.18-3-pve
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
Found linux image: /boot/vmlinuz-5.3.10-1-pve
Found initrd image: /boot/initrd.img-5.3.10-1-pve
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
/usr/sbin/grub-probe: error: disk `lvmid/rT5mdC-XXXX...XXXX-wqrmZe/YhO3eq-XXXX...XXXX-ZpVbQa' not found.
Found memtest86+ image: /boot/memtest86+.bin
Found memtest86+ multiboot image: /boot/memtest86+_multiboot.bin
Adding boot menu entry for EFI firmware configuration
done
XXXX...XXXX is just to reduce the string message
uname -a from a non-working node
Code:
# uname -a
Linux Cluster-STB-1 5.11.22-1-pve #1 SMP PVE 5.11.22-2 (Fri, 02 Jul 2021 16:22:45 +0200) x86_64 GNU/Linux
Message from @fabian
- details about your node's storage setup (especially regarding the / partition/filesystem - which filesystem, any hardware or software raid, ...)
Filesystem in on nvme ssd without raid (lvm).
- contents of /etc/pve/storage.cfg
- output of 'pvs', 'vgs', 'lvs' and 'lsblk' from a working and non-working node
Information from a working node
Code:
# cat /etc/pve/storage.cfg
dir: local
disable
path /var/lib/vz
content backup,vztmpl,iso
maxfiles 3
shared 0
lvmthin: local-lvm
thinpool data
vgname pve
content images,rootdir
rbd: Ceph-STB
content images,rootdir
krbd 0
pool Ceph-STB
cephfs: Cephfs-STB
path /mnt/pve/Cephfs-STB
content backup,iso,vztmpl
prune-backups keep-last=7
# pvs
PV VG Fmt Attr PSize PFree
/dev/nvme0n1p3 pve lvm2 a-- <1.82t 16.37g
/dev/sda ceph-4f8eba3b-2842-41be-99c5-cc0f4c08e0c1 lvm2 a-- <3.64t 0
/dev/sdb ceph-27eea1dd-ce95-48a8-9df6-69e853171b0a lvm2 a-- <3.64t 0
# vgs
VG #PV #LV #SN Attr VSize VFree
ceph-27eea1dd-ce95-48a8-9df6-69e853171b0a 1 1 0 wz--n- <3.64t 0
ceph-4f8eba3b-2842-41be-99c5-cc0f4c08e0c1 1 1 0 wz--n- <3.64t 0
pve 1 10 0 wz--n- <1.82t 16.37g
# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 3.6T 0 disk
└─ceph--4f8eba3b--2842--41be--99c5--cc0f4c08e0c1-osd--block--6c773935--1191--4248--a5b2--a3ec683b81a5 253:14 0 3.6T 0 lvm
sdb 8:16 0 3.6T 0 disk
└─ceph--27eea1dd--ce95--48a8--9df6--69e853171b0a-osd--block--afc5544c--d799--447e--815f--abdda5e9b171 253:13 0 3.6T 0 lvm
sr0 11:0 1 1024M 0 rom
nvme0n1 259:0 0 1.8T 0 disk
├─nvme0n1p1 259:1 0 1007K 0 part
├─nvme0n1p2 259:2 0 512M 0 part /boot/efi
└─nvme0n1p3 259:3 0 1.8T 0 part
├─pve-swap 253:0 0 8G 0 lvm [SWAP]
├─pve-root 253:1 0 96G 0 lvm /
├─pve-data_tmeta 253:2 0 15.8G 0 lvm
│ └─pve-data-tpool 253:4 0 1.7T 0 lvm
│ ├─pve-data 253:5 0 1.7T 1 lvm
│ ├─pve-vm--220--disk--0 253:6 0 8G 0 lvm
│ ├─pve-vm--220--disk--1 253:7 0 8G 0 lvm
│ ├─pve-vm--223--disk--0 253:8 0 4G 0 lvm
│ ├─pve-vm--51111--disk--0 253:9 0 20G 0 lvm
│ ├─pve-vm--51111--cloudinit 253:10 0 4M 0 lvm
│ ├─pve-vm--49032--disk--0 253:11 0 20G 0 lvm
│ └─pve-vm--49032--cloudinit 253:12 0 4M 0 lvm
└─pve-data_tdata 253:3 0 1.7T 0 lvm
└─pve-data-tpool 253:4 0 1.7T 0 lvm
├─pve-data 253:5 0 1.7T 1 lvm
├─pve-vm--220--disk--0 253:6 0 8G 0 lvm
├─pve-vm--220--disk--1 253:7 0 8G 0 lvm
├─pve-vm--223--disk--0 253:8 0 4G 0 lvm
├─pve-vm--51111--disk--0 253:9 0 20G 0 lvm
├─pve-vm--51111--cloudinit 253:10 0 4M 0 lvm
├─pve-vm--49032--disk--0 253:11 0 20G 0 lvm
└─pve-vm--49032--cloudinit 253:12 0 4M 0 lvm
Information from a non-working node
Code:
# cat /etc/pve/storage.cfg
dir: local
disable
path /var/lib/vz
content backup,vztmpl,iso
maxfiles 3
shared 0
lvmthin: local-lvm
thinpool data
vgname pve
content images,rootdir
rbd: Ceph-STB
content images,rootdir
krbd 0
pool Ceph-STB
cephfs: Cephfs-STB
path /mnt/pve/Cephfs-STB
content backup,iso,vztmpl
prune-backups keep-last=7
# pvs
PV VG Fmt Attr PSize PFree
/dev/nvme0n1p3 pve lvm2 a-- <1.82t 16.37g
/dev/sda ceph-407c2264-d843-4a6d-bbe1-578d60f8ca8b lvm2 a-- <3.64t 0
/dev/sdb ceph-015f0e4a-1978-4b85-a13f-f37c4d3ffabb lvm2 a-- <3.64t 0
# vgs
VG #PV #LV #SN Attr VSize VFree
ceph-015f0e4a-1978-4b85-a13f-f37c4d3ffabb 1 1 0 wz--n- <3.64t 0
ceph-407c2264-d843-4a6d-bbe1-578d60f8ca8b 1 1 0 wz--n- <3.64t 0
pve 1 29 0 wz--n- <1.82t 16.37g
# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 3.6T 0 disk
└─ceph--407c2264--d843--4a6d--bbe1--578d60f8ca8b-osd--block--b702c7b8--3b1e--40a0--972f--d4583e3eb243 253:33 0 3.6T 0 lvm
sdb 8:16 0 3.6T 0 disk
└─ceph--015f0e4a--1978--4b85--a13f--f37c4d3ffabb-osd--block--f52bce18--3afb--4f11--b380--80ddcdd0b3ef 253:32 0 3.6T 0 lvm
nvme0n1 259:0 0 1.8T 0 disk
├─nvme0n1p1 259:1 0 1007K 0 part
├─nvme0n1p2 259:2 0 512M 0 part /boot/efi
└─nvme0n1p3 259:3 0 1.8T 0 part
├─pve-swap 253:0 0 8G 0 lvm [SWAP]
├─pve-root 253:1 0 96G 0 lvm /
├─pve-data_tmeta 253:2 0 15.8G 0 lvm
│ └─pve-data-tpool 253:4 0 1.7T 0 lvm
│ ├─pve-data 253:5 0 1.7T 1 lvm
│ ├─pve-vm--51102--disk--0 253:6 0 20G 0 lvm
│ ├─pve-vm--51102--cloudinit 253:7 0 4M 0 lvm
│ ├─pve-vm--48003--disk--0 253:8 0 40G 0 lvm
│ ├─pve-vm--48004--disk--0 253:9 0 20G 0 lvm
│ ├─pve-vm--48002--disk--0 253:10 0 10G 0 lvm
│ ├─pve-vm--49002--disk--0 253:11 0 10G 0 lvm
│ ├─pve-vm--50002--disk--0 253:12 0 10G 0 lvm
│ ├─pve-vm--51002--disk--0 253:13 0 10G 0 lvm
│ ├─pve-vm--52002--disk--0 253:14 0 10G 0 lvm
│ ├─pve-vm--53002--disk--0 253:15 0 10G 0 lvm
│ ├─pve-vm--54002--disk--0 253:16 0 10G 0 lvm
│ ├─pve-vm--55002--disk--0 253:17 0 10G 0 lvm
│ ├─pve-vm--48005--disk--0 253:18 0 4G 0 lvm
│ ├─pve-vm--48005--cloudinit 253:19 0 4M 0 lvm
│ ├─pve-vm--53010--disk--0 253:20 0 10G 0 lvm
......
......
└─pve-data_tdata 253:3 0 1.7T 0 lvm
└─pve-data-tpool 253:4 0 1.7T 0 lvm
├─pve-data 253:5 0 1.7T 1 lvm
├─pve-vm--51102--disk--0 253:6 0 20G 0 lvm
├─pve-vm--51102--cloudinit 253:7 0 4M 0 lvm
├─pve-vm--48003--disk--0 253:8 0 40G 0 lvm
├─pve-vm--48004--disk--0 253:9 0 20G 0 lvm
......
......
Do you know how to fix this grub issue ?
Any advice welcome,
Kind regards,
YAGA