[SOLVED] Update von 7 auf 8 Fehlgeschlagen

ibit

Member
Dec 21, 2020
17
3
8
39
Hallo Zusammen,

nach dem Update von PVE 7 auf PVE 8 habe ich einen Fehler im Kernel erhalten. Weiß jemand, was ich tun kann?

Bash:
apt install pve-kernel-6.2
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
pve-kernel-6.2 is already the newest version (8.0.2).
0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded.
3 not fully installed or removed.
After this operation, 0 B of additional disk space will be used.
Do you want to continue? [Y/n] Y
Setting up pve-kernel-6.2.16-3-pve (6.2.16-3) ...
Examining /etc/kernel/postinst.d.
run-parts: executing /etc/kernel/postinst.d/dkms 6.2.16-3-pve /boot/vmlinuz-6.2.16-3-pve
dkms: running auto installation service for kernel 6.2.16-3-pve.
Sign command: /lib/modules/6.2.16-3-pve/build/scripts/sign-file
Signing key: /var/lib/dkms/mok.key
Public certificate (MOK): /var/lib/dkms/mok.pub

Building module:
Cleaning build area...(bad exit status: 2)
make -j8 KERNELRELEASE=6.2.16-3-pve all KPVER=6.2.16-3-pve...(bad exit status: 2)
Error! Bad return status for module build on kernel: 6.2.16-3-pve (x86_64)
Consult /var/lib/dkms/kernel-mft-dkms/4.18.0/build/make.log for more information.
Error! One or more modules failed to install during autoinstall.
Refer to previous errors for more information.
dkms: autoinstall for kernel: 6.2.16-3-pve failed!
run-parts: /etc/kernel/postinst.d/dkms exited with return code 11
Failed to process /etc/kernel/postinst.d at /var/lib/dpkg/info/pve-kernel-6.2.16-3-pve.postinst line 20.
dpkg: error processing package pve-kernel-6.2.16-3-pve (--configure):
 installed pve-kernel-6.2.16-3-pve package post-installation script subprocess returned error exit status 2
dpkg: dependency problems prevent configuration of pve-kernel-6.2:
 pve-kernel-6.2 depends on pve-kernel-6.2.16-3-pve; however:
  Package pve-kernel-6.2.16-3-pve is not configured yet.

dpkg: error processing package pve-kernel-6.2 (--configure):
 dependency problems - leaving unconfigured
dpkg: dependency problems prevent configuration of proxmox-ve:
 proxmox-ve depends on pve-kernel-6.2; however:
  Package pve-kernel-6.2 is not configured yet.

dpkg: error processing package proxmox-ve (--configure):
 dependency problems - leaving unconfigured
Errors were encountered while processing:
 pve-kernel-6.2.16-3-pve
 pve-kernel-6.2
 proxmox-ve
E: Sub-process /usr/bin/dpkg returned an error code (1)

Momentan läuft:
Linux s00 6.2.6-1-pve #1 SMP PREEMPT_DYNAMIC PVE 6.2.6-1 (2023-03-14T17:08Z) x86_64 GNU/Linux

Vielen Dank!
 
Hi,
sieht so aus als wäre da ein Kernel module mft welches mittels dkms gebaut wird das fehlschlägt. Bitte poste mal den Inhalt von /var/lib/dkms/kernel-mft-dkms/4.18.0/build/make.log
 
Hier ist die Logdatei

Code:
DKMS make.log for kernel-mft-dkms-4.18.0 for kernel 6.2.16-3-pve (x86_64)
Thu Jun 22 03:33:25 PM CEST 2023
/bin/sh: 1: Syntax error: Unterminated quoted string
/bin/sh: 1: [: -lt: unexpected operator
make -C /lib/modules/6.2.16-3-pve/build M=/var/lib/dkms/kernel-mft-dkms/4.18.0/build CONFIG_CTF= CONFIG_CC_STACKPROTECTOR_STRONG=  modules
make[1]: warning: jobserver unavailable: using -j1.  Add '+' to parent make rule.
make[1]: Entering directory '/usr/src/linux-headers-6.2.16-3-pve'
/bin/sh: 1: Syntax error: Unterminated quoted string
/bin/sh: 1: [: -lt: unexpected operator
  CC [M]  /var/lib/dkms/kernel-mft-dkms/4.18.0/build/mst_pci.o
  CC [M]  /var/lib/dkms/kernel-mft-dkms/4.18.0/build/mst_pciconf.o
/var/lib/dkms/kernel-mft-dkms/4.18.0/build/mst_pciconf.c: In function ‘close_dma’:
/var/lib/dkms/kernel-mft-dkms/4.18.0/build/mst_pciconf.c:601:13: error: implicit declaration of function ‘pci_unmap_single’; did you mean ‘dma_unmap_single’? [-Werror=implicit-function-declaration]
  601 |             pci_unmap_single(dev->pci_dev, dev->dma_props[i].dma_map, DMA_MBOX_SIZE, DMA_BIDIRECTIONAL);
      |             ^~~~~~~~~~~~~~~~
      |             dma_unmap_single
/var/lib/dkms/kernel-mft-dkms/4.18.0/build/mst_pciconf.c: In function ‘ioctl.isra’:
/var/lib/dkms/kernel-mft-dkms/4.18.0/build/mst_pciconf.c:1269:1: warning: the frame size of 1184 bytes is larger than 1024 bytes [-Wframe-larger-than=]
 1269 | }
      | ^
cc1: some warnings being treated as errors
make[2]: *** [scripts/Makefile.build:260: /var/lib/dkms/kernel-mft-dkms/4.18.0/build/mst_pciconf.o] Error 1
make[1]: *** [Makefile:2026: /var/lib/dkms/kernel-mft-dkms/4.18.0/build] Error 2
make[1]: Leaving directory '/usr/src/linux-headers-6.2.16-3-pve'
make: *** [Makefile:53: all] Error 2
 
Yes, so it seems this kernel module is not compatible with the v6.2 kernel. You will have to remove it before installing the kernel. Do you have a card which depends on this module or is this a leftover? Maybe the most recent version [0] of the kernel module might work, but there is no guarantee.

[0] https://linux.mellanox.com/public/r...3.04/x86_64/kernel-mft-dkms_4.24.0-72_all.deb


There is the mlx4_core module shipped with the pve kernel which should work, no need for the DKMS module.

Edit: Oops, bin mit den Sprachen durcheinander gekommen, bitte um nachfrage falls die Antwort unverständlich ist
 
Last edited:
In my system runs a mellanox connect x-3 networtcard. At the moment, the system runs with 6.2.6-1-pve but 6.2.16-3-pve makes the Problem.
 
In my system runs a mellanox connect x-3 networtcard. At the moment, the system runs with 6.2.6-1-pve but 6.2.16-3-pve makes the Problem.
Well according to the Downloads tab in https://network.nvidia.com/products/ethernet-drivers/linux/mlnx_en/
Code:
Note: MLNX_EN 4.9-x LTS should be used by customers who would like to utilize one of the following:

* ConnectX-3 Pro
* ConnectX-3

Note: All of the above are not available on MLNX_OFED 5.x branch.

Note: MLNX_OFED 5.4-x//5.8-x LTS should be used by customers who would like to keep using stable 5.4-x//5.8-x deployment and get:           

* Critical bug fixes
* Support for new major OSs

it seems MLNX_EN 4.9 got it last update in 2020 https://linux.mellanox.com/public/repo/mlnx_en/4.9-2.2.4.0/

Which repository/source did you use to install the dkms module? Maybe you can find a patched version there?
 
Last edited:
Yes, so it seems this kernel module is not compatible with the v6.2 kernel. You will have to remove it before installing the kernel. Do you have a card which depends on this module or is this a leftover? Maybe the most recent version [0] of the kernel module might work, but there is no guarantee.

[0] https://linux.mellanox.com/public/r...3.04/x86_64/kernel-mft-dkms_4.24.0-72_all.deb

Edit: Oops, bin mit den Sprachen durcheinander gekommen, bitte um nachfrage falls die Antwort unverständlich ist
This helps.

Code:
Linux s00 6.2.16-3-pve #1 SMP PREEMPT_DYNAMIC PVE 6.2.16-3 (2023-06-17T05:58Z) x86_64 GNU/Linux

I updated two computer with the same Mellanox Nic. But these two computer had not this issue.
 
This helps.

Code:
Linux s00 6.2.16-3-pve #1 SMP PREEMPT_DYNAMIC PVE 6.2.16-3 (2023-06-17T05:58Z) x86_64 GNU/Linux

I updated two computer with the same Mellanox Nic. But these two computer had not this issue.
Maybe there you had a different version of the dkms module installed?
 
Wasn't able to find the source, but I'm also using ConnectX-3 and I've read that Nvidia dropped support, so there will be no new drivers.
That would be very annoying if PVE8 wouldn't run on my servers because they really need the ConnectX-3 NICs. Doesn't PVE come with MLNX_EN 4 drivers integrated? I can't remember needing to install my own drivers when setting up the servers.
 
Wasn't able to find the source, but I'm also using ConnectX-3 and I've read that Nvidia dropped support, so there will be no new drivers.
That would be very annoying if PVE8 wouldn't run on my servers because they really need the ConnectX-3 NICs. Doesn't PVE come with MLNX_EN 4 drivers integrated? I can't remember needing to install my own drivers when setting up the servers.
Not that I am aware of, could you check which kernel driver is in use on your PVE7 system by running lspci -v? Maybe the dkms module is not needed after all.
 
Not that I am aware of, could you check which kernel driver is in use on your PVE7 system by running lspci -v? Maybe the dkms module is not needed after all.
Sure.

Code:
01:00.0 Ethernet controller: Mellanox Technologies MT27500 Family [ConnectX-3]
        Subsystem: Mellanox Technologies MT27500 Family [ConnectX-3]
        Physical Slot: 4
        Flags: bus master, fast devsel, latency 0, IRQ 24, IOMMU group 48
        Memory at fb400000 (64-bit, non-prefetchable) [size=1M]
        Memory at f9000000 (64-bit, prefetchable) [size=8M]
        Expansion ROM at fb300000 [disabled] [size=1M]
        Capabilities: [40] Power Management version 3
        Capabilities: [48] Vital Product Data
        Capabilities: [9c] MSI-X: Enable+ Count=128 Masked-
        Capabilities: [60] Express Endpoint, MSI 00
        Capabilities: [c0] Vendor Specific Information: Len=18 <?>
        Capabilities: [100] Alternative Routing-ID Interpretation (ARI)
        Capabilities: [148] Device Serial Number <redacted>
        Capabilities: [154] Advanced Error Reporting
        Capabilities: [18c] Secondary PCI Express
        Kernel driver in use: mlx4_core
        Kernel modules: mlx4_core

Code:
modinfo mlx4_core

filename:       /lib/modules/5.15.107-2-pve/kernel/drivers/net/ethernet/mellanox/mlx4/mlx4_core.ko
version:        4.0-0
license:        Dual BSD/GPL
description:    Mellanox ConnectX HCA low-level driver
author:         Roland Dreier
srcversion:     C0CE0417CE61D0498534360
alias:          pci:v000015B3d00001010sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Fsv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Esv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Dsv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Csv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Bsv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Asv*sd*bc*sc*i*
alias:          pci:v000015B3d00001009sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001008sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001007sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001006sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001005sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001004sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001003sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001002sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000676Esv*sd*bc*sc*i*
alias:          pci:v000015B3d00006746sv*sd*bc*sc*i*
alias:          pci:v000015B3d00006764sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000675Asv*sd*bc*sc*i*
alias:          pci:v000015B3d00006372sv*sd*bc*sc*i*
alias:          pci:v000015B3d00006750sv*sd*bc*sc*i*
alias:          pci:v000015B3d00006368sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000673Csv*sd*bc*sc*i*
alias:          pci:v000015B3d00006732sv*sd*bc*sc*i*
alias:          pci:v000015B3d00006354sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000634Asv*sd*bc*sc*i*
alias:          pci:v000015B3d00006340sv*sd*bc*sc*i*
depends:
retpoline:      Y
intree:         Y
name:           mlx4_core
vermagic:       5.15.107-2-pve SMP mod_unload modversions
parm:           debug_level:Enable debug tracing if > 0 (int)
parm:           msi_x:0 - don't use MSI-X, 1 - use MSI-X, >1 - limit number of MSI-X irqs to msi_x (int)
parm:           num_vfs:enable #num_vfs functions if num_vfs > 0
num_vfs=port1,port2,port1+2 (array of byte)
parm:           probe_vf:number of vfs to probe by pf driver (num_vfs > 0)
probe_vf=port1,port2,port1+2 (array of byte)
parm:           log_num_mgm_entry_size:log mgm size, that defines the num of qp per mcg, for example: 10 gives 248.range: 7 <= log_num_mgm_entry_size <= 12. To activate device managed flow steering when available, set to -1 (int)
parm:           enable_64b_cqe_eqe:Enable 64 byte CQEs/EQEs when the FW supports this (default: True) (bool)
parm:           enable_4k_uar:Enable using 4K UAR. Should not be enabled if have VFs which do not support 4K UARs (default: false) (bool)
parm:           log_num_mac:Log2 max number of MACs per ETH port (1-7) (int)
parm:           log_num_vlan:Log2 max number of VLANs per ETH port (0-7) (int)
parm:           use_prio:Enable steering by VLAN priority on ETH ports (deprecated) (bool)
parm:           log_mtts_per_seg:Log2 number of MTT entries per segment (0-7) (default: 0) (int)
parm:           port_type_array:Array of port types: HW_DEFAULT (0) is default 1 for IB, 2 for Ethernet (array of int)
parm:           enable_qos:Enable Enhanced QoS support (default: off) (bool)
parm:           internal_err_reset:Reset device on internal errors if non-zero (default 1) (int)
 
Last edited:
Sure.

Code:
01:00.0 Ethernet controller: Mellanox Technologies MT27500 Family [ConnectX-3]
        Subsystem: Mellanox Technologies MT27500 Family [ConnectX-3]
        Physical Slot: 4
        Flags: bus master, fast devsel, latency 0, IRQ 24, IOMMU group 48
        Memory at fb400000 (64-bit, non-prefetchable) [size=1M]
        Memory at f9000000 (64-bit, prefetchable) [size=8M]
        Expansion ROM at fb300000 [disabled] [size=1M]
        Capabilities: [40] Power Management version 3
        Capabilities: [48] Vital Product Data
        Capabilities: [9c] MSI-X: Enable+ Count=128 Masked-
        Capabilities: [60] Express Endpoint, MSI 00
        Capabilities: [c0] Vendor Specific Information: Len=18 <?>
        Capabilities: [100] Alternative Routing-ID Interpretation (ARI)
        Capabilities: [148] Device Serial Number <redacted>
        Capabilities: [154] Advanced Error Reporting
        Capabilities: [18c] Secondary PCI Express
        Kernel driver in use: mlx4_core
        Kernel modules: mlx4_core

Are there any commands to get more infos on the used driver?
Yes I see the symbols also for the 6.2.16-3 kernel. modinfo shows:
Code:
# modinfo mlx4_core
filename:       /lib/modules/6.2.16-3-pve/kernel/drivers/net/ethernet/mellanox/mlx4/mlx4_core.ko
version:        4.0-0
license:        Dual BSD/GPL
description:    Mellanox ConnectX HCA low-level driver
author:         Roland Dreier
srcversion:     2057ACB62A9DF549F4B6AB3
alias:          pci:v000015B3d00001010sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Fsv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Esv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Dsv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Csv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Bsv*sd*bc*sc*i*
alias:          pci:v000015B3d0000100Asv*sd*bc*sc*i*
alias:          pci:v000015B3d00001009sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001008sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001007sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001006sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001005sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001004sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001003sv*sd*bc*sc*i*
alias:          pci:v000015B3d00001002sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000676Esv*sd*bc*sc*i*
alias:          pci:v000015B3d00006746sv*sd*bc*sc*i*
alias:          pci:v000015B3d00006764sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000675Asv*sd*bc*sc*i*
alias:          pci:v000015B3d00006372sv*sd*bc*sc*i*
alias:          pci:v000015B3d00006750sv*sd*bc*sc*i*
alias:          pci:v000015B3d00006368sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000673Csv*sd*bc*sc*i*
alias:          pci:v000015B3d00006732sv*sd*bc*sc*i*
alias:          pci:v000015B3d00006354sv*sd*bc*sc*i*
alias:          pci:v000015B3d0000634Asv*sd*bc*sc*i*
alias:          pci:v000015B3d00006340sv*sd*bc*sc*i*
depends:
retpoline:      Y
intree:         Y
name:           mlx4_core
vermagic:       6.2.16-3-pve SMP preempt mod_unload modversions
parm:           debug_level:Enable debug tracing if > 0 (int)
parm:           msi_x:0 - don't use MSI-X, 1 - use MSI-X, >1 - limit number of MSI-X irqs to msi_x (int)
parm:           num_vfs:enable #num_vfs functions if num_vfs > 0
num_vfs=port1,port2,port1+2 (array of byte)
parm:           probe_vf:number of vfs to probe by pf driver (num_vfs > 0)
probe_vf=port1,port2,port1+2 (array of byte)
parm:           log_num_mgm_entry_size:log mgm size, that defines the num of qp per mcg, for example: 10 gives 248.range: 7 <= log_num_mgm_entry_size <= 12. To activate device managed flow steering when available, set to -1 (int)
parm:           enable_64b_cqe_eqe:Enable 64 byte CQEs/EQEs when the FW supports this (default: True) (bool)
parm:           enable_4k_uar:Enable using 4K UAR. Should not be enabled if have VFs which do not support 4K UARs (default: false) (bool)
parm:           log_num_mac:Log2 max number of MACs per ETH port (1-7) (int)
parm:           log_num_vlan:Log2 max number of VLANs per ETH port (0-7) (int)
parm:           use_prio:Enable steering by VLAN priority on ETH ports (deprecated) (bool)
parm:           log_mtts_per_seg:Log2 number of MTT entries per segment (0-7) (default: 0) (int)
parm:           port_type_array:Array of port types: HW_DEFAULT (0) is default 1 for IB, 2 for Ethernet (array of int)
parm:           enable_qos:Enable Enhanced QoS support (default: off) (bool)
parm:           internal_err_reset:Reset device on internal errors if non-zero (default 1) (int)
 
The ConnectX-3 runs without any extra drivers on Debian. I bought my NICs used for my homelab. The firmware on my NICs was very old. Therefore, I installed the driver and updated the firmware on the NICs.
 
Last edited:

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!