Problem in detecting two equal nvme

codedmind

Active Member
Oct 19, 2020
11
0
41
35
Hello,

I have a hpe dl20 gen11 with proxmox 9.
As this server don't support bifurcation i get an axagon pcem2-nd pcie. I install two nvme and configure a zfspool in proxmox, late july.
But now, after some time i get a warning that the pool is degrated because one of the disk are missing.
In bios and in the boot menu the two disk are visible, in proxmox i still only get one. I try remove one of this from the slot and boot to see if the missing disk is detected, but no luck.

I dig into the boot log and get this information...
Code:
Oct 30 10:56:07 pvemmi kernel: pci 0000:05:08.0:   bridge window [mem 0xd5d00000-0xd5dfffff]
Oct 30 10:56:07 pvemmi kernel: pci 0000:05:08.0: enabling Extended Tags
Oct 30 10:56:07 pvemmi kernel: pci 0000:05:08.0: PME# supported from D0 D3hot D3cold
Oct 30 10:56:07 pvemmi kernel: pci 0000:04:00.0: PCI bridge to [bus 05-07]
Oct 30 10:56:07 pvemmi kernel: pci 0000:05:00.0: bridge configuration invalid ([bus 06-06]), reconfiguring
Oct 30 10:56:07 pvemmi kernel: pci 0000:07:00.0: [c0a9:5426] type 00 class 0x010802 PCIe Endpoint
Oct 30 10:56:07 pvemmi kernel: pci 0000:07:00.0: BAR 0 [mem 0xd5d00000-0xd5d03fff 64bit]
Oct 30 10:56:07 pvemmi kernel: pci 0000:07:00.0: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0000:00:01.0 (capable of 63.012 Gb/s with 16.0 GT/s PCIe x4 link)
Oct 30 10:56:07 pvemmi kernel: pci 0000:05:08.0: PCI bridge to [bus 07]
Oct 30 10:56:07 pvemmi kernel: pci_bus 0000:08: busn_res: can not insert [bus 08-07] under [bus 05-07] (conflicts with (null) [bus 05-07])
Oct 30 10:56:07 pvemmi kernel: pci 0000:05:00.0: PCI bridge to [bus 08-07]
Oct 30 10:56:07 pvemmi kernel: pci_bus 0000:08: busn_res: [bus 08-07] end is updated to 08
Oct 30 10:56:07 pvemmi kernel: pci_bus 0000:08: busn_res: can not insert [bus 08] under [bus 05-07] (conflicts with (null) [bus 05-07])
Oct 30 10:56:07 pvemmi kernel: pci 0000:05:00.0: devices behind bridge are unusable because [bus 08] cannot be assigned for them
Oct 30 10:56:07 pvemmi kernel: pci 0000:04:00.0: bridge has subordinate 07 but max busn 08
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.0: [14e4:1657] type 00 class 0x020000 PCIe Endpoint
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.0: BAR 0 [mem 0x40000b0000-0x40000bffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.0: BAR 2 [mem 0x40000a0000-0x40000affff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.0: BAR 4 [mem 0x4000090000-0x400009ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.0: ROM [mem 0x00000000-0x0003ffff pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.0: PME# supported from D0 D3hot D3cold
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.0: 8.000 Gb/s available PCIe bandwidth, limited by 5.0 GT/s PCIe x2 link at 0000:00:1c.0 (capable of 16.000 Gb/s with 5.0 GT/s PCIe x4 link)
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.1: [14e4:1657] type 00 class 0x020000 PCIe Endpoint
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.1: BAR 0 [mem 0x4000080000-0x400008ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.1: BAR 2 [mem 0x4000070000-0x400007ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.1: BAR 4 [mem 0x4000060000-0x400006ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.1: ROM [mem 0x00000000-0x0003ffff pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.1: PME# supported from D0 D3hot D3cold
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.2: [14e4:1657] type 00 class 0x020000 PCIe Endpoint
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.2: BAR 0 [mem 0x4000050000-0x400005ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.2: BAR 2 [mem 0x4000040000-0x400004ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.2: BAR 4 [mem 0x4000030000-0x400003ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.2: ROM [mem 0x00000000-0x0007ffff pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.2: PME# supported from D0 D3hot D3cold
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.3: [14e4:1657] type 00 class 0x020000 PCIe Endpoint
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.3: BAR 0 [mem 0x4000020000-0x400002ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.3: BAR 2 [mem 0x4000010000-0x400001ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.3: BAR 4 [mem 0x4000000000-0x400000ffff 64bit pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.3: ROM [mem 0x00000000-0x0007ffff pref]
Oct 30 10:56:07 pvemmi kernel: pci 0000:01:00.3: PME# supported from D0 D3hot D3cold
Oct 30 10:56:07 pvemmi kernel: pci 0000:00:1c.0: PCI bridge to [bus 01]

Then i try the card in other computer (with windows) and both nvme are detected... the last test i did was using two different nvmes in the proxmox server and then the two nvmes are again detected...

So i don't know what happen because when i install the proxmox the both disk are detected and i'm able to create a zpool and then after some time (i do all the pve upgrades) i lose one nvme... but if i use two different brands it works...

Anyone can help?
 
Solved with other pcie card.
Startech PEX8M2E2 with ASM2824 chip with works.
The other pcie card have the asm2812 chip that don't work hpe dl20 and proxmox
 
Last edited: