We have 2 labs setup with Proxmox VE 4.0 from latest ISO Download.
In one lab we have HP proliant servers with massive kernel panic on Module hpwdt.ko.
Unfortunately we do not have the trace due to HP's dammed ILO :-( but I will give mor Info when catched it up.
We have a ceph cluster with 3 hosts, 3 monitors up and running on this lab and erverything seems to be quite good.
We can start VM's, also migrate them but as soon you activate HA for any VM we receive a kernel panic on the hhwdt.ko module.
We have DL 360 G6 (lates Bios patches) and a DL380 G( running in this lab.
'This are the versions we are running.
proxmox-ve: 4.0-16 (running kernel: 4.2.2-1-pve)
pve-manager: 4.0-50 (running version: 4.0-50/d3a6b7e5)
pve-kernel-4.2.2-1-pve: 4.2.2-16
lvm2: 2.02.116-pve1
corosync-pve: 2.3.5-1
libqb0: 0.17.2-1
pve-cluster: 4.0-23
qemu-server: 4.0-31
pve-firmware: 1.1-7
libpve-common-perl: 4.0-32
libpve-access-control: 4.0-9
libpve-storage-perl: 4.0-27
pve-libspice-server1: 0.12.5-1
vncterm: 1.2-1
pve-qemu-kvm: 2.4-10
pve-container: 1.0-10
pve-firewall: 2.0-12
pve-ha-manager: 1.0-10
ksm-control-daemon: 1.2-1
glusterfs-client: 3.5.2-2+deb8u1
lxc-pve: 1.1.3-1
lxcfs: 0.9-pve2
cgmanager: 0.37-pve2
criu: 1.6.0-1
zfsutils: 0.6.5-pve4~jessie
Anything known about this kernel panics?
I found some hints googling around.
- blacklisting hpwdt was suggested but not the solution for VE, since we need the watchdog interfaces.
- I also tried grub parameters:
-- noautogroup and
-- intel_idle.max_cstates=0
with no success.
Since we have no debug symbols for the kernel (I did not find any package about this....), I could not use kdump to catch the panic up.
Any advise which could help or anone having problem like this.
In one lab we have HP proliant servers with massive kernel panic on Module hpwdt.ko.
Unfortunately we do not have the trace due to HP's dammed ILO :-( but I will give mor Info when catched it up.
We have a ceph cluster with 3 hosts, 3 monitors up and running on this lab and erverything seems to be quite good.
We can start VM's, also migrate them but as soon you activate HA for any VM we receive a kernel panic on the hhwdt.ko module.
We have DL 360 G6 (lates Bios patches) and a DL380 G( running in this lab.
'This are the versions we are running.
proxmox-ve: 4.0-16 (running kernel: 4.2.2-1-pve)
pve-manager: 4.0-50 (running version: 4.0-50/d3a6b7e5)
pve-kernel-4.2.2-1-pve: 4.2.2-16
lvm2: 2.02.116-pve1
corosync-pve: 2.3.5-1
libqb0: 0.17.2-1
pve-cluster: 4.0-23
qemu-server: 4.0-31
pve-firmware: 1.1-7
libpve-common-perl: 4.0-32
libpve-access-control: 4.0-9
libpve-storage-perl: 4.0-27
pve-libspice-server1: 0.12.5-1
vncterm: 1.2-1
pve-qemu-kvm: 2.4-10
pve-container: 1.0-10
pve-firewall: 2.0-12
pve-ha-manager: 1.0-10
ksm-control-daemon: 1.2-1
glusterfs-client: 3.5.2-2+deb8u1
lxc-pve: 1.1.3-1
lxcfs: 0.9-pve2
cgmanager: 0.37-pve2
criu: 1.6.0-1
zfsutils: 0.6.5-pve4~jessie
Anything known about this kernel panics?
I found some hints googling around.
- blacklisting hpwdt was suggested but not the solution for VE, since we need the watchdog interfaces.
- I also tried grub parameters:
-- noautogroup and
-- intel_idle.max_cstates=0
with no success.
Since we have no debug symbols for the kernel (I did not find any package about this....), I could not use kdump to catch the panic up.
Any advise which could help or anone having problem like this.