New 2.6.32 Kernel with stable OpenVZ (pvetest)

Greetings to all. I have now something to report.

1. After upgrading my HP 380DL G5 Server to the new kernel 2.6.32-43, I can see the next issue in my syslog and on the console:

Code:
Sep 25 20:34:13 bond kernel: e1000e 0000:10:00.1: eth4: Detected Hardware Unit Hang:
Sep 25 20:34:13 bond kernel:  TDH                  <18>
Sep 25 20:34:13 bond kernel:  TDT                  <19>
Sep 25 20:34:13 bond kernel:  next_to_use          <19>
Sep 25 20:34:13 bond kernel:  next_to_clean        <18>
Sep 25 20:34:13 bond kernel: buffer_info[next_to_clean]:
Sep 25 20:34:13 bond kernel:  time_stamp           <10ac83994>
Sep 25 20:34:13 bond kernel:  next_to_watch        <18>
Sep 25 20:34:13 bond kernel:  jiffies              <10ac83e00>
Sep 25 20:34:13 bond kernel:  next_to_watch.status <0>
Sep 25 20:34:13 bond kernel: MAC Status             <80387>
Sep 25 20:34:13 bond kernel: PHY Status             <792d>
Sep 25 20:34:13 bond kernel: PHY 1000BASE-T Status  <3800>
Sep 25 20:34:13 bond kernel: PHY Extended Status    <3000>
Sep 25 20:34:13 bond kernel: PCI Status             <10>

I have a network adapter based on 82571EB chip installed (e1000e):
Code:
bond:/var/log# lspci | grep Ethernet
03:00.0 Ethernet controller: Broadcom Corporation NetXtreme II BCM5708 Gigabit Ethernet (rev 12)
05:00.0 Ethernet controller: Broadcom Corporation NetXtreme II BCM5708 Gigabit Ethernet (rev 12)
10:00.0 Ethernet controller: Intel Corporation 82571EB Gigabit Ethernet Controller (rev 06)
10:00.1 Ethernet controller: Intel Corporation 82571EB Gigabit Ethernet Controller (rev 06)
11:00.0 Ethernet controller: Intel Corporation 82571EB Gigabit Ethernet Controller (rev 06)
11:00.1 Ethernet controller: Intel Corporation 82571EB Gigabit Ethernet Controller (rev 06)

Code:
bond:/var/log# modinfo e1000e
filename:       /lib/modules/2.6.32-6-pve/kernel/drivers/net/e1000e/e1000e.ko
version:        1.5.1-NAPI
license:        GPL
description:    Intel(R) PRO/1000 Network Driver
author:         Intel Corporation, <linux.nics@intel.com>
srcversion:     2EE586A2A1672B0452748CD

Code:
bond:/var/log# pveversion -v
pve-manager: 1.9-24 (pve-manager/1.9/6542)
running kernel: 2.6.32-6-pve
proxmox-ve-2.6.32: 1.9-43
pve-kernel-2.6.32-6-pve: 2.6.32-43
qemu-server: 1.1-32
pve-firmware: 1.0-13
libpve-storage-perl: 1.0-19
vncterm: 0.9-2
vzctl: 3.0.28-1pve5
vzdump: 1.2.6-1
vzprocps: 2.0.11-2
vzquota: 3.0.12-3
pve-qemu-kvm: 0.15.0-1
ksm-control-daemon: 1.0-6

I have found something similiar here. As this is my test machine, I could not encounter any negative consequences of such behaviour. But I'm afraid to migrate real production load onto this box to avoid troubles.

2. I have noticed that bridged interfaces vmbr+ now go up much slower than before. So, my samba shares mentioned in /etc/fstab, cannot be mounted automatically during boot process. Before 2.6.32-6-pve I has not met such a trouble.

3. I don't know if this is a bug. But Debian's udev 164-3 is now incompatible with kernel or lvm. Boot process terminates abnormally during initramfs initialization stage, attempting to find and mount LVM volumes. At the same time udev 125-7 works well. I try to perform further tests now.
 
pls try the latest kernel from pvetest (pve-kernel-2.6.32-6-pve: 2.6.32-47), includes the latest e1000e from http://sourceforge.net/projects/e1000/files/e1000e stable/
Code:
modinfo e1000e
filename:       /lib/modules/2.6.32-6-pve/kernel/drivers/net/e1000e/e1000e.ko
version:        1.6.2-NAPI
license:        GPL
description:    Intel(R) PRO/1000 Network Driver
author:         Intel Corporation, <linux.nics@intel.com>
srcversion:     E4B90BE4DC230817DCFB165
 
pls try the latest kernel from pvetest (pve-kernel-2.6.32-6-pve: 2.6.32-47)
OK, I have tried it. The same effect exists, nothing changed.

Code:
bond:~# modinfo e1000e
filename:       /lib/modules/2.6.32-6-pve/kernel/drivers/net/e1000e/e1000e.ko
version:        1.6.2-NAPI
license:        GPL
description:    Intel(R) PRO/1000 Network Driver
author:         Intel Corporation, <linux.nics@intel.com>
srcversion:     E4B90BE4DC230817DCFB165

Code:
bond:~# pveversion -v
pve-manager: 1.9-24 (pve-manager/1.9/6542)
running kernel: 2.6.32-6-pve
proxmox-ve-2.6.32: 1.9-47
pve-kernel-2.6.32-6-pve: 2.6.32-47
qemu-server: 1.1-32
pve-firmware: 1.0-14
libpve-storage-perl: 1.0-19
vncterm: 0.9-2
vzctl: 3.0.29-2pve1
vzdump: 1.2.6-1
vzprocps: 2.0.11-2
vzquota: 3.0.12-3
pve-qemu-kvm: 0.15.0-1
ksm-control-daemon: 1.0-6
 
Some more things to report.

1. I have discovered, that magic spell "pcie_aspm=off" resolves my trouble with "Detected Hardware Unit Hang" on e1000e NIC (pve-kernel-2.6.32-6-pve: 2.6.32-47). I advise to add this unit into FAQ to avoid similiar issues in the future. But the only thing I cant understand is why everything worked well in older debian-built kernels.

2. The issue with too slow bridge initialization and CIFS I have resolved in a way like this:
Code:
#!/bin/sh
BRNAME="vmbr0"

sleep 30

# Let's bring up shares from fstab if not yet mounted
for share in $(/usr/bin/awk '$1 ~ /^\t*\/\// {print $2}' /etc/fstab)
do
 grep "$share" /proc/mounts > /dev/null || /bin/mount "$share"
done

exit 0

I do know, that it is not the best approach. And I would like to find a more robust solution.

3. The issue with udev seems to be my own mistake. Occasionally I have removed "/lib/udev/rules.d/55-dm.rules" and forgotten about this.
 
Another strange thing, that I have noticed in the new ".47" kernel.
Code:
~# dmesg | grep pcie
Command line: BOOT_IMAGE=/vmlinuz-2.6.32-6-pve root=UUID=020c4296-7c73-488c-b88c-7eb8d7f115d8 ro elevator=noop pcie_aspm=off
Kernel command line: BOOT_IMAGE=/vmlinuz-2.6.32-6-pve root=UUID=020c4296-7c73-488c-b88c-7eb8d7f115d8 ro elevator=noop pcie_aspm=off

And, at she same time on the same system, just after boot:
Code:
~# cat /sys/module/pcie_aspm/parameters/policy
[default] performance powersave

Can someone explain this?
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!