We are using Dell servers - the servers that are failing are using Dell PERC H700 RAID controllers with write / read caching enabled. We have a cluster of 6 new Dell 620 servers that use the Dell PERC H310 controller - these servers are working fine and backups are happening without any issue. Below I will post the details of the working and non-working hardware.
We have tried both schedulers with no luck. I did do the latest kernel on the failing server and will reboot it this weekend and attempt backups again and post my results. Thanks everyone for trying to sort this problem out, it's a painful one. Hopefully this information will be useful to someone else for comparison. I have also posted other kernel logs / info in this and other threads. I agree with the ProxMox team that it now appears kernel / RAID or filesystem driver related. I am digging more into that as well.
Cheers,
Joe Jenkins
All Servers are DELL -
FAILING SERVER - please note it is now running the new kernel / pve released a day or two ago - I have NOT yet rebooted the box to retest VMs, but will soon!
megaraid_sas 0000:05:00.0: irq 100 for MSI/MSI-X
ahci 0000:00:11.0: version 3.0
alloc irq_desc for 22 on node 0
alloc kstat_irqs on node 0
ahci 0000:00:11.0: PCI INT A -> GSI 22 (level, low) -> IRQ 22
ahci 0000:00:11.0: AHCI 0001.0100 32 slots 4 ports 3 Gbps 0xf impl SATA mode
ahci 0000:00:11.0: flags: 64bit ncq sntf ilck pm led clo pmp pio slum part ccc
scsi1 : ahci
scsi2 : ahci
scsi3 : ahci
scsi4 : ahci
ata1: SATA max UDMA/133 abar m1024@0xef2ff800 port 0xef2ff900 irq 22
ata2: SATA max UDMA/133 abar m1024@0xef2ff800 port 0xef2ff980 irq 22
ata3: SATA max UDMA/133 abar m1024@0xef2ff800 port 0xef2ffa00 irq 22
ata4: SATA max UDMA/133 abar m1024@0xef2ff800 port 0xef2ffa80 irq 22
megasas_init_mfi: fw_support_ieee=67108864
megasas: INIT adapter done
scsi0 : LSI SAS based MegaRAID driver
scsi 0:0:0:0: Direct-Access SEAGATE ST9146803SS FS64 PQ: 0 ANSI: 5
Refined TSC clocksource calibration: 1900.022 MHz.
Switching to clocksource tsc
scsi 0:0:1:0: Direct-Access SEAGATE ST9146803SS FS64 PQ: 0 ANSI: 5
scsi 0:0:2:0: Direct-Access SEAGATE ST9146803SS FS64 PQ: 0 ANSI: 5
scsi 0:0:32:0: Enclosure DP BACKPLANE 1.07 PQ: 0 ANSI: 5
scsi 0:2:0:0: Direct-Access DELL PERC H700 2.10 PQ: 0 ANSI: 5
sd 0:2:0:0: [sda] 570949632 512-byte logical blocks: (292 GB/272 GiB)
sd 0:2:0:0: [sda] Write Protect is off
sd 0:2:0:0: [sda] Mode Sense: 1f 00 00 08
sd 0:2:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA
sda: sda1 sda2
sd 0:2:0:0: [sda] Attached SCSI disk
pve-manager: 3.0-23 (pve-manager/3.0/957f0862)
running kernel: 2.6.32-20-pve
proxmox-ve-2.6.32: 3.0-107
pve-kernel-2.6.32-22-pve: 2.6.32-107
pve-kernel-2.6.32-20-pve: 2.6.32-100
lvm2: 2.02.95-pve3
clvm: 2.02.95-pve3
corosync-pve: 1.4.5-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.0-1
pve-cluster: 3.0-4
qemu-server: 3.0-20
pve-firmware: 1.0-23
libpve-common-perl: 3.0-4
libpve-access-control: 3.0-4
libpve-storage-perl: 3.0-8
vncterm: 1.1-4
vzctl: 4.0-1pve3
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 1.4-13
ksm-control-daemon: 1.1-1
primary mounts:
/dev/sda1 on /boot type ext3 (rw,relatime,errors=continue,user_xattr,acl,barrier=0,data=ordered)
/dev/mapper/pve-root on / type ext3 (rw,relatime,errors=remount-ro,user_xattr,acl,barrier=0,data=ordered)
/dev/fuse on /etc/pve type fuse (rw,nosuid,nodev,relatime,user_id=0,group_id=0,default_permissions,allow_other)
/dev/mapper/xx5-vz on /var/lib/vz type ext4 (rw,noatime,barrier=1,nodelalloc,data=ordered,_netdev)
All VMs run on an iSCSI mount / ext4 (see last line above)
----------------------------------------------------------------------------------------------------
WORKING SERVERS - Dell 620, brand new, no issues with backups as of yet.
scsi1 : ahci
scsi2 : ahci
scsi3 : ahci
scsi4 : ahci
scsi5 : ahci
scsi6 : ahci
ata1: SATA max UDMA/133 abar m2048@0xdf8ff000 port 0xdf8ff100 irq 105
ata2: SATA max UDMA/133 abar m2048@0xdf8ff000 port 0xdf8ff180 irq 105
ata3: SATA max UDMA/133 abar m2048@0xdf8ff000 port 0xdf8ff200 irq 105
ata4: SATA max UDMA/133 abar m2048@0xdf8ff000 port 0xdf8ff280 irq 105
ata5: SATA max UDMA/133 abar m2048@0xdf8ff000 port 0xdf8ff300 irq 105
ata6: SATA max UDMA/133 abar m2048@0xdf8ff000 port 0xdf8ff380 irq 105
megasas_init_mfi: fw_support_ieee=67108864
megasas: INIT adapter done
scsi0 : LSI SAS based MegaRAID driver
scsi 0:0:0:0: Direct-Access ATA ST9250610NS AA09 PQ: 0 ANSI: 5
scsi 0:0:1:0: Direct-Access ATA ST9250610NS AA09 PQ: 0 ANSI: 5
scsi 0:0:2:0: Direct-Access ATA ST9250610NS AA09 PQ: 0 ANSI: 5
scsi 0:0:3:0: Direct-Access ATA ST9250610NS AA09 PQ: 0 ANSI: 5
scsi 0:0:32:0: Enclosure DP BP12G+ 1.00 PQ: 0 ANSI: 5
scsi 0:2:0:0: Direct-Access DELL PERC H310 2.12 PQ: 0 ANSI: 5
sd 0:2:0:0: [sda] 974651392 512-byte logical blocks: (499 GB/464 GiB)
sd 0:2:0:0: [sda] Write Protect is off
sd 0:2:0:0: [sda] Mode Sense: 1f 00 10 08
sd 0:2:0:0: [sda] Write cache: disabled, read cache: disabled, supports DPO and FUA
sda: sda1 sda2
sd 0:2:0:0: [sda] Attached SCSI disk
pve-manager: 3.0-23 (pve-manager/3.0/957f0862)
running kernel: 2.6.32-20-pve
proxmox-ve-2.6.32: 3.0-100
pve-kernel-2.6.32-20-pve: 2.6.32-100
lvm2: 2.02.95-pve3
clvm: 2.02.95-pve3
corosync-pve: 1.4.5-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.0-1
pve-cluster: 3.0-4
qemu-server: 3.0-20
pve-firmware: 1.0-22
libpve-common-perl: 3.0-4
libpve-access-control: 3.0-4
libpve-storage-perl: 3.0-8
vncterm: 1.1-4
vzctl: 4.0-1pve3
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 1.4-13
ksm-control-daemon: 1.1-1
primary mounts:
/dev/sda1 on /boot type ext3 (rw,relatime,errors=continue,user_xattr,acl,barrier=0,data=ordered)
/dev/mapper/pve-root on / type ext3 (rw,relatime,errors=remount-ro,user_xattr,acl,barrier=0,data=ordered)
/dev/mapper/pve-data on /var/lib/vz type ext3 (rw,relatime,errors=continue,user_xattr,acl,barrier=0,data=ordered)
/dev/fuse on /etc/pve type fuse (rw,nosuid,nodev,relatime,user_id=0,group_id=0,default_permissions,allow_other)
All VMs on this server run on a local EXT3 mount on the RAID controller (3rd line of mounts above)
---------------------------------------
This is likely not a vzdump problem (although looks related), since vzdump is a userland process which is probably unable to cause a system-wide IO freeze by itself.
The problem most likely lies in the kernel, and is influenced by several factors:
- raid controller (several raid controllers exhibited the problem, to varying degrees)
- logical volume manager (so far, lvm only)
- filesystem (ext3 less likely, ext4 more likely)
- io scheduler (deadline and cfq, so probably scheduler is not related)
- vzdump (problem shows up during vzdump backups, surely related)
- specific OpenVZ VE's (simfs ?)
For us it looks like this: vzdump snapshot backups start at 11 pm, by 3am they reach VE 215, and the entire system freezes (all disk IO stops, load keeps climbing to the sky). If VE 215 is moved off the server then the problem does not appear.
We run Proxmox VE 3.0 on Core i7 servers, Adaptec 6805E RAID10, ext4 filesystem, deadline scheduler.