High wait on raid-1 on one completly fine disk

iso23270

New Member
Mar 18, 2020
4
0
1
Germany
eholtz.de
Hi folks,

I've been using proxmox for years now and I never had any problems ... until recently.

I have a root server rented at a hoster and installed proxmox on it. There is a vanilla debian installation with a software raid on top and I used that to install proxmox. Everything was fine until last week, when one of the two hard drives failed. I got it replaced and shut everything down to give the raid rebuild top prio, but even then it was quite slow and took a couple of days. After the rebuild was complete I restarted my vms and containers and everything was really slow and IOwait jumped to 80%. That was odd. I tried tuning some parameters, but nothing helped. With a simple dd I tried to read from both disks and I could see, that the new disk was quite slow, when reading.
To verify that I booted into the rescue system and I could not verify this problem with the rescue system (which was an ubuntu, but I don't remember the actual kernel version). So I booted in my proxmox again and my problem was still there. I tried a couple other things but nothing. So I gave up and reinstalled the whole system, just to be sure. But even after a complete reinstallation the problem persists. When I have the new hard disk in the RAID-1 the system is terribly slow, when I mark is as faulty the system is back to normal state. I checked smart, but no failures there. Of course I also waited for the raid sync to complete before trying to measure anything.

Long story short: with both drives in my RAID the system is very slow. With the new drive disabled the system behaves normal. Both drives show no smart errors. I can't verify the behaviour with an ubuntu rescue system. I tried the no-subscription kernel and the new 5.4 kernel, but behaviour is the same.

Some measurements:

# both drives in the raid => r_await and w_await > 100 for the loop device (my one container that's running)
Bash:
eholtz@titan719:~$ iostat -xyz 30 -c 1
Linux 5.4.24-1-pve (titan719)   03/18/2020      _x86_64_        (8 CPU)

avg-cpu:  %user   %nice %system %iowait  %steal   %idle
           0.09    0.00    0.05    1.57    0.00    1.69

Device            r/s     w/s     rkB/s     wkB/s   rrqm/s   wrqm/s  %rrqm  %wrqm r_await w_await aqu-sz rareq-sz wareq-sz  svctm  %util                                                         
loop0           25.33    8.67    370.80     48.13     0.00     0.00   0.00   0.00  362.46  153.33  10.47    14.64     5.55   1.44   4.89                                                         
sda             15.93   29.77    375.20    205.87     0.03     7.83   0.21  20.83    4.04   14.13   0.43    23.55     6.92   1.49   6.79                                                         
md2             23.83   51.10    560.13    286.40     0.00     0.00   0.00   0.00    0.00    0.00   0.00    23.50     5.60   0.00   0.00                                                         
sdb              7.83   29.03    184.93    202.53     0.03     7.87   0.42  21.32   90.83   84.23   3.10    23.61     6.98   1.36   5.03


# new disk disabled => r_await and w_await < 10, same load situation as above
Bash:
eholtz@titan719:~$ iostat -xyz 30 -c 1
Linux 5.4.24-1-pve (titan719)   03/18/2020      _x86_64_        (8 CPU)

avg-cpu:  %user   %nice %system %iowait  %steal   %idle
           0.72    0.00    0.19    0.15    0.00    2.19

Device            r/s     w/s     rkB/s     wkB/s   rrqm/s   wrqm/s  %rrqm  %wrqm r_await w_await aqu-sz rareq-sz wareq-sz  svctm  %util                                                         
loop0           87.90   27.60   1283.60    131.20     0.00     0.00   0.00   0.00    5.54    5.22   0.56    14.60     4.75   1.72  19.91                                                         
sda             82.17   61.63   1854.27    381.48     0.07    11.20   0.08  15.38    3.45   12.08   0.87    22.57     6.19   1.65  23.69                                                         
md2             82.20   69.97   1857.60    376.80     0.00     0.00   0.00   0.00    0.00    0.00   0.00    22.60     5.39   0.00   0.00

Do you have any idea what this could be?

Best regards,
Eike
 
Long story short: with both drives in my RAID the system is very slow. With the new drive disabled the system behaves normal. Both drives show no smart errors. I can't verify the behaviour with an ubuntu rescue system. I tried the no-subscription kernel and the new 5.4 kernel, but behaviour is the same.
What model are those two disks?

Aside, Proxmox das not support mdraid, while it is technically doable.
 
Hi, Alwin,

both are seagate models:

Bash:
titan719:~# smartctl --all /dev/sda | head
smartctl 7.1 2019-12-30 r5022 [x86_64-linux-5.4.24-1-pve] (local build)
Copyright (C) 2002-19, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF INFORMATION SECTION ===
Model Family:     Seagate Barracuda 7200.14 (AF)
Device Model:     ST2000DM001-9YN164
Serial Number:   
LU WWN Device Id: 5 000c50 03fe3e486
Firmware Version: CC4B
User Capacity:    2,000,398,934,016 bytes [2.00 TB]

titan719:~# smartctl --all /dev/sdb | head
smartctl 7.1 2019-12-30 r5022 [x86_64-linux-5.4.24-1-pve] (local build)
Copyright (C) 2002-19, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF INFORMATION SECTION ===
Device Model:     ST2000DM008-2FR102
Serial Number:   
LU WWN Device Id: 5 000c50 0cd4701ee
Firmware Version: 0001
User Capacity:    2,000,398,934,016 bytes [2.00 TB]
Sector Sizes:     512 bytes logical, 4096 bytes physical

I thought about dropping the raid and replacing it with something else, but unfortunately I don't really have the choice on what to install on the root server and having the raid saved me from a completly crashed system.

Best regards,
Eike
 
I thought about dropping the raid and replacing it with something else, but unfortunately I don't really have the choice on what to install on the root server and having the raid saved me from a completly crashed system.
Understandable. Just to note, if possible, you could use ZFS or an onboard raid controller.

The sector sizes may be different for the disks (512E vs 4kn). Also there might be a firmware for those two drives. And last but not least, it may just be a faulty one. ;)
 
Thanks for your insight. Unfortunately I missed the details about the sector size. They are the same.

I'll try re-partitioning the disks to have separate partitions for the proxmox local storage that I can format with ZFS, but that will need a moment. I'll post a follow up if I was successful.

Best regards,
Eike

Just to be complete:
Code:
titan719:~# smartctl --all /dev/sda | head -n 16
smartctl 7.1 2019-12-30 r5022 [x86_64-linux-5.4.24-1-pve] (local build)
Copyright (C) 2002-19, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF INFORMATION SECTION ===
Model Family:     Seagate Barracuda 7200.14 (AF)
Device Model:     ST2000DM001-9YN164
Serial Number:   
LU WWN Device Id: 5 000c50 03fe3e486
Firmware Version: CC4B
User Capacity:    2,000,398,934,016 bytes [2.00 TB]
Sector Sizes:     512 bytes logical, 4096 bytes physical
Rotation Rate:    7200 rpm
Device is:        In smartctl database [for details use: -P show]
ATA Version is:   ATA8-ACS T13/1699-D revision 4
SATA Version is:  SATA 3.0, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is:    Wed Mar 18 15:34:09 2020 UTC

titan719:~# smartctl --all /dev/sdb | head -n 16
smartctl 7.1 2019-12-30 r5022 [x86_64-linux-5.4.24-1-pve] (local build)
Copyright (C) 2002-19, Bruce Allen, Christian Franke, www.smartmontools.org

=== START OF INFORMATION SECTION ===
Device Model:     ST2000DM008-2FR102
Serial Number:   
LU WWN Device Id: 5 000c50 0cd4701ee
Firmware Version: 0001
User Capacity:    2,000,398,934,016 bytes [2.00 TB]
Sector Sizes:     512 bytes logical, 4096 bytes physical
Rotation Rate:    7200 rpm
Form Factor:      3.5 inches
Device is:        Not in smartctl database [for details use: -P showall]
ATA Version is:   ACS-3 T13/2161-D revision 5
SATA Version is:  SATA 3.1, 6.0 Gb/s (current: 3.0 Gb/s)
Local Time is:    Wed Mar 18 15:34:12 2020 UTC
 
It was a bit tricky, but I managed to resize the partitions and create a zpool with mirror for the disk images of the vms and containers. The wait rates went back to normal and everything is running smooth again. If someone needs help for resizing running raids just PM me :-)
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!