Hey y'all,
So I wrote a post about this in a post before but it was far too much information and hard to read through, so I am removing the old post and consolidating the information into this post now that I have some more information after exploring a bit. So let's get started.
For my current setup I am using a R720XD that has a PERC H710P Mini flashed to IT mode for pass-through functionality. I have 2 1TB SSDs that are used for Proxmox, LXCs, and VMs, these drives are totally fine. I also have 10 1.2 TB 10K SAS drives (with 2 more on the way) all of the same model (ST1200MM0108 Seagate Secure SED FIPS 140-2) used for media storage, backup, NAS, etc.
When I initially received the server it came with 8 drives. Of these 8 drives, 1 didnt show up in proxmox GUI from the get-go, 1 was in a ZFS single disk array, 3 were in a RAIDz1 config, and the last 3 were also in a RAIDz1 config. The disk that was in a ZFS single disk array is fine and works well (
The SAS Drives
So these drives are the ones I am having a massive pain with. As of right now here is the status for the drives (based on what i can access with
So for these 6 disks that I am having issues with, directly after destroying the ZFS pools these disks displayed the following
In addition to that, whenever I tried to access a disk with
fsck
I also ran
boot errors
In addition, another error I saw with these disks was during boot and for a little while after proxmox fully booted, you can see this code repeat for the different disks several times pretty much the same for each one:
Unfortunately I have tried to understand what this means but I am still trying to figure it out. I really don't know what is going on here other than noticing the protection error which could be related to the fact that these drives are self-encrypting drives. Not sure though.
In addition, I also tried using
Now the strange part!
So after doing all the above and doing a bunch of googling, i ended up trying to re-write the filesystem to the disk to see if that changed anything, sure enough it actually made 4 of the 6 disks show up in the proxmox GUI. I ran
Clearly, there is an issue right at the end when writing the superblocks and final filesystem stuff. That being said, of the 6 disks,
So to conclude, right now of the 6 disks that are having issues 4 seem to be usable (show up in proxmox GUI when creating ZFS pool) and 2 do not.
That's pretty much it, I'm sorry it's long but I figure every piece of information here could help. I really would love to understand what is going on here so if anyone has any idea, any and all comments are helpful! Are the 4 "usable" disks ok to use, or are they likely going to fail soon? What about the 2 that still do not show up in proxmox, is there anything I can do to make them appear?
Thank y'all so much in advance!
So I wrote a post about this in a post before but it was far too much information and hard to read through, so I am removing the old post and consolidating the information into this post now that I have some more information after exploring a bit. So let's get started.
For my current setup I am using a R720XD that has a PERC H710P Mini flashed to IT mode for pass-through functionality. I have 2 1TB SSDs that are used for Proxmox, LXCs, and VMs, these drives are totally fine. I also have 10 1.2 TB 10K SAS drives (with 2 more on the way) all of the same model (ST1200MM0108 Seagate Secure SED FIPS 140-2) used for media storage, backup, NAS, etc.
When I initially received the server it came with 8 drives. Of these 8 drives, 1 didnt show up in proxmox GUI from the get-go, 1 was in a ZFS single disk array, 3 were in a RAIDz1 config, and the last 3 were also in a RAIDz1 config. The disk that was in a ZFS single disk array is fine and works well (
sda
). One of the disks in the 3 disk arrays is also fine (sdd
).The SAS Drives
So these drives are the ones I am having a massive pain with. As of right now here is the status for the drives (based on what i can access with
fdisk
):- 4 of these drives (
sda, sdd, sdh, sdi
) work fine and i can access the disks with fdisk while also being able to select them when creating a ZFS pool from the proxmox GUIsdh
andsdi
I recently purchased and work finesda
was the disk in the single disk ZFS arraysdd
was one of the disks in the 2 3 disk RAIDz1 configurations
- The other 6 drives (
sdl, sdf, sdk, sde, sdj, sdg
) have issues, these are the disks that I will be referring to for the remainder of this post.sdl
was the disk that didnt show from the getgo, it has never been in a ZFS configurationsdf, sdk, sde, sdj, sdg
were in the 2 3 disk RAIDz1 configurations
So for these 6 disks that I am having issues with, directly after destroying the ZFS pools these disks displayed the following
fdisk -l
output, all the output was the same for each disk except for the IDs.In addition to that, whenever I tried to access a disk with
fdisk /dev/sdX
i would get an error that said fdisk: cannot open /dev/sdj: Input/output error
. Therefore I was unable to do anything with fdisk
to these disks.fsck
I also ran
fsck /dev/sdX
on the disks to see what information I got there and you can see the output in the following message, again all the output for the 6 disks was pretty much the same:boot errors
In addition, another error I saw with these disks was during boot and for a little while after proxmox fully booted, you can see this code repeat for the different disks several times pretty much the same for each one:
Code:
sd 0:0:11:0: [sdl] tag#9050 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s
sd 0:0:11:0: [sdl] tag#9050 Sense Key : Aborted Command [current]
sd 0:0:11:0: [sdl] tag#9050 Add. Sense: Logical block guard check failed
sd 0:0:11:0: [sdl] tag#9050 CDB: Read(32)
sd 0:0:11:0: [sdl] tag#9050 CDB[00]: 7f 00 00 00 00 00 00 18 00 09 20 00 00 00 00 00
sd 0:0:11:0: [sdl] tag#9050 CDB[10]: 8b ba 0b a8 8b ba 0b a8 00 00 00 00 00 00 00 08
blk_update_request: protection error, dev sdl, sector 2344225704 op 0x0:(READ) flags 0x0 phys_seg 1 prio class 0
Buffer I/O error on dev sdl, logical block 293028213, async page read
sd 0:0:11:0: [sdl] tag#9249 FAILED Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE cmd_age=0s
sd 0:0:11:0: [sdl] tag#9249 Sense Key : Aborted Command [current]
sd 0:0:11:0: [sdl] tag#9249 Add. Sense: Logical block guard check failed
sd 0:0:11:0: [sdl] tag#9249 CDB: Read(32)
sd 0:0:11:0: [sdl] tag#9249 CDB[00]: 7f 00 00 00 00 00 00 18 00 09 20 00 00 00 00 00
sd 0:0:11:0: [sdl] tag#9249 CDB[10]: 8b ba 0b a8 8b ba 0b a8 00 00 00 00 00 00 00 08
blk_update_request: protection error, dev sdl, sector 2344225704 op 0x0:(READ) flags 0x80700 phys_seg 1 prio class 0
In addition, I also tried using
dd
to wipe the disk and fill it with zeros but that didn't seem to do anything and the partitions still remained after. I also ran smrtctl -a /dev/sdX
which spat out output with no errors and marked the drives as ok.Now the strange part!
So after doing all the above and doing a bunch of googling, i ended up trying to re-write the filesystem to the disk to see if that changed anything, sure enough it actually made 4 of the 6 disks show up in the proxmox GUI. I ran
mkfs.ext2 /dev/sdX
on each of the disks and obtained the following output for each one:Clearly, there is an issue right at the end when writing the superblocks and final filesystem stuff. That being said, of the 6 disks,
sdf, sdj, sdg, sdl
now appear in the proxmox GUI, but sde, sdk
still do not show up (these 2 disks did show up when I created the first RAIDz1 arrays). Furthermore, I still cant access any of these 6 disks with fdisk
, but, when I run fdisk -l
the disks no longer show the paritions or the red "corrupt GPT error".So to conclude, right now of the 6 disks that are having issues 4 seem to be usable (show up in proxmox GUI when creating ZFS pool) and 2 do not.
That's pretty much it, I'm sorry it's long but I figure every piece of information here could help. I really would love to understand what is going on here so if anyone has any idea, any and all comments are helpful! Are the 4 "usable" disks ok to use, or are they likely going to fail soon? What about the 2 that still do not show up in proxmox, is there anything I can do to make them appear?
Thank y'all so much in advance!