Hello
Unclear behavior most probably hardware related, may be config related.
Supermicro barebone with 8 x 3.5 " SAS disks on LSI SAS 3008 in IT-mode.
While reboot, the raidz looses the devices on phy 6 and 7. They appear as "faulted" with a number insted of device.
Example:
The disks are appearing at boot, at least they where seen in dmesg with no problems.
I was able to clear the disk and to rejoin with
The device was ok for some hours and now again shows CKSUM errors. But the disk itself has been replaced.
What could be a problem, the write cache of the disks is "ON" and I dunno how to reconfigure to "OFF".
The behavior is just since we had a power outage, so it might also be a problem with the backplane.
Or perhaps, PBS does a unclean shutdown? I don't think.[/CODE]
Unclear behavior most probably hardware related, may be config related.
Supermicro barebone with 8 x 3.5 " SAS disks on LSI SAS 3008 in IT-mode.
While reboot, the raidz looses the devices on phy 6 and 7. They appear as "faulted" with a number insted of device.
Example:
Bash:
config:
NAME STATE READ WRITE CKSUM
backup-01-pool01 DEGRADED 0 0 0
raidz2-0 DEGRADED 0 0 0
sda ONLINE 0 0 0
sdb ONLINE 0 0 0
sdc ONLINE 0 0 0
sdf ONLINE 0 0 0
sdd ONLINE 0 0 0
sdg1 DEGRADED 0 0 51 too many errors
sde ONLINE 0 0 0
18086653234219275637 FAULTED 0 0 0 was /dev/sdh1
The disks are appearing at boot, at least they where seen in dmesg with no problems.
I was able to clear the disk and to rejoin with
Bash:
zpool labelclear -f /dev/sdg1
zpool replace backup-01-pool01 /dev/sdg1
The device was ok for some hours and now again shows CKSUM errors. But the disk itself has been replaced.
What could be a problem, the write cache of the disks is "ON" and I dunno how to reconfigure to "OFF".
The behavior is just since we had a power outage, so it might also be a problem with the backplane.
Or perhaps, PBS does a unclean shutdown? I don't think.[/CODE]