A drive on one of my nodes is constantly throwing ZFS write errors on my rpool (triple mirror, so not currently worried about data loss). Based on SMART self test, age of the drive, and the fact that I think it started when I was moving some cabling around, I'm pretty confident the issue is due to a bad SATA cable.
However this drive also has a partition used as a Ceph OSD, and I can't find any indication of any problems. I've searched the ceph-osd.#.log and ceph.log files and I haven't found any indication of any errors. I'm pretty sure that if ZFS is seeing write errors, then ceph is seeing them too. I trust that Ceph is taking care of things in the background, and I'm not worried about data loss, but if this was to happen on a drive not in a ZFS pool, I would like to see the issue and take action before it becomes a larger issue.
Does anyone know of any specific log entries or other places I can look to see if there are write errors? Has anyone else run into ceph being too good at self repairing in the background so you aren't aware of minor issues until it becomes a major one?
Thanks!
However this drive also has a partition used as a Ceph OSD, and I can't find any indication of any problems. I've searched the ceph-osd.#.log and ceph.log files and I haven't found any indication of any errors. I'm pretty sure that if ZFS is seeing write errors, then ceph is seeing them too. I trust that Ceph is taking care of things in the background, and I'm not worried about data loss, but if this was to happen on a drive not in a ZFS pool, I would like to see the issue and take action before it becomes a larger issue.
Does anyone know of any specific log entries or other places I can look to see if there are write errors? Has anyone else run into ceph being too good at self repairing in the background so you aren't aware of minor issues until it becomes a major one?
Thanks!