I live-migrated most of the Windows machines last night. I still see the crc errors (235 in total over all the osd's on one particular node). How can I check if they Win VM's are actually now not using krbd anymore?
Update: I found how to: qm...
The error message I get looks somewhat different:
I see in ceph-osd.7.log:
2025-11-03T15:45:22.665+0200 7fbcdbcd7700 0 bad crc in data 1513571956 != exp 3330889006 from v1:192.168.131.3:0/3917894537
In that post the error indicates it's...
All 8 SSD's record crc errors, between 5 and 20 per day. There's no hardware raid involved. The carrier card is the one supplied by SuperMicro, not even a 3rd party one.
We do use krbd on our ceph storage pool for the improved performance it...
We have 4 nodes (dual Xeon CPU's, 256G RAM, 4 NVMe SSD's, 4 HDD's and dual Melanox 25Gb/s sfp's) in a cluster. Randomly I have started noticing crc errors in the osd logs.
Node B, osd.6
2025-10-23T10:32:59.808+0200 7f22a75bf700 0 bad crc in...
That's inconsequential. That Node was down and had started up, but PBS1 and InfluxDB were not started yet.
I have attached all the records from journalctl between 12:00 and the completed shutdown of the Node. I don't see any reason why the...
We had a node go down two days ago and I'm at a loss figuring out why.
I attached the log. This happened at 12:30. The other nodes simply show that the OSD's when down and feverishly started rebalancing the cluster.
Is there any indication...
I have relatively new Samsung Enterprise NVMe in a node that is generating the following error:
...
2025-08-26T15:56:43.870+0200 7fe8ac968700 0 bad crc in data 3326000616 != exp 1246001655 from v1:192.168.131.4:0/1799093090...
No, @birdflewza. I didn't pursue this any further, since the customer that requested it didn't want it anymore. It's on our list though, so we'll visit this again some time.