You need to have the discard flag set on the disk before you run the fstrim command.
If you don't have it enabled the fstrim command will run but won't actually be passed through to the underlying disk.
You need to make sure discard is enabled, VM stopped and started and then fstirm enable...
If you enable discard and then run a fstrim on the VM after it’s enabled and you have applied it to the VM it will clear down the space.
This is a limitation of QEMU live migration
It will ignore any differences between the two copy’s the the PG’s will bring the PG online but the data within it will be corrupt and/or missing.
But being you only have two copies of the data as was running replication of 2 you don’t have too many options.
This is impossible to answer without knowing your setup. But there is 0 reasons why you wouldn’t be able to use the same IP.
However going by your limited expirence and fact this is a hobby project ZFS may not be the best option due to the extra complexity you will need to learn and manage...
You would need to wipe and start from fresh as the underlying setup of the Filesystem would need to change from what ever your currently running to ZFS.
You'd need to backup your VM's wipe and run through the installer to setup ZFS: https://pve.proxmox.com/wiki/ZFS_on_Linux
No, it will allow you to repair and or tell CEPH they are PG's with lost data and to continue.
But I think at this point you have data that was inflight to the disks that has been lost, hence you have PG's in the state they are, one of the core downsides to a size of 2 vs the recommended 3.
No problem as such, only issue with making RAID 0's instead of mixed mode with HBA is your underlying OS won't get full access to the disk and it's health as they are hidden behind the RAID.
Smartctl normally can connect via raid controllers but takes some extra steps.
What RAID controller is...
It's not quite clear exactly what you did at the start when this all happened?
Did you physically remove a disk whilst CEPH was online, all it takes is a write to not have been completed when your running 2/1 and you have a corrupt PG, if you did something on one of the disks outside of CEPH...
One question I have you state your aiming for 99% uptime, this is around 3 days a year downtime.
If that's the case sounds like a Active node with backups/data sync to a secondary source and the downtime of bringing up that secondary source is fine.
If you was aiming for more 99's then your...
What you want is a RAID card that supports mixed mode, where some disk's are put through to the OS as a HBA would and some are under RAID aka RAID5.
If your current card does not support mixed mode then your need to do separate RAID0's for the remaining disks for them to appear in the OS.
I can see 17.2.4 is now live, however Ceph released 17.2.5 as they missed some patches into 17.2.4.
Has Proxmox back ported these into the 17.2.4 or shall we expected a 17.2.5 soon?
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.