As has already been said, this is not due to Proxmox.
@tgx You will notice on this forum, that Proxmox VE is entirely flawless product.
Only consumer SSDs without PLP are subject to such quick wear.
@tgx If you are curious,
apt install iotop
, then create e.g.
10 nodes (you can completely virtualise this), then create
10 resources, even just containers,
best with shared storage off the nodes, even something completely idling by is fine. Then
activate High Availability on them. Then check (on any single node that you won't be taking down during this exercise) -
iotop -oP
(once interactive press
a
for cumulative results), watch for (amongst others)
pmxcfs, and start migrating those resources around a bit, you can simulate some nodes dying in the process, bring them back up, etc.). You can compare this with what's going through to
systemd-journald. And make up your mind, imagine how it scales.
Were the SSDs previously operated on a raid controller with battery cache? The SSDs are extremely protected and the I/O is also written in an optimized way to keep wear to a minimum. If you then operate the same SSDs with a RaidZ1 or RaidZ2, for example, you have the highest possible write amplification and therefore much more wear.
@tgx And
combine it with this piece of information and the
choice of ZFS.
If you want to turn your existing ESXi servers into PVE hosts and have cheap boot SSDs on a raid controller, simply leave the Raid1 in place and install PVE with ext4.
@tgx Yes. And check if you mind the iotop numbers you get with the above experiment.