Hello all,
with one customer I run a Proxmox 5.4 Ceph cluster with 4 nodes. Two of the nodes have 24 GB of memory and 5 OSDs each, and these nodes are tasked only with one (important) VM each.
On both nodes we see a disturbing pattern:
- after reboot, all the ceph-osd processes start out below 1G of memory usage which would be ok
- then with time, the processes grow up to 4G of memory usage
- some out of memory conditions happen, sometimes only failed backups, sometimes crashes of the VM failing to get all of the allocated (in VM configuration) memory
To recap, servers with 24 GB memory and running only one (1) VM with about 8 GB main memory, can run only stable for about 1-2 weeks until crashes and out-of-memory-situations happen.
Why are the ceph-osd processes growing in their memory usage so much? How to limit their memory usage and how to solve this problem?
Both nodes should be able to take 2 VMs (16 GB) and only require 8 GB for proxmox system and Ceph.
Please help,
Andreas Bauer
with one customer I run a Proxmox 5.4 Ceph cluster with 4 nodes. Two of the nodes have 24 GB of memory and 5 OSDs each, and these nodes are tasked only with one (important) VM each.
On both nodes we see a disturbing pattern:
- after reboot, all the ceph-osd processes start out below 1G of memory usage which would be ok
- then with time, the processes grow up to 4G of memory usage
- some out of memory conditions happen, sometimes only failed backups, sometimes crashes of the VM failing to get all of the allocated (in VM configuration) memory
To recap, servers with 24 GB memory and running only one (1) VM with about 8 GB main memory, can run only stable for about 1-2 weeks until crashes and out-of-memory-situations happen.
Why are the ceph-osd processes growing in their memory usage so much? How to limit their memory usage and how to solve this problem?
Both nodes should be able to take 2 VMs (16 GB) and only require 8 GB for proxmox system and Ceph.
Please help,
Andreas Bauer