Search results

  1. B

    CEPH shared SSD for DB/WAL?

    You can make ceph automatically handle this, it nicely create partitions accordingly.
  2. B

    CEPH shared SSD for DB/WAL?

    A common recommendation is to have 1 SSD per 4-6 HDDs since it can speedup writes dramatically. If I understand correctly, it is indeed so that if the SSD would fail, all HDDs it holds data for will need rebuilding. In CEPH docs, I've seen 4% recommended, I think the recommended minimum is...
  3. B

    Benchmark: 3 node AMD EPYC 7742 64-Core, 512G RAM, 3x3 6,4TB Micron 9300 MAX NVMe

    It looks like the amount of OSDs per NVMe does not influence the results too much then? I'm looking to run similar drives at 1 OSD per NVMe to save CPU power (64C/128T for 20-24 drives).
  4. B

    Proxmox VE Ceph Benchmark 2020/09 - hyper-converged with NVMe

    We'd put 24 drives in a chassis with 128 CPU threads, giving a ratio of 5.33 CPU threads per NVMe. I've read before that 4 threads per NVMe is the recommended minimum, with 8+ seeming standard practice. It looks like we'll run in to a CPU bottleneck as far as I can find, however it would be not...
  5. B

    Proxmox VE Ceph Benchmark 2020/09 - hyper-converged with NVMe

    Does this depend on the type of drive or the CPU architecture or something else? It would be super helpful to see CPU load during benchmarking, I'm considering using a 64C/128T single socket for 24 U.2 drives, at 5.3 threads per drive, would it be able to keep up?
  6. B

    CEPH cluster planning

    Thanks for your extended reply, greatly appreciated! When going over 50-60% we intend to start adding drives/servers. Sequential write of a single disk would be around ~200-250MB/s, our drives (HC550) have fairly low IOPS/TB so I would think the performance will suffer. I've read a case where...
  7. B

    CEPH cluster planning

    Hi people! I'm planning a CEPH cluster which will go in production at some point but first will serve as a testing setup. We need 125TB usable storage initially, with a cap of about 2PB. The cluster will feed 10 intensive users initially, up to 100 later on. The loads are generally read heavy...
  8. B

    Delete old disks

    I'd like to clear my drive from old leftovers which remained when updating my install. There's several drives/partitions listed like this: Disk /dev/mapper/pve--OLD--C576DBC9-vm--104--disk--0: 200 GiB, 214748364800 bytes, 419430400 sectors Units: sectors of 1 * 512 = 512 bytes Sector size...
  9. B

    Hugepage allocation failed

    That's what I suspected, I also found it difficult to find the correct words to describe it. :D No idea. It appears it doesn't handle Numa nodes without memory properly. This thread seems to be abandoned/ignored though so I suggest starting a new thread. My problem with the normal settings...
  10. B

    Hugepage allocation failed

    memory: 102400 cores: 32 numa0: cpus=0-7;32-39,hostnodes=0,memory=51200,policy=bind numa2: cpus=8-15;40-47,hostnodes=2,memory=51200,policy=bind This fails due to what I explained earlier: You give your VM 32 cores, you can't select/pin the cores of the host, you can only choose from what...
  11. B

    Hugepage allocation failed

    I think you made a mistake here: numa0: cpus=0;1;2;3;4;5;6;7;32;33;34;35;36;37;38;39,hostnodes=0,memory=65536,policy=preferred The cpu's listed after CPU should be numbered sequentially with the core id from the VM, not the host. In other words, you should define that as: numa0...
  12. B

    Hugepage allocation failed

    Bump Problem still exists in latest 5.4-5
  13. B

    Hugepage allocation failed

    Is there something I can do so I don't need to set every VM manually? It makes automated deployment impossible, some kind of workaround would be convenient.
  14. B

    Hugepage allocation failed

    Another issue found with my configuration: When starting 4 VM's taking 30000MB from every numa (machine has 32GB of ram per numa node on an EPYC system), all seems to work fine. When I start the 4VM's with hugepages enabled it fails every time at some point. This snapshot is taken when the 3rd...
  15. B

    Hugepage allocation failed

    This is how it looks when starting 4 VM's with 20GB RAM each: root@ProxS01:~# numactl -H available: 4 nodes (0-3) node 0 cpus: 0 1 2 3 4 5 24 25 26 27 28 29 node 0 size: 32140 MB node 0 free: 11273 MB node 1 cpus: 6 7 8 9 10 11 30 31 32 33 34 35 node 1 size: 32232 MB node 1 free: 31516 MB node 2...
  16. B

    Hugepage allocation failed

    I believe I found a bug in how the memory is allocated. When running numactl -H before running any VM I get this: root@ProxS01:~# numactl -H available: 4 nodes (0-3) node 0 cpus: 0 1 2 3 4 5 24 25 26 27 28 29 node 0 size: 32140 MB node 0 free: 31248 MB node 1 cpus: 6 7 8 9 10 11 30 31 32 33 34...
  17. B

    Hugepage allocation failed

    I'm thinking this is not related to hugespages, I can start all VM's when giving them 8 vCPU's per machine and 20GB of RAM with NUMA=0, but I can't start at least one when enabling NUMA. It's a different error when not using hugepages but the same issue I think. The error I get instead of the...
  18. B

    Hugepage allocation failed

    I have a machine with 128GB of RAM with an EPYC 7401P (24C 48T) and want to run 4 VM's. Initially I had issues setting up the first machine and giving it varying amounts of RAM, but that seemed fixed when using hugepages and configuring more sockets. Setup of the template went well but now I hit...
  19. B

    (DNS) routing questions

    I would like to make some of my VM's accessible remotely. When I rented VPS's in the past i got access via an address like vps1058.178.133.22.68 (random example). Given I only have 1 IPv4 atm, how could I do something similar? For example machine 1 would be reachable with SSH via...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!