Search results

  1. B

    3 node cluster with nodes in two diferent server rooms

    Thank you for your advice. IINerd , it is going to be a similar setup but with replication on Proxmox.
  2. B

    3 node cluster with nodes in two diferent server rooms

    Does anybody have any experience with putting cluster nodes in different server rooms ? I have several buildings and was wondering what is acceptable latency for a cluster to operate without any issues. The buildings are connected via 10Gbps fiber and latency is very low 1-2 ms. What is max...
  3. B

    pvestatd keeps restarting

    I see in that the pvestatd keeps restarting itself. I have 3 servers in 6 server cluster showing that log. See example below: pvestatd[2352]: restarting server after 72 cycles to reduce memory usage Everything seems to be working properly. How do we troubleshoot it or stop it ? Thank you...
  4. B

    Pooling local backup between two PBSs from the same datastore

    The scenario I have is with 2 Proxmox Backup servers doing sync between each other. There are no naming conflicts. In one location there is one naming convention in the other there is different one. When I pull the backups I just want to backup the the VM that are not on the other side. I just...
  5. B

    Pooling local backup between two PBSs from the same datastore

    I have two locations and backup to PBS in each location. Is it possible to sync between locations only VMs that are local to that PBS but using the same datastore ? In other words sync between locations but only VMs that are not in the other PBS datastore ? Thank you
  6. B

    Using auto scaler for the first time

    So I set it back to "on" from "warn" and now the warning of "too many pgs per old" disappeared with Health status being OK. The number of pgs is still showing 2048. Is this because of what you wrote in the first post: "The autoscaler will only start changing the pg_num automatically if the...
  7. B

    Using auto scaler for the first time

    Thank you again. I did not think about that: If you have a replicated pool with a size of 3, you will have pg_num * 3 replicas in total. 30 OSDs and with ~100 PGs per OSD -> 30 * 100 / 3 = 1000 -> 1024 , that makes more sense now. In other words I forgot about the * 3 replicas. fortunately...
  8. B

    Using auto scaler for the first time

    after exhaustive reading about pgs and how to calculate them , I decided to turn off the autoscaler (mainly because it can start rebalancing during the business hours) , I set the pgs to 2048 (was tempted to use 4096 per the 100pgs per ssd) . I have enough resources CPU/RAM to handle the OSDs...
  9. B

    Using auto scaler for the first time

    Thank you Aaron this is all good advice. Should the target ration be increased gradually , lets say from 0.0 now to 0.2 , then 0.5 then 0.7etc. up to 1. I assume that 1 is the final ratio in my case as this is going to be only one pool in this cluster , is that a correct assumption? Thank you
  10. B

    Using auto scaler for the first time

    I have 1 pool and will have just one pool in this cluster. 5 node cluster, 30 osds (1.6TB drives). I might add within 2 years one node with 6 additional osds but it is not 100%. I will have about 75-80% of the cluster full. The cluster I just decommissioned had pgs assigned statically with 24...
  11. B

    Using auto scaler for the first time

    so i should turn it off and set it up manually to 1024 ? With 27% should I do this gradually and increase by 128 (or perhaps 64) and wait , then increase by another 64 or 128 ? Thank you
  12. B

    Using auto scaler for the first time

    I tested the auto scaling option on a test system and now I am using it in production. I have 5 nodes and 30 osd (all ssds), I setup the target size to 80% of the total size of the pool. Ceph shows the pool has 512 pg and that the optimal # of pgs is 1024 , the autoscaler is on , I checked...
  13. B

    Removing 3 nodes from 6 node cluster

    I did pvecm expected 3 which gave me quorum. I guess what I did not realize was that the expected command with x nodes is available only after you remove in my case it was the 3rd node. Before I was getting an error. Once the 3rd node was removed (in 6 node cluster) pvecm expected 3 command did...
  14. B

    Removing 3 nodes from 6 node cluster

    I need to remove 3 nodes from a 6 node cluster and they are not coming back , the remaining 3 nodes will be gradually retiring after moving all the VMs. I have done that before few years ago with pvecm expected command. Is there any other way ? There will be no HA on remaining nodes , is...
  15. B

    Ceph Crush map question

    I had a need at some point to create a secondary HDD pool on our ceph cluster, I added two rules: ceph osd crush rule create-replicated replicated_rule_ssd default host ssd - for SSDs ceph osd crush rule create-replicated replicated_rule_hdd default host hdd - for HDD # rules rule...
  16. B

    question about updates

    I scheduled adding 3 nodes to a working 3 node cluster on 8.3.0. I see today 4 packages popped up in the updated so the new nodes that will join will be fully updated. I usually schedule maintenance to do updated with reboots (when kernel is updated) but I am wondering if these packages can be...
  17. B

    Storage migration failed io_uring known to cause issues

    Thx , for your advice I changed from LVM to directory. The local storage is only used as a backup/maintenance for Ceph.
  18. B

    Error while restoring VM, PVE version 8.3.0

    Thank you, it was a one-time thing. Not happening anymore. I did change the disk from LVM to Direcorty tho.
  19. B

    Storage migration failed io_uring known to cause issues

    Thank you for your response, I think I should start with it but is the io_uring option best/optimal for Ceph and other storage, in general ? BTW now it makes sense our other cluster uses directory for local hard drive storage and the new one is using LVM. Depending on your answer with the...