Search results

  1. B

    3 node cluster with nodes in two diferent server rooms

    Does anybody have any experience with putting cluster nodes in different server rooms ? I have several buildings and was wondering what is acceptable latency for a cluster to operate without any issues. The buildings are connected via 10Gbps fiber and latency is very low 1-2 ms. What is max...
  2. B

    pvestatd keeps restarting

    I see in that the pvestatd keeps restarting itself. I have 3 servers in 6 server cluster showing that log. See example below: pvestatd[2352]: restarting server after 72 cycles to reduce memory usage Everything seems to be working properly. How do we troubleshoot it or stop it ? Thank you...
  3. B

    Pooling local backup between two PBSs from the same datastore

    I have two locations and backup to PBS in each location. Is it possible to sync between locations only VMs that are local to that PBS but using the same datastore ? In other words sync between locations but only VMs that are not in the other PBS datastore ? Thank you
  4. B

    Using auto scaler for the first time

    I tested the auto scaling option on a test system and now I am using it in production. I have 5 nodes and 30 osd (all ssds), I setup the target size to 80% of the total size of the pool. Ceph shows the pool has 512 pg and that the optimal # of pgs is 1024 , the autoscaler is on , I checked...
  5. B

    Removing 3 nodes from 6 node cluster

    I need to remove 3 nodes from a 6 node cluster and they are not coming back , the remaining 3 nodes will be gradually retiring after moving all the VMs. I have done that before few years ago with pvecm expected command. Is there any other way ? There will be no HA on remaining nodes , is...
  6. B

    Ceph Crush map question

    I had a need at some point to create a secondary HDD pool on our ceph cluster, I added two rules: ceph osd crush rule create-replicated replicated_rule_ssd default host ssd - for SSDs ceph osd crush rule create-replicated replicated_rule_hdd default host hdd - for HDD # rules rule...
  7. B

    question about updates

    I scheduled adding 3 nodes to a working 3 node cluster on 8.3.0. I see today 4 packages popped up in the updated so the new nodes that will join will be fully updated. I usually schedule maintenance to do updated with reboots (when kernel is updated) but I am wondering if these packages can be...
  8. B

    Error while restoring VM, PVE version 8.3.0

    I am getting that message when moving VMs from old PVE cluster to the new PVE cluster using PBS: restore image complete (bytes=34359738368, duration=150.76s, speed=217.35MB/s) can't deactivate LV '/dev/local2-2.4TB-HDD/vm-324056-disk-0': Logical volume local2-2.4TB-HDD/vm-324056-disk-0 in...
  9. B

    Storage migration failed io_uring known to cause issues

    I am getting this error when trying to move drive from Ceph storage to PVE local hard drive: TASK ERROR: storage migration failed: target storage is known to cause issues with aio=io_uring (used by current drive) This started to happen on 8.2.x/8.3.0 , I have a 7.3 and 7.4 cluster and that...
  10. B

    HELP with ver 8.3.0 , downgrade possibly reinstall

    I am putting 2 new clusters in , they are not in production (be ion production next week) and I noticed new packages in the update section for Enterprise Repo. I did the update as I like to keep clusters updated to the newest firmware at least right before going live. After the update I notice...
  11. B

    Proxmox 8.2 with Ceph , time server

    What are you doing for time server for Ceph on Prox 8.2 ? I used to have two dedicated time serer on Ceph under Proxmox 6 otherwise I got clock skew errors. On Proxmox 7.x I tested Ceph with default setting and for last several years I had no problems. Is this the case with 8.2, I see it is...
  12. B

    Ceph cluster with one node having no hard drives for specific pool.

    I have 5 node Ceph cluster run under Proxmox. I am moving VMs from an old PVE/Ceph clusters to a new one PVE/Ceph clusters. During this time my new 5 node Ceph cluster will not have all the drives installed. I will have only 8 SSDs, planning to put 2 per server which will cover 4 servers, the...
  13. B

    Multiple Proxmox clusters on the same subnet

    Just wanted to confirm , on 8.2 version we can have multiple PVE clusters on the same subnet as long as the "Cluster Name" is different , right ? We are adding servers and creating new cluster , we will have 5 PVE cluster on the same subnet for about 2 week, after 2 week we will go down to 3...
  14. B

    Moving pve cluster to a new switch

    I have a cluster in production that I need to move to a new switch. We are using MCLAG redundant interfaces and need to move 6 nodes from on switch cluster to the other. That should not take long but I was wondering if I should execute on each node pvecm expected -1 before I do that, any advice...
  15. B

    Test, nested Ceph cluster SSD latency up to 2000ms

    I configured a nested Proxmox Ceph cluster (3 nodes) for testing on my Proxmox server. 3 nodes, plenty of RAM, CPU power etc. I used 3 good SAS SSDs , 1 per virtual machine. Currently there is nothing else running on this Proxmox serer. All networking works fine , I have 2 x 10Gbps ports and...
  16. B

    PBS backing up relatively big virtual machines

    I am using PBS to backup about 300 systems running on PVE then sync it to a backup location to another PBS. It works GREAT , have been doing it for last year with daily and weekly backups, with no issues. This VMs are relatively small in size, the cumulative size is about 13-14TB. I have a VM...
  17. B

    VNC argument/port question

    I have the VNC working with args: -vnc 0.0.0.0:xx added to the config file. What happens if the VM is migrated to another node with another VM using the same port ? Is there a wider range of ports that can be used instead of 59xx or perhaps extend that range so I can add it to more VMs from...
  18. B

    Strange network behavior on one port in a bond

    proxmox-ve: 7.4-1 ceph: 17.2.6-pve1 I have 4 node Ceph cluster using Proxmox as a host. Public and Private networks are separated using 2 x 10Gbps ports each (2 cards per node , 4 ports total). All nodes are setup in exactly the same way. Here is an example of Ceph Private config: auto...
  19. B

    Ceph PG #

    Proxmox 7.4.16. I am getting confused by all the numbers. I have 24 OSDs , SSD 1.46TB across 4 nodes, 3 replicas , total size of the pool 12TB and it is going to be 80-85% full. I did the calculation from ceph calc and it gets me 800 , rounded 1024 PG which is also the number that Ceph...
  20. B

    PBS restore command syntax

    I am playing with restore command for PBS and I cannot seem to get it right. I have local LVM storage called "local1-SSD" where I am trying to restore. proxmox-backup-client restore --repository username@pbs@1.1.1.1:Datastore vm/1000/2023-08-13T23:00:39Z drive-virtio0.img.fidx...