Search results

  1. B

    corosync entries in journalctl link: 0 is down etc.

    We had a node 5 in a 6 node cluster fenced due to excessive ram ecc errors. HA worked great and all vms started on other nodes. The cluster worked with no corosync issues for last year since it was put it in production (we had ecc errors in February but ram was replaced). I should mention that...
  2. B

    3 node cluster with nodes in two diferent server rooms

    Does anybody have any experience with putting cluster nodes in different server rooms ? I have several buildings and was wondering what is acceptable latency for a cluster to operate without any issues. The buildings are connected via 10Gbps fiber and latency is very low 1-2 ms. What is max...
  3. B

    pvestatd keeps restarting

    I see in that the pvestatd keeps restarting itself. I have 3 servers in 6 server cluster showing that log. See example below: pvestatd[2352]: restarting server after 72 cycles to reduce memory usage Everything seems to be working properly. How do we troubleshoot it or stop it ? Thank you...
  4. B

    Pooling local backup between two PBSs from the same datastore

    I have two locations and backup to PBS in each location. Is it possible to sync between locations only VMs that are local to that PBS but using the same datastore ? In other words sync between locations but only VMs that are not in the other PBS datastore ? Thank you
  5. B

    Using auto scaler for the first time

    I tested the auto scaling option on a test system and now I am using it in production. I have 5 nodes and 30 osd (all ssds), I setup the target size to 80% of the total size of the pool. Ceph shows the pool has 512 pg and that the optimal # of pgs is 1024 , the autoscaler is on , I checked...
  6. B

    Removing 3 nodes from 6 node cluster

    I need to remove 3 nodes from a 6 node cluster and they are not coming back , the remaining 3 nodes will be gradually retiring after moving all the VMs. I have done that before few years ago with pvecm expected command. Is there any other way ? There will be no HA on remaining nodes , is...
  7. B

    Ceph Crush map question

    I had a need at some point to create a secondary HDD pool on our ceph cluster, I added two rules: ceph osd crush rule create-replicated replicated_rule_ssd default host ssd - for SSDs ceph osd crush rule create-replicated replicated_rule_hdd default host hdd - for HDD # rules rule...
  8. B

    question about updates

    I scheduled adding 3 nodes to a working 3 node cluster on 8.3.0. I see today 4 packages popped up in the updated so the new nodes that will join will be fully updated. I usually schedule maintenance to do updated with reboots (when kernel is updated) but I am wondering if these packages can be...
  9. B

    Error while restoring VM, PVE version 8.3.0

    I am getting that message when moving VMs from old PVE cluster to the new PVE cluster using PBS: restore image complete (bytes=34359738368, duration=150.76s, speed=217.35MB/s) can't deactivate LV '/dev/local2-2.4TB-HDD/vm-324056-disk-0': Logical volume local2-2.4TB-HDD/vm-324056-disk-0 in...
  10. B

    Storage migration failed io_uring known to cause issues

    I am getting this error when trying to move drive from Ceph storage to PVE local hard drive: TASK ERROR: storage migration failed: target storage is known to cause issues with aio=io_uring (used by current drive) This started to happen on 8.2.x/8.3.0 , I have a 7.3 and 7.4 cluster and that...
  11. B

    HELP with ver 8.3.0 , downgrade possibly reinstall

    I am putting 2 new clusters in , they are not in production (be ion production next week) and I noticed new packages in the update section for Enterprise Repo. I did the update as I like to keep clusters updated to the newest firmware at least right before going live. After the update I notice...
  12. B

    Proxmox 8.2 with Ceph , time server

    What are you doing for time server for Ceph on Prox 8.2 ? I used to have two dedicated time serer on Ceph under Proxmox 6 otherwise I got clock skew errors. On Proxmox 7.x I tested Ceph with default setting and for last several years I had no problems. Is this the case with 8.2, I see it is...
  13. B

    Ceph cluster with one node having no hard drives for specific pool.

    I have 5 node Ceph cluster run under Proxmox. I am moving VMs from an old PVE/Ceph clusters to a new one PVE/Ceph clusters. During this time my new 5 node Ceph cluster will not have all the drives installed. I will have only 8 SSDs, planning to put 2 per server which will cover 4 servers, the...
  14. B

    Multiple Proxmox clusters on the same subnet

    Just wanted to confirm , on 8.2 version we can have multiple PVE clusters on the same subnet as long as the "Cluster Name" is different , right ? We are adding servers and creating new cluster , we will have 5 PVE cluster on the same subnet for about 2 week, after 2 week we will go down to 3...
  15. B

    Moving pve cluster to a new switch

    I have a cluster in production that I need to move to a new switch. We are using MCLAG redundant interfaces and need to move 6 nodes from on switch cluster to the other. That should not take long but I was wondering if I should execute on each node pvecm expected -1 before I do that, any advice...
  16. B

    Test, nested Ceph cluster SSD latency up to 2000ms

    I configured a nested Proxmox Ceph cluster (3 nodes) for testing on my Proxmox server. 3 nodes, plenty of RAM, CPU power etc. I used 3 good SAS SSDs , 1 per virtual machine. Currently there is nothing else running on this Proxmox serer. All networking works fine , I have 2 x 10Gbps ports and...
  17. B

    PBS backing up relatively big virtual machines

    I am using PBS to backup about 300 systems running on PVE then sync it to a backup location to another PBS. It works GREAT , have been doing it for last year with daily and weekly backups, with no issues. This VMs are relatively small in size, the cumulative size is about 13-14TB. I have a VM...
  18. B

    VNC argument/port question

    I have the VNC working with args: -vnc 0.0.0.0:xx added to the config file. What happens if the VM is migrated to another node with another VM using the same port ? Is there a wider range of ports that can be used instead of 59xx or perhaps extend that range so I can add it to more VMs from...
  19. B

    Strange network behavior on one port in a bond

    proxmox-ve: 7.4-1 ceph: 17.2.6-pve1 I have 4 node Ceph cluster using Proxmox as a host. Public and Private networks are separated using 2 x 10Gbps ports each (2 cards per node , 4 ports total). All nodes are setup in exactly the same way. Here is an example of Ceph Private config: auto...
  20. B

    Ceph PG #

    Proxmox 7.4.16. I am getting confused by all the numbers. I have 24 OSDs , SSD 1.46TB across 4 nodes, 3 replicas , total size of the pool 12TB and it is going to be 80-85% full. I did the calculation from ceph calc and it gets me 800 , rounded 1024 PG which is also the number that Ceph...