Search results

  1. R

    Running snapshot backup stalls VM - Backup throughput is low

    Hi Dominik, so this is the output directly on the pbs01 This is from a local PVE cluster connected via 10G. On I have not yet experienced problems like mentioned above on it. And this is the output of a node of the 5-node Cluster. As mentioned the traffic goes via "The Internet". This...
  2. R

    [SOLVED] PVE ZFS mirror installation without 512MByte Partition - how to convert to UEFI boot?

    Hi, I have a older installation of a 3-node Proxmox Ceph cluster - probably 5.3 or older. The OS disks look like this: So there is no 512M partition I could use for the EFI boot partition. I would like to switch to a UEFI boot, but how do I cut out that 512M UEFI partition?
  3. R

    Running snapshot backup stalls VM - Backup throughput is low

    Hi, we are trying to replace out current Backup solution with PBS. So I installed a PBS on an older ThomasKrenn 4HE Intel Dual-CPU RI2436 with some disks... One Proxmox Cluster is already running backups into this pbs01 for some time without problems. Now I try to run backups with our...
  4. R

    nfs share from lxc

    Might want to have a look here: https://forum.proxmox.com/threads/ha-nfs-service-for-kvm-vms-on-a-proxmox-cluster-with-ceph.80967/post-365321 But I can not recommend to run a nfs-kernel-server within a LXC container on Proxmox - as soon as the nfs-kernel-server hangs it might also hang your...
  5. R

    HA NFS service for KVM VMs on a Proxmox Cluster with Ceph

    I can not recommend using nfs-kernel-server with a cephfs kernel client when using CephFS snaphots. As soon as the MDS service fails over from active to a standby MDS the NFS clients die with kernel panic or services running on them just die. It seems to be a caching problem. This problem...
  6. R

    nfs error in lxc

    nfs-ganesha runs in user space, so if it hangs or crashes in the CT (LXC) container it will not take the Proxmox Host down with it. If the nfs-kernel-server in a CT container has problems it might hang the Proxmox host. It is better to run a nfs-kernel-server within a Proxmox VM (KVM).
  7. R

    Proxmox on Ceph: After PG_NUM increase high read io on NVMEs - VMs unusable

    Deep scrubbing is not happening at all at the moment - and probably not untill this is finished... () 9 pgs not deep-scrubbed in time pg 4.76f not deep-scrubbed since 2021-01-31 13:12:33.368818 pg 4.72f not deep-scrubbed since 2021-01-31 16:48:10.470599 pg 4.649 not deep-scrubbed since...
  8. R

    Proxmox on Ceph: After PG_NUM increase high read io on NVMEs - VMs unusable

    One of the six remaining PGs finished today... and again two OSDs hat 100% Read IO usage as soon as the PG finished.
  9. R

    Proxmox on Ceph: After PG_NUM increase high read io on NVMEs - VMs unusable

    Yes, i can check them, but what should I be looking for??? root@proxmox07:~# ceph pg ls | head -n1 PG OBJECTS DEGRADED MISPLACED UNFOUND BYTES OMAP_BYTES* OMAP_KEYS* LOG STATE SINCE VERSION REPORTED UP ACTING SCRUB_STAMP...
  10. R

    Proxmox on Ceph: After PG_NUM increase high read io on NVMEs - VMs unusable

    No, the pg_num and pgp_num are now at 128 and I changed the autoscaler to "warn" for these pools. root@proxmox07:~# ceph osd pool ls detail pool 2 'cephfs_data' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 autoscale_mode warn last_change 15199 lfor...
  11. R

    Proxmox on Ceph: After PG_NUM increase high read io on NVMEs - VMs unusable

    Adjusted settings after reading this. root@proxmox07:~# ceph config show-with-defaults osd.0 | grep 'osd_max_backfills\|osd_recovery_max_active\|osd_recovery_sleep' osd_max_backfills 3 mon osd_recovery_max_active...
  12. R

    Proxmox on Ceph: After PG_NUM increase high read io on NVMEs - VMs unusable

    So, the monday was horrible, our customers started to hate us again... Carefully(!!!) restarting the OSD processes and using the recoveries to keep the backfilling from starting up got us over the day, and late in the evening we found the bluefs_buffered_io setting. So that shifted the read IO...
  13. R

    Proxmox on Ceph: After PG_NUM increase high read io on NVMEs - VMs unusable

    Hi, the autoscaler increased the number of PGs on our Ceph storage (Hardware like this but 5 nodes). As soon as the backfill starts the VMs become unusable and we startet killing OSD processes that cause high read io load. So as in this picture we would kill the ceph-osd process working on...
  14. R

    PBS server full: two days later almost empty?!?!?!?

    So, here comes the next uncertainity from my side: I generally disable atime on ZFS pools and this becomes inherited on the volumes then. If you rely on atime does this still work?
  15. R

    PBS server full: two days later almost empty?!?!?!?

    So I let the log output continue to run and after about 2 hours and about one GByte I stopped it again. The content is just about the same for the whole file.
  16. R

    PBS server full: two days later almost empty?!?!?!?

    Here are parts of the log: starting garbage collection on store pve-prod task triggered by schedule '15:40' Start GC phase1 (mark used chunks) marked 1% (17 of 1651 index files) WARN: warning: unable to access chunk 2cd5a53b5d8aa3c9d530c2ee2b89ccf7ed238ad7bf97afb1a3424666784656d0, required by...
  17. R

    PBS server full: two days later almost empty?!?!?!?

    Still venting air here... Underlying storage is a ZFS pool, which looks ok: There are so many warnings that the Browser stops working, when I tried to have a look at them: Mount looks ok: root@pbs01:~# mount | grep pbs rpool/ROOT/pbs-1 on / type zfs (rw,relatime,xattr,noacl) backup/pbs on...
  18. R

    PBS server full: two days later almost empty?!?!?!?

    No idea how that happened! But 206GByte ist definately not the size of my source five node PVE cluster. Any idea how that could happen? So on 27.01.2021 it went full and two days later it is almost empty... I am pretty shocked right now... Rainer
  19. R

    HA NFS service for KVM VMs on a Proxmox Cluster with Ceph

    But does not support properly exporting .snap directories... At least a bug report was created for this missing feature on NFS-Ganesha : https://tracker.ceph.com/issues/48991

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!