Search results

  1. A

    Ceph unable to sustain 1 OSD out with size=4

    I hear you, but I've already got the default profile that prioritizes client ops, yet during a recovery multiple slow OSD ops are all that it takes to freeze VMs. It seems like regardless of my choice, I have to set the nobackfill and norecover flags and so on and wait until the evening to...
  2. A

    Ceph unable to sustain 1 OSD out with size=4

    Thanks Aaron. I am considering going back to wpq until the next release. I don't want to get into an open-ended situation of adjusting tunables all the time.
  3. A

    Ceph Quincy - rbd: listing images failed after 17.2.4 to 17.2.5 upgrade

    By default RBD is not even a permitted application on the .mgr pool. You definitely want to create an RBD-dedicated pool.
  4. A

    Ceph unable to sustain 1 OSD out with size=4

    Tonight, rebooting a single-OSD host with the OSD flags set went better. There was a brief appearance of slow OSD ops but they were completed within a few seconds. I fear how Quincy would respond to an unplanned failure. One would need to react quickly to disable recover, balance, and backfill...
  5. A

    Ceph Quincy - rbd: listing images failed after 17.2.4 to 17.2.5 upgrade

    I am beyond confused, you are using the .mgr pool for RBD? In my experience disks are expressed by the path poolName/vm-id-disk-0 etc
  6. A

    Ceph unable to sustain 1 OSD out with size=4

    This sounds closest to what we are experiencing. What else do you know about it, and what else can be done besides using the norebalance flag? To me, the 17.2.4 changelog suggests that these things have been fixed, when they have not.
  7. A

    Ceph unable to sustain 1 OSD out with size=4

    With regards to the ceph status, don't worry about the 1 mon being down. They are on comparatively slower storage and spend a lot of time on get_health_metrics. That is one reason we have 7 mons, they are all active and running, but they come and go when they get bogged down with stats but we...
  8. A

    Ceph unable to sustain 1 OSD out with size=4

    root@virtual41:~# pveversion -v proxmox-ve: 7.2-1 (running kernel: 5.15.53-1-pve) pve-manager: 7.2-11 (running version: 7.2-11/b76d3178) pve-kernel-helper: 7.2-12 pve-kernel-5.15: 7.2-10 pve-kernel-5.4: 6.4-18 pve-kernel-5.15.53-1-pve: 5.15.53-1 pve-kernel-5.15.39-1-pve: 5.15.39-1...
  9. A

    Ceph unable to sustain 1 OSD out with size=4

    We have 15 OSD hosts, and 22 OSDs. The servers physically have 2 drive bays. Of course the OSDs are not distributed perfectly evenly. Some servers have 1 OSD and some servers have 2 OSDs, but we are always adding drives to the system as time/availability allows. OSD utilization according to the...
  10. A

    Single ring failure causes cluster reboot? (AKA: We hates the fencing my precious.. we hates it..)

    just more testing. read the corosync redundancy part about setting ring priorities https://pve.proxmox.com/pve-docs/pve-admin-guide.html#pvecm_redundancy
  11. A

    Single ring failure causes cluster reboot? (AKA: We hates the fencing my precious.. we hates it..)

    I had a couple Windows VMs get irreparably corrupt after a spontaneous reboot of all my nodes. Also using Ceph. We ended up disabling HA entirely because of it. If this is merely a matter of taking some switches offline, and you are already familiar with editing corosync.conf, then you can...
  12. A

    Terminal Server Windows Server 2019/2022 on Proxmox Virtualiziation

    PVE is excellent for hosting just about any Windows use case.
  13. A

    PVE 7.2 and Corosync, maximum hosts per cluster

    The pvecm section hasn't changed since I first read it over 4 years ago. There are still notes in there relating to PVE 3.x and 4.x. I was hoping for some real world accounts from admins of large clusters but it seems large clusters are rare.
  14. A

    Multiple PVE clusters sharing a Ceph cluster

    Wow I can't believe I missed that, I went through all the datacenter tabs. A cluster prefix would be nice too, and cleaner looking, but this certainly gets the job done.
  15. A

    Multiple PVE clusters sharing a Ceph cluster

    Due to Corosync there is clearly a finite and rather small number of nodes that a single Proxmox cluster can support, yet Ceph clusters have no such size limit. The only thing preventing multiple Proxmox clusters going hog-wild on a shared Ceph cluster is the certainty of overlap and collision...
  16. A

    PVE 7.2 and Corosync, maximum hosts per cluster

    We have a 16 node cluster and are preparing to expand to 32 nodes. Because of the hemmed in architecture of the Dell blades and the M1000e modular chassis, we had some interesting choices to make as far as the physical networking. Each server has 6x 10 GbE ports and 8x 1 GbE ports. The 10 GbE...
  17. A

    Is it possible to rename the cluster

    The cluster name appears to be fixed on the cluster tab as well as the Datacenter root object in the sidebar tree. Just curious if it's possible to rename this. I see clearly how the GUI doesn't provide for such a thing but I am open to hacky suggestions as well.
  18. A

    Windows VMs stuck on boot after Proxmox Upgrade to 7.0

    FYI this issue came up with me on a 2016 VM, Proxmox 7.2-7, and Qemu 6.2.0-11, and adding args: -machine smm=off to the VM config had no effect. Will try the downgrade later on, this is my only environment and it's production so unfortunately I can't afford to play with it.
  19. A

    Restrict pvemanager default_views per-user

    Thanks for responding. Of course it is not a true security measure, it was just meant to control the perspective of our users and provide no more than the data pertinent to them. These people can only control the power on/off/reset, console, and snapshots, nothing to do with any particular...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!