Search results

  1. A

    VMs crash on migration to host with downgraded CPU

    We are introducing newer servers to our environment with Xeon Gold 6148 in combination with our E5-2697v2 hosts. So now we have a mix of Ivy Bridge-EP and Skylake-IP. I expected the requirement was to configure the VM's CPU to "IvyBridge" to fix the CPU flags, instruction set extensions, etc to...
  2. A

    Ceph unable to sustain 1 OSD out with size=4

    We have 15 OSD hosts, and 22 OSDs. The servers physically have 2 drive bays. Of course the OSDs are not distributed perfectly evenly. Some servers have 1 OSD and some servers have 2 OSDs, but we are always adding drives to the system as time/availability allows. OSD utilization according to the...
  3. A

    Multiple PVE clusters sharing a Ceph cluster

    Due to Corosync there is clearly a finite and rather small number of nodes that a single Proxmox cluster can support, yet Ceph clusters have no such size limit. The only thing preventing multiple Proxmox clusters going hog-wild on a shared Ceph cluster is the certainty of overlap and collision...
  4. A

    PVE 7.2 and Corosync, maximum hosts per cluster

    We have a 16 node cluster and are preparing to expand to 32 nodes. Because of the hemmed in architecture of the Dell blades and the M1000e modular chassis, we had some interesting choices to make as far as the physical networking. Each server has 6x 10 GbE ports and 8x 1 GbE ports. The 10 GbE...
  5. A

    Is it possible to rename the cluster

    The cluster name appears to be fixed on the cluster tab as well as the Datacenter root object in the sidebar tree. Just curious if it's possible to rename this. I see clearly how the GUI doesn't provide for such a thing but I am open to hacky suggestions as well.
  6. A

    Restrict pvemanager default_views per-user

    With rare exception we give some third party partners a login to our PVE cluster. Of course we restrict their permissions to which resource pools they can see and give them only the bare minimum permissions on their VMs. However, when these restricted users login to the PVE GUI, the default...
  7. A

    nodes eventually lose access to CephFS and have to unmount

    15 node cluster, just upgraded all to 6.4-13. Did not have this problem with previous 6.x PVE. I can always "fix it" by running umount /mnt/pve/CephFS but within an hour or so, at least 4 or 5 of the nodes will lose it again. On the PVE GUI the error is "unable to activate storage 'CephFS' -...
  8. A

    Lost quorum and all nodes were reset during a disjoin!

    Hello, I have a 15-node cluster with about 50 VM/CT's running, all configured for HA and using the software watchdog. It's hyperconverged using Ceph RBD for the VM storage. All nodes run PVE 6.4-13 and Ceph 15.2.9. The hostnames in the cluster are "node36" through "node50." Yesterday, I removed...
  9. A

    pve 6.3 and 7.0 mixed environment?

    I have 15 hosts running 6.3. I want to upgrade them to 7.0 and at the same time switch from a single ext4 root FS to a mirrored ZFS setup, so I want to do a clean install on all the hosts, one at a time, migrating VMs around so they are not affected. Ceph version is 15.2.9 on all nodes. If I...
  10. A

    fdisk and sfdisk missing from pve 7.0 installation iso?

    I'm just pointing this out for others, since I use the 6.x ISOs as well, but it is nice sometimes to be able to boot the ISO into debug mode and have all the usual disk utilities there. On the new ISO, fdisk and sfdisk are missing.... just saying! Was it necessary to remove them?
  11. A

    ZFS rebuild questions

    I'm experimenting with installing Proxmox on a ZFS RAID1 setup, and have successfully tested failing and rebuilding drives. The host remains operational and responsive during the loss of the drive. However, I am noticing the rebuilt drive is missing the "BIOS Boot" and "EFI" partitions. It only...
  12. A

    PVE system mail list

    By default, only the email address assigned to root@pam recieves system alerts about fencing hosts. In my cluster there are 6 other users that are administrators (although they use the pve realm), and their email addresses are configured, but they do not receive email notifications about host...
  13. A

    HA failover issues reconfigures VM after upgrade to 6.3

    A host was fenced today and all of the VMs (all Windows 2016 or 2019) were successfully powered up on other hosts. Only trouble was that none of them were reachable on the network, and they all had services that failed to start upon boot. After investigating, we found that all of the network...
  14. A

    mysterious reboots

    I have had the same node mysteriously reboot twice over the weekend. Once on a saturday morning and once on a sunday morning. It was fenced properly and VMs were brought up elsewhere, and I have roughly the exact time the host rebooted (or reset). How would I go about tracking down the root...
  15. A

    After 6.3 upgrade, VM boot device missing if member of an HA group

    If a virtual machine is not a member of an HA group, you can add a disk, ISO, or network boot device, and expect it will boot to them in the order given on the options screen. If you power on the machine and interrupt POST with the escape key, you can make a one-time selection to manually boot...
  16. A

    Performance decrease after Octopus upgrade

    Just throwing this out there to see if anyone has experienced anything similar. Under Nautilus, our Windows VMs were able to do about 1.5 GB/sec sequential read, and 1.0 GB/sec sequential write. Under Nautilus, our rados bench was showing us 2.0 GB/s sequential read and write, and this was...
  17. A

    Is qemu guest agent required to TRIM thin provisioned storage?

    I have a 15GB pfSense machine that has about 1 GB used on its UFS file system. Back-end is NVMe Ceph RBD. Ceph shows the disk is 15 GB with 14 GB used. There is currently no option for qemu guest agent on pfSense, and I've noticed in the few hours its been running that it has not been TRIMmed...
  18. A

    PVE and Ceph on Mellanox infiniband parts, what is the current state of support?

    We are using PVE and Ceph in Dell blades using the M1000e modular chassis. We are currently using dual mezzanine cards with 2x 10 GbE ports, one for Ceph front-end and one for Ceph back-end. Public LANs, guest LANs, and Corosync are handled by 4x 10GbE cards on 40GbE MXL switches, so all is...
  19. A

    Choose between Samsung PM983 or Intel DC P4510

    These are the 2 drives I'm looking at right now for a 16-32 node PVE6+Ceph RBD setup. Referring to the 2018 performance doc, https://www.proxmox.com/en/downloads/item/proxmox-ve-ceph-benchmark They ran an fio command to represent expected 4KQD1 performance as it pertains to OSD journal...
  20. A

    redirecting intense write activity, Folder2RAM, etc...

    Has anyone ever looked at Folder2RAM: https://github.com/bobafetthotmail/folder2ram I see it more often used in conjunction with OMV but as it's also Debian-based I wanted to mention it here to get your thoughts. Consider a PVE/Ceph setup comprised of blade servers such as Dell M610, M620...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!