Search results

  1. VictorSTS

    [SOLVED] CEPH OSDs Full, Unbalanced PGs, and Rebalancing Issues in Proxmox VE 8

    A lot of things here... You are using 6 monitors, which isn't supported. Use either 3 or 5 (preferred as it would allow 2 mons to fail and still keep quorum). You have a 3/2 pool set to drive class "hdd", but only have 2 servers with "hdd" drives. This is the main origin of your problem. All...
  2. VictorSTS

    [SOLVED] CEPH OSDs Full, Unbalanced PGs, and Rebalancing Issues in Proxmox VE 8

    Need the ouput requested to be sure. That was helpful in that case due to using just 3 hosts. With 6 it will probably help, but does not completely avoid the main issue and over time unbalanced distribution will arise, not to mention how unsafe is to use just 2 replicas ;)
  3. VictorSTS

    [SOLVED] CEPH OSDs Full, Unbalanced PGs, and Rebalancing Issues in Proxmox VE 8

    Checking this again, with the already provided data, pretty sure you are using crush rule(s) that do not use device class and mixing 1T with 10T drives in the same pools with so few PGs will cause such imbalace.
  4. VictorSTS

    [SOLVED] CEPH OSDs Full, Unbalanced PGs, and Rebalancing Issues in Proxmox VE 8

    Would need the output of: ceph osd pool ls detail ceph osd crush rule dump ceph pg dump (this last might get huge!)
  5. VictorSTS

    Minimum permissions for users for file restore

    Be careful: that would allow a user to download any file from the VM, including sensitive information that normally is protected by the OS and it's permissions, not to mention the ability to see the retention policy of backups which can be used in an ATP attack to increase the potential damage.
  6. VictorSTS

    Adding a New Server to Existing Proxmox Cluster - Network Configuration and VM Communication

    Nothing personal at all and wasn't meant to be rude with the OP in any way. Sorry if it looks like that. Thanks for pointing it out. I've been there too: tasked with a job that could not do because I lacked knowledge/experience/practice. What OP is asking to do isn't exactly dead simple and...
  7. VictorSTS

    High CPU

    I would: Try ticking NUMA in the VM CPU settings. Install updated microcode [1]. I don't see any mention about your PVE version, running kernel, etc, so another option could be to use an updated kernel or even a previous one (maybe after the crash it booted with a different kernel? Check...
  8. VictorSTS

    Adding a New Server to Existing Proxmox Cluster - Network Configuration and VM Communication

    You could read the documentation so you can give your CEO an informed opinion about what should be done and how to properly deploy Proxmox software. Again, if in doubt, refer your CEO to a partner or ask for official training [1]. A very personal opinion: If your CEO is a wise one, should hire...
  9. VictorSTS

    High CPU

    The snapshot you posted seems to indicated that balloon was off and/or the balloon service wasn't running correctly. That may not be related to the issue, just got my attention the high mem usage shown. Change CPU type to host and try again. Also try with CPU type "x86-64-v2-AES". Why such a...
  10. VictorSTS

    Adding a New Server to Existing Proxmox Cluster - Network Configuration and VM Communication

    No, a PVE node can't be on different providers. As stated in the link I sent you before [1], nodes must be at max 10ms from each other, IME being under 5ms is recommended. Maybe you could use professional support on this and/or training, please check the official Partners list [2] [1]...
  11. VictorSTS

    High CPU

    Do you have qemu-guest-tools enabled in the VM config, installed and running in the guest OS? Do you have the balloon driver installed and running in the guest OS? The former is needed for QEMU to give accurate memory usage information to PVE with Windows OS and the later for QEMU to be able to...
  12. VictorSTS

    Rebooting one node causes other node to reboot

    Supermicro Twin? Ceph wise I would create a custom crush map to define a "chasis" level and ensure that a single chasis will not hold two copies. Replace "migh" with "will" for around one third of your PGs. It's a fact, not a chance ;). The fourth copy ensures that you have to lose at least 4...
  13. VictorSTS

    Rebooting one node causes other node to reboot

    Are you planning on shutting down 3 of your servers often or are you simply planning for a "just in case 3 servers break"? On the first case, use a qdevice somewhere (PBS is fine, although I also install PVE in my PBS servers and prefer to use QDevice in a VM instead of installing in bare...
  14. VictorSTS

    VM was silently skipped for backup with: config locked (backup)

    When you backup a VM, it's config is locked so no change can be made until backup is done. When a backup task doesn't end properly (user stops it, destination full and the likes), PVE does not clean the status properly (yet) and the VM stays in "locked" state. Use qm unlock <VMID> to remove the...
  15. VictorSTS

    No HA?

    You don't mention it, but you have to add to HA every VM you want to be in high availability. If you did, and you did issue a "reboot" of the PVE host, remember that there is a setting in Datacenter -> Options called "HA settings" where you can set the "Shutdown policy". By default it will...
  16. VictorSTS

    dirty-bitmap status: existing bitmap was invalid and has been cleared - every.single.time!

    IMHO it's better to use PBS sync to copy backups from one PBS to another instead of doing backup of the VM multiple times, even if PVE could use multiple dirty-maps for a few PBS. Less load on source PVE, less load on VMs and you offload the responsibility of doing the second backup replica to...
  17. VictorSTS

    pve-cluster service wont start

    Would be nice if you provided that little detail ;) For reference, there is a long thread regarding host name change with different procedures, some for older versions [1] You probably had some configurations in the host too. Some combinations of changes end up with a database error like the...
  18. VictorSTS

    pve-cluster service wont start

    Have you changed the host's name? What's in your /etc/hostname and /etc/hosts ?
  19. VictorSTS

    LVM-Thin can be used in multiple-node clusters?

    No, you can't use LVM-Thin with a shared LUN, as LVM-Thin uses metadata for things like snapshots and thin provision. If you configure your shared LUN with LVM-thin you will end up with data corruption sooner than later. If your storage supports it, you may enable some kind of compression/thin...
  20. VictorSTS

    Sync

    Thats the functionality needed for all this to work :). Previously, even if you set a datastore offline, PBS service still kept some open files in it and you could not unmount it unless you stopped PBS services, which at the very least is a PITA and in most cases simply not possible as that...