Search results

  1. K

    All resources in HA

    Hi We have been doing some testing on HA with PVE and Ceph, which has given very possitive results in terms of our goals with exploring the setup. From our testing, we have found two issues that we had not really expected 1. When a host has failed and comes back online, VMs are not migrated...
  2. K

    [SOLVED] HA with ZFS

    Thank you for the replies We will be looking in to how we can use Ceph. Some initial testing shows that it gives the necessary HA, so we just need to figure out the monitoring of disks
  3. K

    [SOLVED] HA with ZFS

    Thank you for the reply Are there any recommended tools that can help monitor disk health when Ceph has control over the disks? With zfs-zed, our system-administrators get a notification when a disk failure is detected
  4. K

    [SOLVED] HA with ZFS

    Hi We are looking into hardening our infrastructure to better handle outages due to network or hardware failtures. We use OVH dedicated physical servers for our infrastructure and currently have 3 node PVE cluster. This cluster is currently configured with 2x4TB nVME drives per node, with one of...
  5. K

    Zabbix as VM looses connection

    Hi For some time now, we have had a strange issue with the VM running our Zabbix monitoring server. Zabbix runs as a VM on our PVE cluster and monitors all running virtual machines on the cluster. Zabbix VM is running CentOS 7 as the guest OS. Whenever a backup of this VM is triggered, it seems...
  6. K

    Replace disk on OVH PVE installation

    Hi For a few years, we have been running PVE on OVH dedicated servers in Europe and changed servers as we outgrew the ones we had. At this point we have their HG-servers, where components can be changed from month to month. In preparation for a drive replacement on our servers (larger capacity...
  7. K

    Cluster unable to add new node from Canada

    So the high latency between Canada and Europe could be the limiting factor? We have tried to manually copy corosyn.conf, which got the cluster to start and show up on the other nodes, but it shows up as having no connection (the red circle with a cross over it and the grayed out node name) If...
  8. K

    Cluster unable to add new node from Canada

    Hi again Some more, maybe relevant information. All pre-existing nodes are located in the OVH datacenters across Europe, while the new node is located in Canada. Latency of a ping is between 150-300ms. But would this make it impossible to have it join the cluster?
  9. K

    Cluster unable to add new node from Canada

    Hi Today we provisioned a OVH dedicated server as a new node for our PVE cluster. Upon adding the node from the UI and setting the ring0 address/link0 address, we could not get the node to join. We tried to figure out why and finally got it to join. However, there is a strange difference between...
  10. K

    VZDump locks VM

    VMs are running CentOS 7 with cPanel on top. But we have 3 other servers with the exact same configuration and they do not have the issue. We do not have remote syslog configured, but if it could help in the given situation, I might be able to configure it.
  11. K

    VZDump locks VM

    As mentioned previously, the VM locks completely and no messages show at all. If we try to log in using the console, we can type the username and after that there is just a blinking cursor/underscore and nothing happens. All services running on the server can no longer be accessed once the VM...
  12. K

    VZDump locks VM

    I have done migration to a different host in the cluster and still the same result. I have quite a few similar sized VMs on both hosts and when I start the backup of those, then it stops at the "creating archive" step for at most a few seconds and then it continues
  13. K

    VZDump locks VM

    Hi I am experiencing a strange issue with 2 VMs on our 4 node proxmox cluster. The 2 VMs are on the same host. When either scheduled or manual VZDump starts, it locks the VM completely and nothing more happens, Only way to get past the problem is to stop the vzdump job and do a unlock and reset...
  14. K

    VM looses network

    Hi Right around the same time as PVE 5.2 was released, we deployed a new cluster of PVE servers running on PVE 5.2.0 Everything went fine and all VMs are working with no problems. For system monitoring we use Zabbix 3.0 LTS and we have a zabbix proxy on one of the nodes. A few months ago we...
  15. K

    NIC suddenly changed name

    Strange. The three nodes are actually quite recent installs and they were deployed with PVE 5.2. Maybe it is something with either the network driver or perhaps the installation image provided by OVH
  16. K

    NIC suddenly changed name

    Hi Today we experienced something strange. During the night, we got this output from the cronjobs on 3 out of 4 nodes: /etc/cron.daily/logrotate: Job for pveproxy.service failed. See "systemctl status pveproxy.service" and "journalctl -xe" for details. When we started working in the morning, we...
  17. K

    Removed PVE 5.2 cluster node still shows

    Hi Yesterday we took a server in our PVE cluster out of production. As per the documentation we can "pvecm delnode <nodename>" to remove the node from the cluster. Immediately after removing the node, it was no longer visible from "pvecm status" and "pvecm nodes", but we could still see it in...
  18. K

    Performance issues with ZFS

    The pool only has 800GB used on the node running the Windows servers NAME SIZE ALLOC FREE EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT rpool 3.62T 807G 2.84T - 43% 21% 1.00x ONLINE - The secondary node only has around 600GB used: NAME SIZE ALLOC FREE...
  19. K

    Performance issues with ZFS

    Hi We are running a three node PVE 5.1 cluster. Two nodes for running virtual machines and a backup node for storing VM backups and use it for virtual machines in case of an emergency. All servers use locally attached storage with 2x 4TB SATA drives in a ZFS mirror and 2x 450GB NVMe drives, on...
  20. K

    Strange behaviour with memory hotplug

    Hi During some testing of Proxmox VE (Currently still running with PVE 4.4), we have noticed some strange behaviour with some systems when memory hotplug is enabled. First thing is Ubuntu 16.04. When we enable memory hotplug, the guest OS only sees 1GB of RAM even if a higher value was...