Search results

  1. P

    Some Windows guests hanging every night

    HP ProLiant DL365 each with 6x 300GB SATA disks in RAID 1+0 (striping across 3 mirrored pairs) on P400i RAID controllers with 512MB BBWC (one has a disconnected battery so is using the cache for read-only. The other has a problem with one of the physical drive slots and keeps seeing the drive...
  2. P

    Some Windows guests hanging every night

    Well, it happened twice now on the VM I moved to different hardware (whose other Windows VMs are working fine.) So it's _something_ with the guest but how the heck do I find out what?! Next I will try booting one of the hosts on the previous kernel and see what happens.
  3. P

    Some Windows guests hanging every night

    AAAHHH HELP!!! Even with cache=directsync, two of the three affected VMs hung again last night! This is maddening and I need it to stop!!! What else can I do?! And they are on two different hosts!
  4. P

    Some Windows guests hanging every night

    For anyone else reading this, I was pointed to this article by someone in the IRC channel which explains the cache modes clearly: http://webapp5.rrz.uni-hamburg.de/SuSe-Dokumentation/manual/sles-manuals_en/cha.qemu.cachemodes.html By those descriptions (and the qemu-kvm man page,) it sounds...
  5. P

    Some Windows guests hanging every night

    Nope, it too hung again. But it happened when the system was rebuilding a failed disk, so I/O delay may be the cause here too. I've set all of the Windows VMs to directsync caching and restarted them, verifying in Device manager that it's indeed off. Let's see if that helps. (And all this time...
  6. P

    Some Windows guests hanging every night

    Okay, but https://pve.proxmox.com/wiki/Performance_Tweaks says to "Avoid to use cache=directsync and writethrough with qcow2 files." but it doesn't say why. Also, since I'm using XFS to store the qcow2 files, it has write barrier support enabled by default, FWIW.
  7. P

    Some Windows guests hanging every night

    So far the moved VM is still operating correctly while one still on the problem host just experienced the problem again. I was asking for guidance in ##windows and some people mentioned that storage performance problems can cause weird issues like this in guests. So I checked my RAID controller...
  8. P

    Some Windows guests hanging every night

    Done. Since the other host was secondary anyway, it wasn't a big deal, but there were indeed 340K worth of out-of-sync data that is now corrected. (BTW, do I need to set up automatic DRBD resource verification on all nodes or just one of each pair?)
  9. P

    Some Windows guests hanging every night

    Okay here's the whole storage hierarchy: - Physical disks - Hardware RAID (with BBU cache) - Partitions (LVM) - system, DRBD-backing1 (this host's VMs), DRBD-backing2 (other host's VMs) - DRBD devices - XFS File systems on DRBD devices - qcow2 disk image files, set to "Default (no cache)" Right...
  10. P

    Some Windows guests hanging every night

    I've just been using Default (no cache) on all my KVM VMs on all hosts. Is that not a good idea? Oh wow, that's not intuitive at all! (I need to read the docs again I guess.) Still, one of the affected VMs is set to a fixed 4GB, and the host has enough RAM for all of them at their max set...
  11. P

    Some Windows guests hanging every night

    The really strange part is that none of my other machines in the cluster have the problem and they're all on the same version (and hardware.) However this host is the only one with four Windows VMs running, incase that matters. They use 7 VCPU cores in total though and the physical machine has...
  12. P

    Some Windows guests hanging every night

    That's the only version I have. (I'm not clear on how to get the 3.2 one.) proxmox-ve-2.6.32: 3.3-138 (running kernel: 2.6.32-33-pve) pve-manager: 3.3-2 (running version: 3.3-2/995e687e) pve-kernel-2.6.32-32-pve: 2.6.32-136 pve-kernel-2.6.32-33-pve: 2.6.32-138 lvm2: 2.02.98-pve4 clvm...
  13. P

    Some Windows guests hanging every night

    Some additional information: The problems started shortly after installing the following updates: Start-Date: 2014-10-05 01:30:05 Commandline: apt-get dist-upgrade Install: pve-kernel-2.6.32-33-pve:amd64 (2.6.32-138, automatic) Upgrade: pve-qemu-kvm:amd64 (2.1-8, 2.1-9), librbd1:amd64...
  14. P

    Some Windows guests hanging every night

    Hey everyone. Within the past week, I'm having problems with many (but not all) of my Windows guests freezing/hanging almost every night. (In one case it happened at 8AM.) They're mostly Windows 2008 R2 x64 but one is Windows 2003 R2 x64. What's really weird is that the 2008 ones still respond...
  15. P

    4+ node management cluster with DRBD pairs: Where to do fencing for HA?

    But in my case, I'll have three separate shared storage devices with two nodes attached to each. Each will need HA, but only within its pair. From what I read, I should be able to do this with failover domains, right? Feature request, please, ProxMox developers! ;) Would setting DRBD to...
  16. P

    4+ node management cluster with DRBD pairs: Where to do fencing for HA?

    Hello. I have a ProxMox management cluster that currently has 5 nodes (it will have 6 eventually) and I plan to set up DRBD to mirror data between machine pairs whith a goal of high-availability (via automatic live migration.) But where should I implement fencing? In DRBD, in ProxMox, or both...
  17. P

    Problem to configure HA with Fencing

    I found that for newer iLO firmware, you need to use the fence tool for the next version up. e.g. for iLO2 with firmware from 2013, I have to use fence_ilo3 -o status -a 192.168.0.x
  18. P

    rebooting node, unmounting configfs takes forever...

    I'm having this problem too on all of my nodes and have to bring them down hard. What can we do?
  19. P

    OpenVZ Ploop support

    I'm interested in ploop because I run my VMs on an xfs file system. It's well-known that vzquota doesn't support xfs (https://bugzilla.openvz.org/show_bug.cgi?id=128) and according to that bug, vzquota is now deprecated in favor of ploop. It seems the OpenVZ team would like everyone to migrate...
  20. P

    Bug - PVE Cluster: VNC error on node

    And replying to my own problem, apparently you must manually SSH as root once from the machine whose Web GUI you're accessing to every other cluster node to add the keys to the known_hosts file. Even though I had done this for node3 from node1 already, I hadn't done it for node2 from node1 which...