Recent content by Vasu Sreekumar

  1. V

    "start failed - can't aquire lock '/var/run/pveproxy/pveproxy.pid.lock' - Resource temporarily unava

    I am facing this issue in multiple nodes. Node shows grey in cluster, but ssh works and *the proxy service it is not active* When I start proxy i get the message "start failed - can't aquire lock '/var/run/pveproxy/pveproxy.pid.lock' - Resource temporarily unavailable" Only solution is to...
  2. V

    [SOLVED] Proxmox and ram /swap usage

    I use 4GB for zfs_arc_max I have no issues, i have 25+ nodes running with 5 to 10 guests.
  3. V

    Cluster nodes "offline" but working

    Try all 4 service pve-cluster restart service pveproxy restart service pvedaemon restart service pvestatd restart
  4. V

    Some windows server 2003 are shutting down automatically in Proxmox VE 5.1

    I faced another issue with Promxox 5.1-46 and windows 2003. After few hours it shows blue screen. I tried all possible combinations of CPU, Drive and other settings, still same issue. It was a running KVM on Proxmox 3.4, issue started after migrating to 5.1.46. Any suggestions?
  5. V

    Please update zfs package for 0.7.7 asap

    https://www.servethehome.com/zfs-on-linux-0-7-7-disappearing-file-bug/
  6. V

    LXC not start

    Upgrade to kernel 4.15.10 and see.
  7. V

    Corosync spam

    What is kernel version?
  8. V

    Node with question mark

    Try all 4 service pve-cluster restart service pveproxy restart service pvedaemon restart service pvestatd restart
  9. V

    Node with question mark

    There was no error messages. Running service pve-cluster restart cleared the grey issue. No reboot was required. All 4 nodes are on same CISCO 1000 mbps switch.
  10. V

    Node with question mark

    root@S034:~# pvesh get /cluster/resources 200 OK [ { "cpu" : 0.0297653311349122, "disk" : 0, "diskread" : 9547070464, "diskwrite" : 31136726528, "id" : "qemu/204", "maxcpu" : 8, "maxdisk" : 53687091200, "maxmem" : 8589934592, "mem" ...
  11. V

    LXC 3.0.0 has been released

    When LXC 3.0.0 will be available in Proxmox?
  12. V

    Node with question mark

    I also started facing same issue with 4.15.10 with KVM guests. I have 4 nodes, only one shows green, rest all three grey. But all nodes and guests pinging fine.
  13. V

    4.15 based test kernel for PVE 5.x available

    System crashed and restarted time 20:02:00 (I have 25 live nodes, 1 or 2 nodes crashes like this everyday. ) Log file. Mar 29 19:39:42 Q172 pvedaemon[2841]: <root@pam> successful auth for user 'root@pam' Mar 29 19:39:56 Q172 pvedaemon[9167]: <root@pam> successful auth for user 'root@pam' Mar...
  14. V

    4.15 based test kernel for PVE 5.x available

    With default settings, node crashes when I start the 4th guest. With changed settings node does not crash when i start 4th guest, since it starts KSM early enough. I think it is more like LXC related issues than a bug. In KVM I didn't face any issues.
  15. V

    4.15 based test kernel for PVE 5.x available

    pve-kernel-4.15.10-1-pve also has the above KSM sharing issue. If you have plenty of memory, you will not see it. I have 25+ nodes, and i don't have plenty of memory, so i see it often. But after setting the % of KSM thresh hold., i d didn't face any issue.