Recent content by Vasu Sreekumar

  1. "start failed - can't aquire lock '/var/run/pveproxy/pveproxy.pid.lock' - Resource temporarily unava

    I am facing this issue in multiple nodes. Node shows grey in cluster, but ssh works and *the proxy service it is not active* When I start proxy i get the message "start failed - can't aquire lock '/var/run/pveproxy/pveproxy.pid.lock' - Resource temporarily unavailable" Only solution is to...
  2. [SOLVED] Proxmox and ram /swap usage

    I use 4GB for zfs_arc_max I have no issues, i have 25+ nodes running with 5 to 10 guests.
  3. Cluster nodes "offline" but working

    Try all 4 service pve-cluster restart service pveproxy restart service pvedaemon restart service pvestatd restart
  4. Some windows server 2003 are shutting down automatically in Proxmox VE 5.1

    I faced another issue with Promxox 5.1-46 and windows 2003. After few hours it shows blue screen. I tried all possible combinations of CPU, Drive and other settings, still same issue. It was a running KVM on Proxmox 3.4, issue started after migrating to 5.1.46. Any suggestions?
  5. Please update zfs package for 0.7.7 asap

    https://www.servethehome.com/zfs-on-linux-0-7-7-disappearing-file-bug/
  6. LXC not start

    Upgrade to kernel 4.15.10 and see.
  7. Corosync spam

    What is kernel version?
  8. Node with question mark

    Try all 4 service pve-cluster restart service pveproxy restart service pvedaemon restart service pvestatd restart
  9. Node with question mark

    There was no error messages. Running service pve-cluster restart cleared the grey issue. No reboot was required. All 4 nodes are on same CISCO 1000 mbps switch.
  10. Node with question mark

    root@S034:~# pvesh get /cluster/resources 200 OK [ { "cpu" : 0.0297653311349122, "disk" : 0, "diskread" : 9547070464, "diskwrite" : 31136726528, "id" : "qemu/204", "maxcpu" : 8, "maxdisk" : 53687091200, "maxmem" : 8589934592, "mem" ...
  11. LXC 3.0.0 has been released

    When LXC 3.0.0 will be available in Proxmox?
  12. Node with question mark

    I also started facing same issue with 4.15.10 with KVM guests. I have 4 nodes, only one shows green, rest all three grey. But all nodes and guests pinging fine.
  13. 4.15 based test kernel for PVE 5.x available

    System crashed and restarted time 20:02:00 (I have 25 live nodes, 1 or 2 nodes crashes like this everyday. ) Log file. Mar 29 19:39:42 Q172 pvedaemon[2841]: <root@pam> successful auth for user 'root@pam' Mar 29 19:39:56 Q172 pvedaemon[9167]: <root@pam> successful auth for user 'root@pam' Mar...
  14. 4.15 based test kernel for PVE 5.x available

    With default settings, node crashes when I start the 4th guest. With changed settings node does not crash when i start 4th guest, since it starts KSM early enough. I think it is more like LXC related issues than a bug. In KVM I didn't face any issues.
  15. 4.15 based test kernel for PVE 5.x available

    pve-kernel-4.15.10-1-pve also has the above KSM sharing issue. If you have plenty of memory, you will not see it. I have 25+ nodes, and i don't have plenty of memory, so i see it often. But after setting the % of KSM thresh hold., i d didn't face any issue.

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE and Proxmox Mail Gateway. We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get your own in 60 seconds.

Buy now!