Unexpected reboot behavior

MAMProgr

Member
May 5, 2017
4
0
6
32
Hello,
I have Proxmox 4.4 (8 Core & 32GB RAM), includes:
- VM: CentOS 7 (8 vCore & 16GB).
- VM: Windows Server 2012 (4 vCore & 4GB RAM).

When i run a backup process in CentOS 7, after about 1 hour the host server (Proxmox) reboot!

/var/log/syslog:
Code:
May 16 05:11:57 srv1 pvestatd[2339]: No balloon device has been activated
May 16 05:12:07 srv1 pvestatd[2339]: No balloon device has been activated
May 16 05:12:17 srv1 pvestatd[2339]: No balloon device has been activated
May 16 05:12:27 srv1 pvestatd[2339]: No balloon device has been activated
May 16 05:12:40 srv1 pvestatd[2339]: got timeout
May 16 05:12:48 srv1 pvestatd[2339]: No balloon device has been activated
May 16 05:12:57 srv1 pvestatd[2339]: No balloon device has been activated
May 16 05:13:07 srv1 pvestatd[2339]: No balloon device has been activated
May 16 05:14:22 srv1 rsyslogd: [origin software="rsyslogd" swVersion="8.4.2" x-pid="2180" x-info="www rsyslog com"] start
May 16 05:14:22 srv1 systemd-modules-load[1438]: Module 'fuse' is builtin
May 16 05:14:22 srv1 systemd-modules-load[1438]: Inserted module 'vhost_net'
May 16 05:14:22 srv1 systemd[1]: Mounted Huge Pages File System.
May 16 05:14:22 srv1 systemd[1]: Mounted POSIX Message Queue File System.
May 16 05:14:22 srv1 kernel: [    0.000000] Initializing cgroup subsys cpuset
May 16 05:14:22 srv1 kernel: [    0.000000] Initializing cgroup subsys cpu
May 16 05:14:22 srv1 systemd[1]: Mounted Debug File System.
May 16 05:14:22 srv1 kernel: [    0.000000] Initializing cgroup subsys cpuacct
May 16 05:14:22 srv1 systemd[1]: Mounted FUSE Control File System.
May 16 05:14:22 srv1 kernel: [    0.000000] Linux version 4.4.35-1-pve (root@elsa) (gcc version 4.9.2 (Debian 4.9.2-10) ) #1 SMP Fri Dec 9 11:09:55 CET 2016 ()
May 16 05:14:22 srv1 kernel: [    0.000000] Command line: BOOT_IMAGE=/ROOT/pve-1@/boot/vmlinuz-4.4.35-1-pve root=ZFS=rpool/ROOT/pve-1 ro root=ZFS=rpool/ROOT/pve-1 boot=zfs quiet
May 16 05:14:22 srv1 kernel: [    0.000000] KERNEL supported cpus:
May 16 05:14:22 srv1 systemd[1]: Started Create Static Device Nodes in /dev.
May 16 05:14:22 srv1 kernel: [    0.000000]   Intel GenuineIntel
May 16 05:14:22 srv1 kernel: [    0.000000]   AMD AuthenticAMD
May 16 05:14:22 srv1 kernel: [    0.000000]   Centaur CentaurHauls
May 16 05:14:22 srv1 systemd[1]: Starting udev Kernel Device Manager...
May 16 05:14:22 srv1 kernel: [    0.000000] x86/fpu: xstate_offset[2]:  576, xstate_sizes[2]:  256
May 16 05:14:22 srv1 kernel: [    0.000000] x86/fpu: Supporting XSAVE feature 0x01: 'x87 floating point registers'
May 16 05:14:22 srv1 kernel: [    0.000000] x86/fpu: Supporting XSAVE feature 0x02: 'SSE registers'
May 16 05:14:22 srv1 systemd[1]: Started udev Coldplug all Devices.
May 16 05:14:22 srv1 kernel: [    0.000000] x86/fpu: Supporting XSAVE feature 0x04: 'AVX registers'
May 16 05:14:22 srv1 kernel: [    0.000000] x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format.
May 16 05:14:22 srv1 kernel: [    0.000000] x86/fpu: Using 'eager' FPU context switches.
May 16 05:14:22 srv1 kernel: [    0.000000] e820: BIOS-provided physical RAM map:
May 16 05:14:22 srv1 systemd[1]: Starting udev Wait for Complete Device Initialization...
May 16 05:14:22 srv1 kernel: [    0.000000] BIOS-e820: [mem 0x0000000000000000-0x000000000009d7ff] usable
May 16 05:14:22 srv1 kernel: [    0.000000] BIOS-e820: [mem 0x000000000009d800-0x000000000009ffff] reserved
May 16 05:14:22 srv1 kernel: [    0.000000] BIOS-e820: [mem 0x00000000000e0000-0x00000000000fffff] reserved
May 16 05:14:22 srv1 systemd[1]: Started udev Kernel Device Manager.

How can i detect the problem?
Thanks.
 
Thanks for reply,
I upgraded to latest version:
Code:
root@fh02:~# pveversion -v
proxmox-ve: 4.4-87 (running kernel: 4.4.59-1-pve)
pve-manager: 4.4-13 (running version: 4.4-13/7ea56165)
pve-kernel-4.4.35-1-pve: 4.4.35-77
pve-kernel-4.4.59-1-pve: 4.4.59-87
lvm2: 2.02.116-pve3
corosync-pve: 2.4.2-2~pve4+1
libqb0: 1.0.1-1
pve-cluster: 4.0-49
qemu-server: 4.0-110
pve-firmware: 1.1-11
libpve-common-perl: 4.0-94
libpve-access-control: 4.0-23
libpve-storage-perl: 4.0-76
pve-libspice-server1: 0.12.8-2
vncterm: 1.3-2
pve-docs: 4.4-4
pve-qemu-kvm: 2.7.1-4
pve-container: 1.0-99
pve-firewall: 2.0-33
pve-ha-manager: 1.0-40
ksm-control-daemon: 1.2-1
glusterfs-client: 3.5.2-2+deb8u3
lxc-pve: 2.0.7-4
lxcfs: 2.0.6-pve1
criu: 1.6.0-1
novnc-pve: 0.5-9
smartmontools: 6.5+svn4324-1~pve80
zfsutils: 0.6.5.9-pve15~bpo80
But the problem still appears, server reboots.
Note: When i reduced the RAM for CentOS7 to 12GB. The problem gone but when i increased the RAM server reboots.
 
Hello,
I have the same problem with 4.4.35-1-pve version, the system reboot when I launch a third VM

Hardware : Dell
CPUs : 8 x Intel(R) Xeon(R) CPU E3-1230 v6 @ 3.50GHz (1 Socket)
Kernel Version : Linux 4.4.35-1-pve #1 SMP Fri Dec 9 11:09:55 CET 2016
PVE Manager Version : pve-manager/4.4-1/eb2d6f1e
8 Gb RAM total
---
First VM : Zentyal 4.2
2 Gb RAM
1 socket 2 cores
---
Second VM : Win 7 Pro
4 Gb RAM
1 socket 4 cores

---
Third VM : SME 9
2 Gb RAM
1 socket 2 cores​
 
I tried to downgrade ram on my second VM, and it work's !

We can retain : if all memory used by VMs, a crash reboot occurs !

Thank's hugotor :)
I don't know if 5 version do the same !?
You confirm ... ?
 
  • Like
Reactions: albans
I tried to downgrade ram on my second VM, and it work's !

We can retain : if all memory used by VMs, a crash reboot occurs !

Thank's hugotor :)
I don't know if 5 version do the same !?
You confirm ... ?


Had same issue - reducing ram within VMs is the way to stabilize Proxmox. Thank you for the tips.