Hi, I have a bit of a problem with a test system I am running.
Version details are below but the key factors are that I am booting from a single conventional hard disk, setup as a ZFS pool (called rpool). My reason for using zfs is that I plan to replicate the partition so I have a complete backup.
The 'problem' is that I have only 8GB RAM in the host. So when I migrated (live) a VM onto this host, it didn't have enough RAM to support the additional VM. It couldn't use its swap file, because being a clean ZFS build, it didn't have one!
root@pve1:~# free -h
total used free shared buff/cache available
Mem: 7.7Gi 4.5Gi 129Mi 71Mi 3.2Gi 3.0Gi
Swap: 0B 0B 0B
root@pve1:~#
So I got this
Mar 04 16:12:24 pve1 kernel: Out of memory: Killed process 11295 (kvm) total-vm:4906856kB, anon-rss:4227176kB, file-rss:1560kB, shmem-rss:4kB
Mar 04 16:12:24 pve1 kernel: oom_reaper: reaped process 11295 (kvm), now anon-rss:0kB, file-rss:36kB, shmem-rss:4kB
Mar 04 16:12:24 pve1 systemd[1]: 104.scope: Succeeded.
Mar 04 16:12:24 pve1 qmeventd[772]: Starting cleanup for 104
i.e - It shutdown one of the VMs already running on the host. And it wasn't the one I just started either!
Obviously, this is going to be a problem in a production scenario.
Can anyone offer any advice?
Many thanks
Tom
VERSIONS
proxmox-ve: 6.1-2 (running kernel: 5.3.10-1-pve) pve-manager: 6.1-3 (running version: 6.1-3/37248ce6) pve-kernel-5.3: 6.0-12 pve-kernel-helper: 6.0-12 pve-kernel-5.3.10-1-pve: 5.3.10-1 ceph-fuse: 12.2.11+dfsg1-2.1+b1 corosync: 3.0.2-pve4 criu: 3.11-3 glusterfs-client: 5.5-3 ifupdown: not correctly installed ifupdown2: 1.2.5-1 ksm-control-daemon: 1.3-1 libjs-extjs: 6.0.1-10 libknet1: 1.13-pve1
SYSLOG
Mar 04 16:12:16 pve1 pvedaemon[13362]: start VM 105: UPIDve1:00003432:0007BE93:5E5FD360:qmstart:105:root@pam:
Mar 04 16:12:16 pve1 pvedaemon[1753]: <root@pam> starting task UPIDve1:00003432:0007BE93:5E5FD360:qmstart:105:root@pam:
Mar 04 16:12:16 pve1 systemd[1]: Started 105.scope.
Mar 04 16:12:16 pve1 systemd-udevd[13417]: Using default interface naming scheme 'v240'.
Mar 04 16:12:16 pve1 systemd-udevd[13417]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Mar 04 16:12:16 pve1 systemd-udevd[13417]: Could not generate persistent MAC address for tap105i0: No such file or directory
Mar 04 16:12:16 pve1 kernel: device tap105i0 entered promiscuous mode
Mar 04 16:12:16 pve1 ovs-vsctl[13430]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl del-port tap105i0
Mar 04 16:12:16 pve1 ovs-vsctl[13430]: ovs|00002|db_ctl_base|ERR|no port named tap105i0
Mar 04 16:12:16 pve1 ovs-vsctl[13431]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl del-port fwln105i0
Mar 04 16:12:16 pve1 ovs-vsctl[13431]: ovs|00002|db_ctl_base|ERR|no port named fwln105i0
Mar 04 16:12:16 pve1 ovs-vsctl[13432]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl add-port vmbr0 tap105i0 tag=1201
Mar 04 16:12:16 pve1 pvedaemon[1753]: <root@pam> end task UPIDve1:00003432:0007BE93:5E5FD360:qmstart:105:root@pam: OK
Mar 04 16:12:21 pve1 pvedaemon[7663]: VM 105 qmp command failed - VM 105 qmp command 'guest-ping' failed - got timeout
Mar 04 16:12:24 pve1 kernel: kvm invoked oom-killer: gfp_mask=0x100dca(GFP_HIGHUSER_MOVABLE|__GFP_ZERO), order=0, oom_score_adj=0
Mar 04 16:12:24 pve1 kernel: CPU: 1 PID: 13446 Comm: kvm Tainted: P O 5.3.10-1-pve #1
Mar 04 16:12:24 pve1 kernel: Hardware name: Dell Inc. PowerEdge R210 II/03X6X0, BIOS 2.7.0 11/15/2013
Mar 04 16:12:24 pve1 kernel: 0 pages hwpoisoned
Mar 04 16:12:24 pve1 kernel: Tasks state (memory values in pages):
Mar 04 16:12:24 pve1 kernel: oom-kill:constraint=CONSTRAINT_NONE,nodemask=(null),cpuset=/,mems_allowed=0,global_oom,task_memcg=/qemu.slice/104.scope,task=kvm,pid=11295,uid=0
Mar 04 16:12:24 pve1 kernel: Out of memory: Killed process 11295 (kvm) total-vm:4906856kB, anon-rss:4227176kB, file-rss:1560kB, shmem-rss:4kB
Mar 04 16:12:24 pve1 kernel: oom_reaper: reaped process 11295 (kvm), now anon-rss:0kB, file-rss:36kB, shmem-rss:4kB
Mar 04 16:12:24 pve1 systemd[1]: 104.scope: Succeeded.
Mar 04 16:12:24 pve1 qmeventd[772]: Starting cleanup for 104
Mar 04 16:12:24 pve1 ovs-vsctl[13543]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl del-port fwln104i0
Mar 04 16:12:24 pve1 ovs-vsctl[13543]: ovs|00002|db_ctl_base|ERR|no port named fwln104i0
Mar 04 16:12:24 pve1 ovs-vsctl[13544]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl del-port tap104i0
Mar 04 16:12:25 pve1 qmeventd[772]: Finished cleanup for 104
Version details are below but the key factors are that I am booting from a single conventional hard disk, setup as a ZFS pool (called rpool). My reason for using zfs is that I plan to replicate the partition so I have a complete backup.
The 'problem' is that I have only 8GB RAM in the host. So when I migrated (live) a VM onto this host, it didn't have enough RAM to support the additional VM. It couldn't use its swap file, because being a clean ZFS build, it didn't have one!
root@pve1:~# free -h
total used free shared buff/cache available
Mem: 7.7Gi 4.5Gi 129Mi 71Mi 3.2Gi 3.0Gi
Swap: 0B 0B 0B
root@pve1:~#
So I got this
Mar 04 16:12:24 pve1 kernel: Out of memory: Killed process 11295 (kvm) total-vm:4906856kB, anon-rss:4227176kB, file-rss:1560kB, shmem-rss:4kB
Mar 04 16:12:24 pve1 kernel: oom_reaper: reaped process 11295 (kvm), now anon-rss:0kB, file-rss:36kB, shmem-rss:4kB
Mar 04 16:12:24 pve1 systemd[1]: 104.scope: Succeeded.
Mar 04 16:12:24 pve1 qmeventd[772]: Starting cleanup for 104
i.e - It shutdown one of the VMs already running on the host. And it wasn't the one I just started either!
Obviously, this is going to be a problem in a production scenario.
Can anyone offer any advice?
Many thanks
Tom
VERSIONS
proxmox-ve: 6.1-2 (running kernel: 5.3.10-1-pve) pve-manager: 6.1-3 (running version: 6.1-3/37248ce6) pve-kernel-5.3: 6.0-12 pve-kernel-helper: 6.0-12 pve-kernel-5.3.10-1-pve: 5.3.10-1 ceph-fuse: 12.2.11+dfsg1-2.1+b1 corosync: 3.0.2-pve4 criu: 3.11-3 glusterfs-client: 5.5-3 ifupdown: not correctly installed ifupdown2: 1.2.5-1 ksm-control-daemon: 1.3-1 libjs-extjs: 6.0.1-10 libknet1: 1.13-pve1
SYSLOG
Mar 04 16:12:16 pve1 pvedaemon[13362]: start VM 105: UPIDve1:00003432:0007BE93:5E5FD360:qmstart:105:root@pam:
Mar 04 16:12:16 pve1 pvedaemon[1753]: <root@pam> starting task UPIDve1:00003432:0007BE93:5E5FD360:qmstart:105:root@pam:
Mar 04 16:12:16 pve1 systemd[1]: Started 105.scope.
Mar 04 16:12:16 pve1 systemd-udevd[13417]: Using default interface naming scheme 'v240'.
Mar 04 16:12:16 pve1 systemd-udevd[13417]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Mar 04 16:12:16 pve1 systemd-udevd[13417]: Could not generate persistent MAC address for tap105i0: No such file or directory
Mar 04 16:12:16 pve1 kernel: device tap105i0 entered promiscuous mode
Mar 04 16:12:16 pve1 ovs-vsctl[13430]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl del-port tap105i0
Mar 04 16:12:16 pve1 ovs-vsctl[13430]: ovs|00002|db_ctl_base|ERR|no port named tap105i0
Mar 04 16:12:16 pve1 ovs-vsctl[13431]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl del-port fwln105i0
Mar 04 16:12:16 pve1 ovs-vsctl[13431]: ovs|00002|db_ctl_base|ERR|no port named fwln105i0
Mar 04 16:12:16 pve1 ovs-vsctl[13432]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl add-port vmbr0 tap105i0 tag=1201
Mar 04 16:12:16 pve1 pvedaemon[1753]: <root@pam> end task UPIDve1:00003432:0007BE93:5E5FD360:qmstart:105:root@pam: OK
Mar 04 16:12:21 pve1 pvedaemon[7663]: VM 105 qmp command failed - VM 105 qmp command 'guest-ping' failed - got timeout
Mar 04 16:12:24 pve1 kernel: kvm invoked oom-killer: gfp_mask=0x100dca(GFP_HIGHUSER_MOVABLE|__GFP_ZERO), order=0, oom_score_adj=0
Mar 04 16:12:24 pve1 kernel: CPU: 1 PID: 13446 Comm: kvm Tainted: P O 5.3.10-1-pve #1
Mar 04 16:12:24 pve1 kernel: Hardware name: Dell Inc. PowerEdge R210 II/03X6X0, BIOS 2.7.0 11/15/2013
Mar 04 16:12:24 pve1 kernel: 0 pages hwpoisoned
Mar 04 16:12:24 pve1 kernel: Tasks state (memory values in pages):
Mar 04 16:12:24 pve1 kernel: oom-kill:constraint=CONSTRAINT_NONE,nodemask=(null),cpuset=/,mems_allowed=0,global_oom,task_memcg=/qemu.slice/104.scope,task=kvm,pid=11295,uid=0
Mar 04 16:12:24 pve1 kernel: Out of memory: Killed process 11295 (kvm) total-vm:4906856kB, anon-rss:4227176kB, file-rss:1560kB, shmem-rss:4kB
Mar 04 16:12:24 pve1 kernel: oom_reaper: reaped process 11295 (kvm), now anon-rss:0kB, file-rss:36kB, shmem-rss:4kB
Mar 04 16:12:24 pve1 systemd[1]: 104.scope: Succeeded.
Mar 04 16:12:24 pve1 qmeventd[772]: Starting cleanup for 104
Mar 04 16:12:24 pve1 ovs-vsctl[13543]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl del-port fwln104i0
Mar 04 16:12:24 pve1 ovs-vsctl[13543]: ovs|00002|db_ctl_base|ERR|no port named fwln104i0
Mar 04 16:12:24 pve1 ovs-vsctl[13544]: ovs|00001|vsctl|INFO|Called as /usr/bin/ovs-vsctl del-port tap104i0
Mar 04 16:12:25 pve1 qmeventd[772]: Finished cleanup for 104