VM NOT RUNNING

cr_eis

Member
Feb 3, 2021
20
3
8
50
Hello,
I 'm a newbe in promox's world and i had got a problem with a VM under Windows server 2019 whot it stoppped by itself.
In the syslog file, i found this message :

Code:
Feb  3 13:15:06 promox2 QEMU[1946]: kvm: hw/ide/core.c:724: ide_cancel_dma_sync: Assertion `s->bus->dma->aiocb == NULL' failed.
Feb  3 13:15:09 promox2 pvestatd[1739]: VM 100 qmp command failed - VM 100 not running

Somebody knows what it happens ?
 
what is your 'pveversion -v' ?
 
Code:
proxmox-ve: 6.3-1 (running kernel: 5.4.78-2-pve)
pve-manager: 6.3-3 (running version: 6.3-3/eee5f901)
pve-kernel-5.4: 6.3-3
pve-kernel-helper: 6.3-3
pve-kernel-5.4.78-2-pve: 5.4.78-2
pve-kernel-5.4.60-1-pve: 5.4.60-2
pve-kernel-5.4.34-1-pve: 5.4.34-2
ceph-fuse: 12.2.11+dfsg1-2.1+b1
corosync: 3.0.4-pve1
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: 0.8.35+pve1
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.16-pve1
libproxmox-acme-perl: 1.0.7
libproxmox-backup-qemu0: 1.0.2-1
libpve-access-control: 6.1-3
libpve-apiclient-perl: 3.1-3
libpve-common-perl: 6.3-2
libpve-guest-common-perl: 3.1-4
libpve-http-server-perl: 3.1-1
libpve-storage-perl: 6.3-4
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve4
lxc-pve: 4.0.3-1
lxcfs: 4.0.6-pve1
novnc-pve: 1.1.0-1
proxmox-backup-client: 1.0.6-1
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.4-3
pve-cluster: 6.2-1
pve-container: 3.3-2
pve-docs: 6.3-1
pve-edk2-firmware: 2.20200531-1
pve-firewall: 4.1-3
pve-firmware: 3.1-3
pve-ha-manager: 3.1-1
pve-i18n: 2.2-2
pve-qemu-kvm: 5.1.0-8
pve-xtermjs: 4.7.0-3
qemu-server: 6.3-3
smartmontools: 7.1-pve2
spiceterm: 3.1-1
vncterm: 1.6-2
zfsutils-linux: 0.8.5-pve1
 
Hi,

i have the same Problem :/

Code:
Feb 10 15:46:06 wtf QEMU[1017]: kvm: hw/ide/core.c:724: ide_cancel_dma_sync: Assertion `s->bus->dma->aiocb == NULL' failed.

Any possibility to have a workaround? So the vm doenst stop?

Best Regards

Code:
pveversion -v
proxmox-ve: 6.3-1 (running kernel: 5.4.78-2-pve)
pve-manager: 6.3-3 (running version: 6.3-3/eee5f901)
pve-kernel-5.4: 6.3-3
pve-kernel-helper: 6.3-3
pve-kernel-5.4.78-2-pve: 5.4.78-2
ceph-fuse: 12.2.11+dfsg1-2.1+b1
corosync: 3.0.4-pve1
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: 0.8.35+pve1
libjs-extjs: 6.0.1-10
libknet1: 1.16-pve1
libproxmox-acme-perl: 1.0.7
libproxmox-backup-qemu0: 1.0.2-1
libpve-access-control: 6.1-3
libpve-apiclient-perl: 3.1-3
libpve-common-perl: 6.3-2
libpve-guest-common-perl: 3.1-4
libpve-http-server-perl: 3.1-1
libpve-storage-perl: 6.3-5
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve4
lxc-pve: 4.0.6-1
lxcfs: 4.0.6-pve1
novnc-pve: 1.1.0-1
proxmox-backup-client: 1.0.6-1
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.4-4
pve-cluster: 6.2-1
pve-container: 3.3-2
pve-docs: 6.3-1
pve-edk2-firmware: 2.20200531-1
pve-firewall: 4.1-3
pve-firmware: 3.1-3
pve-ha-manager: 3.1-1
pve-i18n: 2.2-2
pve-qemu-kvm: 5.1.0-8
pve-xtermjs: 4.7.0-3
qemu-server: 6.3-3
smartmontools: 7.1-pve2
spiceterm: 3.1-1
vncterm: 1.6-2
zfsutils-linux: 0.8.5-pve1
 
Last edited:
Hi,

i have the same Problem :/

Code:
Feb 10 15:46:06 wtf QEMU[1017]: kvm: hw/ide/core.c:724: ide_cancel_dma_sync: Assertion `s->bus->dma->aiocb == NULL' failed.

Any possibility to have a workaround? So the vm doenst stop?

Best Regards

Code:
pveversion -v
proxmox-ve: 6.3-1 (running kernel: 5.4.78-2-pve)
pve-manager: 6.3-3 (running version: 6.3-3/eee5f901)
pve-kernel-5.4: 6.3-3
pve-kernel-helper: 6.3-3
pve-kernel-5.4.78-2-pve: 5.4.78-2
ceph-fuse: 12.2.11+dfsg1-2.1+b1
corosync: 3.0.4-pve1
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: 0.8.35+pve1
libjs-extjs: 6.0.1-10
libknet1: 1.16-pve1
libproxmox-acme-perl: 1.0.7
libproxmox-backup-qemu0: 1.0.2-1
libpve-access-control: 6.1-3
libpve-apiclient-perl: 3.1-3
libpve-common-perl: 6.3-2
libpve-guest-common-perl: 3.1-4
libpve-http-server-perl: 3.1-1
libpve-storage-perl: 6.3-5
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve4
lxc-pve: 4.0.6-1
lxcfs: 4.0.6-pve1
novnc-pve: 1.1.0-1
proxmox-backup-client: 1.0.6-1
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.4-4
pve-cluster: 6.2-1
pve-container: 3.3-2
pve-docs: 6.3-1
pve-edk2-firmware: 2.20200531-1
pve-firewall: 4.1-3
pve-firmware: 3.1-3
pve-ha-manager: 3.1-1
pve-i18n: 2.2-2
pve-qemu-kvm: 5.1.0-8
pve-xtermjs: 4.7.0-3
qemu-server: 6.3-3
smartmontools: 7.1-pve2
spiceterm: 3.1-1
vncterm: 1.6-2
zfsutils-linux: 0.8.5-pve1

Mate, I have been having a similar issue for the past few days. My VM is stored on SSD and I had "Discard" option checked. I have unchecked this option and my VM has been running ever since. Not entirely sure why but this has helped out a bit!

1614398568786.png
 
Hello,
I 'm a newbe in promox's world and i had got a problem with a VM under Windows server 2019 whot it stoppped by itself.
In the syslog file, i found this message :

Code:
Feb  3 13:15:06 promox2 QEMU[1946]: kvm: hw/ide/core.c:724: ide_cancel_dma_sync: Assertion `s->bus->dma->aiocb == NULL' failed.
Feb  3 13:15:09 promox2 pvestatd[1739]: VM 100 qmp command failed - VM 100 not running

Somebody knows what it happens ?

Mate, I have been having a similar issue for the past few days. My VM is stored on SSD and I had "Discard" option checked. I have unchecked this option and my VM has been running ever since. Not entirely sure why but this has helped out a bit!

1614398830861.png


1614398779090.png
 
Mate, I have been having a similar issue for the past few days. My VM is stored on SSD and I had "Discard" option checked. I have unchecked this option and my VM has been running ever since. Not entirely sure why but this has helped out a bit!

View attachment 24002

Mate, I have been having a similar issue for the past few days. My VM is stored on SSD and I had "Discard" option checked. I have unchecked this option and my VM has been running ever since. Not entirely sure why but this has helped out a bit!

View attachment 24002
Hi , thank you for the hint! I'll give it a try!

FYI: Also i have configured "write back" as Cache ,like it is written in the Best Practise Guide of the Proxmox Wiki.
 
Hi , thank you for the hint! I'll give it a try!

FYI: Also i have configured "write back" as Cache ,like it is written in the Best Practise Guide of the Proxmox Wiki.
@eXcited thanks mate, will configure "write back".

Upon further investigation of the previous issue/error, I have ascertained it mainly affects VM configured to use "IDE". I have changed all my VM's to use "VirtIO block" and enable discard option again and error fixed.

1615110455613.png

1615110554290.png
 
Hi, i already had it configured with VirtIO SCSI. So i unchecked acutally the discard Feature

virtioscsi.png
But i see you have configured virtio0 and i have confgured ide0. I couldnt install windows server 2019 with the apropiate driver wirh scsi bus type :/

PS: I changed the Bus Type from IDE to VirtIo an reactivated the discard Flag
 
Last edited:
Hi, i already had it configured with VirtIO SCSI. So i unchecked acutally the discard Feature

View attachment 24216
But i see you have configured virtio0 and i have confgured ide0. I couldnt install windows server 2019 with the apropiate driver wirh scsi bus type :/

PS: I changed the Bus Type from IDE to VirtIo an reactivated the discard Flag
@eXcited did the error go away? Is the machine running?

I changed to virtIO as is offer better perfomance over IDE.
 
@eXcited did the error go away? Is the machine running?

I changed to virtIO as is offer better perfomance over IDE.
I didn't have the Problem every Minute. I had it once a week or something. I have to keep an eye now on it , if the error occurs again.
 
Last edited:
Hey Guys, I'm having the same issue once a week, and it only affects Windows VMs.

I'll try unchecking 'discard' and changing to "write back" as a test in one of the VMs and report back.
 
It might be because of my setup (SAN / iSCSI) but I can only use IDE on Windows VMS and BUS SCSI with "VirtIO SCSI" controller on the Linux ones. Any other combination make the disk not available during OS installation.

About the VMs shutting down, I have to wait until the weekend to say if it worked. I'm keeping my fingers crossed.
 
Thanks man! I don't know how I missed that. I'm giving it a try today. If it works I should be able to create a template.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!