Hi Fiona,Hi,
when a VM gets stuck, you can runstrace -c -p $(cat /var/run/qemu-server/<ID>.pid)
with the ID of the VM. Press Ctrl+C after about 10 seconds to get the output.
You can also install debugger and debug symbolsapt install pve-qemu-kvm-dbg gdb
and then rungdb --batch --ex 't a a bt' -p $(cat /var/run/qemu-server/<ID>.pid)
.
When you share this information, please also share the output ofqm config <ID>
andpveversion -v
to make it easier to correlate things.
If we're lucky, those will give some idea where it's stuck.
If you don't have latest microcode and BIOS update installed, please try that first.
we are "lucky" and have the issue on one VM (win).
The output:
Code:
strace -c -p $(cat /var/run/qemu-server/$ID.pid)
strace: Process 38494 attached
^Cstrace: Process 38494 detached
% time seconds usecs/call calls errors syscall
------ ----------- ----------- --------- --------- ----------------
99.71 19.803374 34621 572 ppoll
0.14 0.028442 13 2136 write
0.08 0.016147 29 549 read
0.06 0.012145 23 523 recvmsg
0.00 0.000007 0 10 sendmsg
0.00 0.000002 1 2 close
0.00 0.000001 0 2 accept4
0.00 0.000000 0 2 getsockname
0.00 0.000000 0 4 fcntl
------ ----------- ----------- --------- --------- ----------------
100.00 19.860118 5226 3800 total
gdb --batch --ex 't a a bt' -p $(cat /var/run/qemu-server/$ID.pid)
[New LWP 38495]
[New LWP 38670]
[New LWP 38671]
[New LWP 38672]
[New LWP 38673]
[New LWP 38678]
[New LWP 531848]
[New LWP 532278]
[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
0x00007ff555ef0e26 in internal_fallocate64 (fd=-137380896, offset=80, len=140735731015936) at ../sysdeps/posix/posix_fallocate64.c:36
36 ../sysdeps/posix/posix_fallocate64.c: No such file or directory.
Thread 9 (Thread 0x7ff54b2df040 (LWP 532278) "iou-wrk-38494"):
#0 0x0000000000000000 in ?? ()
Backtrace stopped: Cannot access memory at address 0x0
Thread 8 (Thread 0x7ff54b2df040 (LWP 531848) "iou-wrk-38494"):
#0 0x0000000000000000 in ?? ()
Backtrace stopped: Cannot access memory at address 0x0
Thread 7 (Thread 0x7ff32bfff700 (LWP 38678) "vnc_worker"):
#0 futex_wait_cancelable (private=0, expected=0, futex_word=0x558ff7cf6648) at ../sysdeps/nptl/futex-internal.h:186
#1 __pthread_cond_wait_common (abstime=0x0, clockid=0, mutex=0x558ff7cf6658, cond=0x558ff7cf6620) at pthread_cond_wait.c:508
#2 __pthread_cond_wait (cond=0x558ff7cf6620, mutex=0x558ff7cf6658) at pthread_cond_wait.c:638
#3 0x0000558ff511155b in ?? ()
#4 0x0000558ff4bdf5e3 in ?? ()
#5 0x0000558ff4be02a8 in ?? ()
#6 0x0000558ff5110a19 in ?? ()
#7 0x00007ff555fdeea7 in start_thread (arg=<optimized out>) at pthread_create.c:477
#8 0x00007ff555efca2f in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
Thread 6 (Thread 0x7ff33bdff700 (LWP 38673) "CPU 3/KVM"):
#0 0x00007ff555ef25f7 in preadv64v2 (fd=-137491808, vector=0x558ff4f8f817, count=0, offset=1, flags=44672) at ../sysdeps/unix/sysv/linux/preadv64v2.c:31
#1 0x0000000000000000 in ?? ()
Thread 5 (Thread 0x7ff548fff700 (LWP 38672) "CPU 2/KVM"):
#0 0x00007ff555ef25f7 in preadv64v2 (fd=-137555216, vector=0x558ff4f8f817, count=0, offset=1, flags=44672) at ../sysdeps/unix/sysv/linux/preadv64v2.c:31
#1 0x0000000000000000 in ?? ()
Thread 4 (Thread 0x7ff549bff700 (LWP 38671) "CPU 1/KVM"):
#0 0x00007ff555ef25f7 in preadv64v2 (fd=-137621056, vector=0x558ff4f8f817, count=0, offset=1, flags=44672) at ../sysdeps/unix/sysv/linux/preadv64v2.c:31
#1 0x0000000000000000 in ?? ()
Thread 3 (Thread 0x7ff54a87b700 (LWP 38670) "CPU 0/KVM"):
#0 0x00007ff555ef25f7 in preadv64v2 (fd=-137845280, vector=0x558ff4f8f817, count=0, offset=1, flags=44672) at ../sysdeps/unix/sysv/linux/preadv64v2.c:31
#1 0x0000000000000000 in ?? ()
Thread 2 (Thread 0x7ff54b17d700 (LWP 38495) "call_rcu"):
#0 0x00007ff555ef62e9 in ?? () from /lib/x86_64-linux-gnu/libc.so.6
#1 0x0000558ff5111bda in ?? ()
#2 0x0000558ff511a16a in ?? ()
#3 0x0000558ff5110a19 in ?? ()
#4 0x00007ff555fdeea7 in start_thread (arg=<optimized out>) at pthread_create.c:477
#5 0x00007ff555efca2f in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
Thread 1 (Thread 0x7ff54b2df040 (LWP 38494) "kvm"):
#0 0x00007ff555ef0e26 in internal_fallocate64 (fd=-137380896, offset=80, len=140735731015936) at ../sysdeps/posix/posix_fallocate64.c:36
#1 0x0000000000000000 in ?? ()
[Inferior 1 (process 38494) detached]
Code:
pveversion -v
proxmox-ve: 7.3-1 (running kernel: 5.19.17-1-pve)
pve-manager: 7.3-3 (running version: 7.3-3/c3928077)
pve-kernel-5.15: 7.2-14
pve-kernel-5.19: 7.2-14
pve-kernel-helper: 7.2-14
pve-kernel-5.13: 7.1-9
pve-kernel-5.19.17-1-pve: 5.19.17-1
pve-kernel-5.15.74-1-pve: 5.15.74-1
pve-kernel-5.15.53-1-pve: 5.15.53-1
pve-kernel-5.15.39-1-pve: 5.15.39-1
pve-kernel-5.13.19-6-pve: 5.13.19-15
pve-kernel-5.13.19-2-pve: 5.13.19-4
ceph-fuse: 14.2.21-1
corosync: 3.1.7-pve1
criu: 3.15-1+pve-1
glusterfs-client: 9.2-1
ifupdown: 0.8.36+pve2
ksm-control-daemon: 1.4-1
libjs-extjs: 7.0.0-1
libknet1: 1.24-pve2
libproxmox-acme-perl: 1.4.2
libproxmox-backup-qemu0: 1.3.1-1
libpve-access-control: 7.2-5
libpve-apiclient-perl: 3.2-1
libpve-common-perl: 7.3-1
libpve-guest-common-perl: 4.2-3
libpve-http-server-perl: 4.1-5
libpve-storage-perl: 7.3-1
libqb0: 1.0.5-1
libspice-server1: 0.14.3-2.1
lvm2: 2.03.11-2.1
lxc-pve: 5.0.0-3
lxcfs: 4.0.12-pve1
novnc-pve: 1.3.0-3
openvswitch-switch: 2.15.0+ds1-2+deb11u1
proxmox-backup-client: 2.3.1-1
proxmox-backup-file-restore: 2.3.1-1
proxmox-mini-journalreader: 1.3-1
proxmox-offline-mirror-helper: 0.5.0-1
proxmox-widget-toolkit: 3.5.3
pve-cluster: 7.3-1
pve-container: 4.4-2
pve-docs: 7.3-1
pve-edk2-firmware: 3.20220526-1
pve-firewall: 4.2-7
pve-firmware: 3.5-6
pve-ha-manager: 3.5.1
pve-i18n: 2.8-1
pve-qemu-kvm: 7.1.0-4
pve-xtermjs: 4.16.0-1
pve-zsync: 2.2.3
qemu-server: 7.3-1
smartmontools: 7.2-pve3
spiceterm: 3.2-2
swtpm: 0.8.0~bpo11+2
vncterm: 1.7-1
zfsutils-linux: 2.1.6-pve1
Hope it's helpfull.
Udo