For the past 2 months we've hit this all VMs freezing couple times.That regression came in with 5.2-1 (the first 5.2 release) and was fixed with 5.2-4, so all versions in-between were affected.
Why do you ask?
May 10 22:03:04 pm-pm1.pminner.cluster.eu pvestatd[3545]: VM 82075205 qmp command failed - VM 82075205 qmp command 'query-proxmox-support' failed - unable t
o connect to VM 82075205 qmp socket - timeout after 31 retries
May 10 22:03:07 pm-pm1.pminner.cluster.eu pvestatd[3545]: VM 82075202 qmp command failed - VM 82075202 qmp command 'query-proxmox-support' failed - unable t
o connect to VM 82075202 qmp socket - timeout after 31 retries
May 10 22:03:10 pm-pm1.pminner.cluster.eu pvestatd[3545]: VM 82076201 qmp command failed - VM 82076201 qmp command 'query-proxmox-support' failed - unable t
o connect to VM 82076201 qmp socket - timeout after 31 retries
# dpkg -l|grep pve-qemu-kvm
ii pve-qemu-kvm 5.2.0-3 amd64 Full virtualization on x86 hardware
Jan 20 21:25:05 brabus pvestatd[2071]: VM 1271 qmp command failed - VM 1271 qmp command 'query-proxmox-support' failed - unable to connect to VM 1271 qmp socket - timeout after 31 retries
Jan 20 21:25:05 brabus pvedaemon[2025903]: <root@pam> successful auth for user 'apiproxmox@pve'
Jan 20 21:25:05 brabus pvestatd[2071]: status update time (6.169 seconds)
Jan 20 21:25:15 brabus pvestatd[2071]: VM 1271 qmp command failed - VM 1271 qmp command 'query-proxmox-support' failed - unable to connect to VM 1271 qmp socket - timeout after 31 retries
Jan 20 21:25:15 brabus pvestatd[2071]: status update time (6.151 seconds)
Jan 20 21:25:20 brabus pvedaemon[2025903]: <root@pam> successful auth for user 'apiproxmox@pve'
Jan 20 21:25:25 brabus pvestatd[2071]: VM 1271 qmp command failed - VM 1271 qmp command 'query-proxmox-support' failed - unable to connect to VM 1271 qmp socket - timeout after 31 retries
Jan 20 21:25:26 brabus pvestatd[2071]: status update time (6.166 seconds)
Jan 20 21:25:28 brabus pvedaemon[2026132]: <root@pam> successful auth for user 'zhitomirskiy@hms'
Jan 20 21:25:35 brabus pvestatd[2071]: VM 1271 qmp command failed - VM 1271 qmp command 'query-proxmox-support' failed - unable to connect to VM 1271 qmp socket - timeout after 31 retries
Jan 20 21:25:35 brabus pvestatd[2071]: status update time (6.152 seconds)
Jan 20 21:25:35 brabus pvedaemon[2025903]: <root@pam> successful auth for user 'apiproxmox@pve'
Jan 20 21:25:45 brabus pvestatd[2071]: VM 1271 qmp command failed - VM 1271 qmp command 'query-proxmox-support' failed - unable to connect to VM 1271 qmp socket - timeout after 31 retries
At this moment in Proxmox GUI where is high DISK IO in Dashboard on All VMs and not Host, but doesn't responds only VM one with log records VM 1271 qmp command 'query-proxmox-support' failed - unable to connect to VM 1271 qmp socket - timeout after 31 retries
Can you please also post the config of an affected VM (up .. i have the same error.
qm config VMID
), thanks!YesCan you please also post the config of an affected VM (qm config VMID
), thanks!
balloon: 256
boot: c
bootdisk: sata0
cores: 2
hotplug: disk,network,usb
memory: 1024
name: XPenology-Data
net0: e1000=00:00:00:00:00:01,bridge=vmbr0,link_down=1
net1: e1000=DA:53:41:29:A6:98,bridge=vmbr0
numa: 0
onboot: 1
ostype: l26
sata0: SSD-LVM:vm-1271-disk-0,size=52M
sata1: SSD-LVM:vm-1271-disk-1,cache=writeback,size=6G
sata2: HDD-NFS:1271/vm-1271-disk-0.raw,cache=writeback,size=250G
scsihw: virtio-scsi-pci
serial0: socket
smbios1: uuid=e545f02a-16ce-4d20-bf21-838a98f7c4d0
sockets: 1
startup: order=71,up=5
tablet: 0
So NFS then, maybe even to theBy the way - I'm not using Proxmox Backup Server.
HDD-NFS
storage? Can you please also post the output of pvesm status
?You are right, NFS -> HDD-NFSSo NFS then, maybe even to theHDD-NFS
storage? Can you please also post the output ofpvesm status
?
root@brabus:~# pvesm status
Name Type Status Total Used Available %
Backup dir active 7751366384 1854768748 5505879976 23.93%
HDD-NFS nfs active 3221225472 505371648 2715853824 15.69%
SSD-LVM lvm active 234426368 120905728 113520640 51.58%
Template dir active 7751366384 1854768748 5505879976 23.93%
local dir disabled 0 0 0 N/A
root@crocus:~# pvesm status
Name Type Status Total Used Available %
Backup nfs active 7751366656 1854768128 5505880064 23.93%
HDD-NFS nfs active 3221225472 505371648 2715853824 15.69%
HDD-ZFS zfspool active 13696708589 3039088118 10657620471 22.19%
SSD-LVM lvm active 31436800 19398656 12038144 61.71%
Template nfs active 7751366656 1854768128 5505880064 23.93%
local dir disabled 0 0 0 N/A
root@crocus:~# zpool status
pool: zfs
state: ONLINE
scan: scrub repaired 0B in 03:38:02 with 0 errors on Sun Jan 9 04:02:03 2022
config:
NAME STATE READ WRITE CKSUM
zfs ONLINE 0 0 0
raidz1-0 ONLINE 0 0 0
wwn-0x5000c500b48a21ab-part1 ONLINE 0 0 0
wwn-0x5000c500b48ce0e7-part1 ONLINE 0 0 0
wwn-0x5000c500b4a09e92-part1 ONLINE 0 0 0
wwn-0x5000c500b4a3fd5f-part1 ONLINE 0 0 0
wwn-0x5000c500b4a5c689-part1 ONLINE 0 0 0
errors: No known data errors
root@crocus:~# zfs list
NAME USED AVAIL REFER MOUNTPOINT
zfs 11.6T 9.93T 192K /zfs
zfs/acrus-logs 3.11M 9.93T 3.11M /zfs/acrus-logs
zfs/andy 3.60G 9.93T 3.60G /zfs/andy
zfs/elacrus 9.00G 1015G 9.00G /zfs/elacrus
zfs/iscsi 8.16T 18.1T 9.61G -
zfs/nfs 482G 2.53T 482G /zfs/nfs
zfs/prs 62.9G 437G 62.9G /zfs/prs
zfs/public 49.9G 9.93T 49.9G /zfs/public
zfs/tmp 50.3M 1024G 50.3M /zfs/tmp
zfs/zfs 2.83T 9.93T 179K /zfs/zfs
zfs/zfs/subvol-1320-disk-0 188G 312G 188G /zfs/zfs/subvol-1320-disk-0
zfs/zfs/subvol-1320-disk-1 2.64T 9.36T 2.64T /zfs/zfs/subvol-1320-disk-1
zfs/zfs/subvol-1320-disk-2 3.23G 96.8G 3.23G /zfs/zfs/subvol-1320-disk-2
zfs/zfs/subvol-1320-disk-3 762M 99.3G 762M /zfs/zfs/subvol-1320-disk-3