Hey All,
So this problem I've been having for a little while now is sometimes I will reboot a Virtual Machine and then it will fully shut down and come back and just get stuck 9/10ths of the way on the start boot option and a lot of times if it's say a Windows update scheduled at 12 am, i'll get in the morning at 8 am and it's still there just sitting like this.
Usually a forceful stop or reset will get it working again but I have no clue why this happens.
Another time a colleague of mine rebooted their alarm server during the day at 11 am that I'm hosting on a Proxmox Ceph cluster and an hour later they said that their server never came back up, logging into the cluster I just saw this same screen and i force reset it and it came back alive.
Here's the output of my pveversion
And lastly the specific VM details:
I've removed any information that is considered sensitive from these. I'm welcome to any suggestions and potential troubleshooting steps.
So this problem I've been having for a little while now is sometimes I will reboot a Virtual Machine and then it will fully shut down and come back and just get stuck 9/10ths of the way on the start boot option and a lot of times if it's say a Windows update scheduled at 12 am, i'll get in the morning at 8 am and it's still there just sitting like this.
Usually a forceful stop or reset will get it working again but I have no clue why this happens.
Another time a colleague of mine rebooted their alarm server during the day at 11 am that I'm hosting on a Proxmox Ceph cluster and an hour later they said that their server never came back up, logging into the cluster I just saw this same screen and i force reset it and it came back alive.
Here's the output of my pveversion
Code:
pveversion -v
proxmox-ve: 8.2.0 (running kernel: 6.8.12-2-pve)
pve-manager: 8.2.6 (running version: 8.2.6/414ce79a1d42d6bc)
proxmox-kernel-helper: 8.1.0
proxmox-kernel-6.8: 6.8.12-2
proxmox-kernel-6.8.12-2-pve-signed: 6.8.12-2
ceph: 18.2.2-pve1
ceph-fuse: 18.2.2-pve1
corosync: 3.1.7-pve3
criu: 3.17.1-2
glusterfs-client: 10.3-5
ifupdown2: 3.2.0-1+pmx9
intel-microcode: 3.20231114.1~deb12u1
ksm-control-daemon: 1.5-1
libjs-extjs: 7.0.0-4
libknet1: 1.28-pve1
libproxmox-acme-perl: 1.5.1
libproxmox-backup-qemu0: 1.4.1
libproxmox-rs-perl: 0.3.4
libpve-access-control: 8.1.4
libpve-apiclient-perl: 3.3.2
libpve-cluster-api-perl: 8.0.7
libpve-cluster-perl: 8.0.7
libpve-common-perl: 8.2.2
libpve-guest-common-perl: 5.1.4
libpve-http-server-perl: 5.1.0
libpve-network-perl: 0.9.8
libpve-rs-perl: 0.8.10
libpve-storage-perl: 8.2.4
libspice-server1: 0.15.1-1
lvm2: 2.03.16-2
lxc-pve: 6.0.0-1
lxcfs: 6.0.0-pve2
novnc-pve: 1.4.0-4
proxmox-backup-client: 3.2.7-1
proxmox-backup-file-restore: 3.2.7-1
proxmox-firewall: 0.5.0
proxmox-kernel-helper: 8.1.0
proxmox-mail-forward: 0.2.3
proxmox-mini-journalreader: 1.4.0
proxmox-offline-mirror-helper: 0.6.7
proxmox-widget-toolkit: 4.2.3
pve-cluster: 8.0.7
pve-container: 5.2.0
pve-docs: 8.2.3
pve-edk2-firmware: 4.2023.08-4
pve-esxi-import-tools: 0.7.2
pve-firewall: 5.0.7
pve-firmware: 3.13-2
pve-ha-manager: 4.0.5
pve-i18n: 3.2.3
pve-qemu-kvm: 9.0.2-3
pve-xtermjs: 5.3.0-3
qemu-server: 8.2.4
smartmontools: 7.3-pve1
spiceterm: 3.3.0
swtpm: 0.8.0+pve1
vncterm: 1.8.0
zfsutils-linux: 2.2.6-pve1
Code:
ceph status
cluster:
id: 5fd31f6c-3f31-4fe2-bcc5-1f73aa608f8f
health: HEALTH_OK
services:
mon: 3 daemons, quorum A,B,C (age 44h)
mgr: A(active, since 44h), standbys: B, C
osd: 9 osds: 9 up (since 44h), 9 in (since 6w)
data:
pools: 2 pools, 129 pgs
objects: 335.86k objects, 1.3 TiB
usage: 3.7 TiB used, 4.2 TiB / 7.9 TiB avail
pgs: 129 active+clean
io:
client: 304 KiB/s rd, 432 KiB/s wr, 33 op/s rd, 48 op/s wr
And lastly the specific VM details:
Code:
cat /etc/pve/qemu-server/103.conf
agent: 1
bios: ovmf
boot: order=scsi0;ide0;net0;scsi1
cores: 4
cpu: host
efidisk0: cluster-storage:vm-103-disk-2,efitype=4m,pre-enrolled-keys=1,size=528K
machine: pc-q35-8.1
memory: 16384
meta: creation-qemu=8.1.2,ctime=1704766310
name: 103
net0: virtio=BC:24:11:8C:5A:75,bridge=vmbr0,firewall=1
numa: 1
onboot: 1
ostype: win10
scsi0: cluster-storage:vm-103-disk-0,cache=writeback,discard=on,iothread=1,serial='C-drive',size=60G,ssd=1
scsi1: cluster-storage:vm-103-disk-4,cache=writeback,discard=on,iothread=1,serial='D-drive',size=150G,ssd=1
scsi2: cluster-storage:vm-103-disk-3,cache=writeback,discard=on,iothread=1,serial='E-drive',size=150G,ssd=1
scsi3: cluster-storage:vm-103-disk-5,cache=writeback,discard=on,iothread=1,serial='F-drive',size=210G,ssd=1
scsi4: cluster-storage:vm-103-disk-6,cache=writeback,discard=on,iothread=1,serial='G-drive',size=140G,ssd=1
scsihw: virtio-scsi-single
smbios1: uuid=66c36bf2-6ee8-43d5-b62c-a34c93befbc8
sockets: 2
vmgenid: 215fc570-a1b1-402f-8c93-09eb20805feb
I've removed any information that is considered sensitive from these. I'm welcome to any suggestions and potential troubleshooting steps.