Hi there everyone,
Now here is what happened. This node was running since 2013 perfectly. I have created 8 VMs back then and they were all up, they are still.
When I consoled into the VM using VNC, the vm was stuck at boot with error "boot failed, This hard drive is not bootable"
here is the task log via the web UI.:
Creating snapshot:
Formatting '/var/lib/vz/images/105/vm-105-state-feb.raw', fmt=raw size=7017070592
TASK OK
Removing Snapshot:
TASK OK
Vm config:
root@instant:~# cat /etc/pve/qemu-server/105.conf
balloon: 2048
bootdisk: ide0
cores: 2
cpu: host
cpuunits: 100
ide0: local:105/vm-105-disk-1.qcow2,format=qcow2,cache=writeback,size=76G
ide2: none,media=cdrom
memory: 3096
name: server
net0: e1000=E6:FB:7E:BD:43:06,bridge=vmbr0
onboot: 1
ostype: l26
sockets: 1
Output of pveversion -v:
root@instant:~# pveversion -v
proxmox-ve-2.6.32: 3.2-136 (running kernel: 3.10.0-1-pve)
pve-manager: 3.3-5 (running version: 3.3-5/bfebec03)
pve-kernel-2.6.32-32-pve: 2.6.32-136
pve-kernel-2.6.32-27-pve: 2.6.32-121
pve-kernel-3.10.0-1-pve: 3.10.0-5
pve-kernel-2.6.32-28-pve: 2.6.32-124
pve-kernel-2.6.32-31-pve: 2.6.32-132
pve-kernel-2.6.32-26-pve: 2.6.32-114
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.7-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.10-1
pve-cluster: 3.0-15
qemu-server: 3.3-3
pve-firmware: 1.1-3
libpve-common-perl: 3.0-19
libpve-access-control: 3.0-15
libpve-storage-perl: 3.0-25
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-8
vzctl: 4.0-1pve6
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 2.1-10
ksm-control-daemon: 1.1-1
glusterfs-client: 3.5.2-1
kernel :
root@instant:~# uname -a
Linux instant. 3.10.0-1-pve #1 SMP Tue Dec 17 13:12:13 CET 2013 x86_64 GNU/Linux
file:
root@instant:/var/lib/vz/images/105# file vm-105-disk-1.qcow2
vm-105-disk-1.qcow2: QEMU QCOW Image (unknown version)
Fdisk:
root@instant:/var/lib/vz/images/105# fdisk -l vm-105-disk-1.qcow2
Disk vm-105-disk-1.qcow2: 66.8 GB, 66801762304 bytes
255 heads, 63 sectors/track, 8121 cylinders, total 130472192 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk vm-105-disk-1.qcow2 doesn't contain a valid partition table
ls:
root@instant:/var/lib/vz/images/105# ls -l vm-105-disk-1.qcow2
-rw-r--r-- 1 root root 66801762304 Feb 6 06:31 vm-105-disk-1.qcow2
When I did qemu-check I got thousands of errors (Leak errors etc.)
qemu-img check vm-105-disk-1.qcow2
So I used the -r all flag to fix the disk, that fixed all of the errors.
qemu-img check -r all vm-105-disk-1.qcow2
And ran the qemu-check again
root@instant:/var/lib/vz/images/105# qemu-img check vm-105-disk-1.qcow2
No errors were found on the image.
303104/1245184 = 24.34% allocated, 0.00% fragmented, 0.00% compressed clusters
Image end offset: 66800386048
Still no luck.
After that I've decided to use the qemu-nbd command to mount the disk and run fsck
Here is what happened next.
root@instant:/var/lib/vz/images/105# modprobe nbd
FATAL: Module nbd not found.
root@instant:~# lsmod | grep ndb
root@instant:~#
Though I'm able to view the LVM partitions via tetsdisk.
IDK if im on the right track or not and don't know what to do next either. Can anybody please help me.
Now here is what happened. This node was running since 2013 perfectly. I have created 8 VMs back then and they were all up, they are still.
When I consoled into the VM using VNC, the vm was stuck at boot with error "boot failed, This hard drive is not bootable"
here is the task log via the web UI.:
Creating snapshot:
Formatting '/var/lib/vz/images/105/vm-105-state-feb.raw', fmt=raw size=7017070592
TASK OK
Removing Snapshot:
TASK OK
Vm config:
root@instant:~# cat /etc/pve/qemu-server/105.conf
balloon: 2048
bootdisk: ide0
cores: 2
cpu: host
cpuunits: 100
ide0: local:105/vm-105-disk-1.qcow2,format=qcow2,cache=writeback,size=76G
ide2: none,media=cdrom
memory: 3096
name: server
net0: e1000=E6:FB:7E:BD:43:06,bridge=vmbr0
onboot: 1
ostype: l26
sockets: 1
Output of pveversion -v:
root@instant:~# pveversion -v
proxmox-ve-2.6.32: 3.2-136 (running kernel: 3.10.0-1-pve)
pve-manager: 3.3-5 (running version: 3.3-5/bfebec03)
pve-kernel-2.6.32-32-pve: 2.6.32-136
pve-kernel-2.6.32-27-pve: 2.6.32-121
pve-kernel-3.10.0-1-pve: 3.10.0-5
pve-kernel-2.6.32-28-pve: 2.6.32-124
pve-kernel-2.6.32-31-pve: 2.6.32-132
pve-kernel-2.6.32-26-pve: 2.6.32-114
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.7-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.10-1
pve-cluster: 3.0-15
qemu-server: 3.3-3
pve-firmware: 1.1-3
libpve-common-perl: 3.0-19
libpve-access-control: 3.0-15
libpve-storage-perl: 3.0-25
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-8
vzctl: 4.0-1pve6
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 2.1-10
ksm-control-daemon: 1.1-1
glusterfs-client: 3.5.2-1
kernel :
root@instant:~# uname -a
Linux instant. 3.10.0-1-pve #1 SMP Tue Dec 17 13:12:13 CET 2013 x86_64 GNU/Linux
file:
root@instant:/var/lib/vz/images/105# file vm-105-disk-1.qcow2
vm-105-disk-1.qcow2: QEMU QCOW Image (unknown version)
Fdisk:
root@instant:/var/lib/vz/images/105# fdisk -l vm-105-disk-1.qcow2
Disk vm-105-disk-1.qcow2: 66.8 GB, 66801762304 bytes
255 heads, 63 sectors/track, 8121 cylinders, total 130472192 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk vm-105-disk-1.qcow2 doesn't contain a valid partition table
ls:
root@instant:/var/lib/vz/images/105# ls -l vm-105-disk-1.qcow2
-rw-r--r-- 1 root root 66801762304 Feb 6 06:31 vm-105-disk-1.qcow2
When I did qemu-check I got thousands of errors (Leak errors etc.)
qemu-img check vm-105-disk-1.qcow2
So I used the -r all flag to fix the disk, that fixed all of the errors.
qemu-img check -r all vm-105-disk-1.qcow2
And ran the qemu-check again
root@instant:/var/lib/vz/images/105# qemu-img check vm-105-disk-1.qcow2
No errors were found on the image.
303104/1245184 = 24.34% allocated, 0.00% fragmented, 0.00% compressed clusters
Image end offset: 66800386048
Still no luck.
After that I've decided to use the qemu-nbd command to mount the disk and run fsck
Here is what happened next.
root@instant:/var/lib/vz/images/105# modprobe nbd
FATAL: Module nbd not found.
root@instant:~# lsmod | grep ndb
root@instant:~#
Though I'm able to view the LVM partitions via tetsdisk.
IDK if im on the right track or not and don't know what to do next either. Can anybody please help me.
Last edited: