Hi all
I'm running a Proxmox 1.8 cluster with DRBD, as described here: http://pve.proxmox.com/wiki/DRBD (Versions below)
I think I've found a problem with deleting virtual machines and volumes. I've been able to replicate it on a test cluster, so it's not just my system! My master proxmox host is called coquet, my second proxmox host is called loddont. Steps I took to replicate:
1. Create a new VM on coquet, with a 10Gb hard drive on the DRBD LVM volume group
2. Start the VM and live-migrate it to loddont and back to coquet again, so the LV gets marked as active on both hosts
3. Stop the VM
4. Run this on each host to find out what device mapper node the LV appears as: lvdisplay|awk '/LV Name/{n=$3} /Block device/{d=$3; sub(".*:","dm-",d); print d,n;}'
In my setup, it happens to appear as dm-8.
5. grep for that device mapper node in /proc/partitions in each host: grep dm-8 /proc/partitions
6. Delete the VM
7. Look in /proc/partitions again. You'll find that on coquet, the dm- entry has gone, but it's still there on loddont.
To see how this is a problem:
8. Create a new VM on loddont, but this time give it a 15Gb hard drive
9. Boot the VM and run a linux liveCD or something.
10. Look at dmesg or fdisk, and notice that it detects a 10Gb drive
I can fix this by either:
1) Rebooting the host with the bad mapping, or:
2) Stopping the VM and running 'dmsetup remove /dev/mapper/drbdvg-vm--106--disk--1' and then either re-activate the LV if I've create a new VM, or just remove the LV if I haven't. I don't know if these actions will cause problems though...
Can someone take a look at it please?
Thanks!
Phil
Versions:
running kernel: 2.6.35-1-pve
proxmox-ve-2.6.35: 1.8-11
pve-kernel-2.6.32-4-pve: 2.6.32-33
pve-kernel-2.6.35-1-pve: 2.6.35-11
qemu-server: 1.1-30
pve-firmware: 1.0-11
libpve-storage-perl: 1.0-17
vncterm: 0.9-2
vzctl: 3.0.28-1pve1
vzdump: 1.2-14
vzprocps: 2.0.11-2
vzquota: 3.0.11-1
pve-qemu-kvm: 0.14.1-1
ksm-control-daemon: 1.0-6
I'm running a Proxmox 1.8 cluster with DRBD, as described here: http://pve.proxmox.com/wiki/DRBD (Versions below)
I think I've found a problem with deleting virtual machines and volumes. I've been able to replicate it on a test cluster, so it's not just my system! My master proxmox host is called coquet, my second proxmox host is called loddont. Steps I took to replicate:
1. Create a new VM on coquet, with a 10Gb hard drive on the DRBD LVM volume group
2. Start the VM and live-migrate it to loddont and back to coquet again, so the LV gets marked as active on both hosts
3. Stop the VM
4. Run this on each host to find out what device mapper node the LV appears as: lvdisplay|awk '/LV Name/{n=$3} /Block device/{d=$3; sub(".*:","dm-",d); print d,n;}'
In my setup, it happens to appear as dm-8.
5. grep for that device mapper node in /proc/partitions in each host: grep dm-8 /proc/partitions
6. Delete the VM
7. Look in /proc/partitions again. You'll find that on coquet, the dm- entry has gone, but it's still there on loddont.
To see how this is a problem:
8. Create a new VM on loddont, but this time give it a 15Gb hard drive
9. Boot the VM and run a linux liveCD or something.
10. Look at dmesg or fdisk, and notice that it detects a 10Gb drive
I can fix this by either:
1) Rebooting the host with the bad mapping, or:
2) Stopping the VM and running 'dmsetup remove /dev/mapper/drbdvg-vm--106--disk--1' and then either re-activate the LV if I've create a new VM, or just remove the LV if I haven't. I don't know if these actions will cause problems though...
Can someone take a look at it please?
Thanks!
Phil
Versions:
running kernel: 2.6.35-1-pve
proxmox-ve-2.6.35: 1.8-11
pve-kernel-2.6.32-4-pve: 2.6.32-33
pve-kernel-2.6.35-1-pve: 2.6.35-11
qemu-server: 1.1-30
pve-firmware: 1.0-11
libpve-storage-perl: 1.0-17
vncterm: 0.9-2
vzctl: 3.0.28-1pve1
vzdump: 1.2-14
vzprocps: 2.0.11-2
vzquota: 3.0.11-1
pve-qemu-kvm: 0.14.1-1
ksm-control-daemon: 1.0-6