[SOLVED] Migration Issue

prova9

New Member
Jul 4, 2024
4
0
1
Hello everyone,
today I experienced a very strange and, so far, hard to solve problem with my proxmox cluster.
I had a power outage and after everything went back online I found one of my VM (the most important one, of course) to be splitted in the 2 nodes…the VM itself is in node 2 and the VM disk still in node 1.
I can’t power up the VM, I can’t migrate it and I can’t replicate it…
I really don’t know what to do next…

Thanks for the help
 
Hi @prova9 , welcome to forum.

At the very least you need to provide:
- pveversion
- qm config [vmid]
- pvesm status
- cat /etc/pve/storage.cfg
- pvecm status
- ls -al /etc/pve/nodes/[node-1]
- ls -al /etc/pve/nodes/[node-2]


Blockbridge : Ultra low latency all-NVME shared storage for Proxmox - https://www.blockbridge.com/proxmox
-pve version: 8.2.2

-qm config:
agent: 1
balloon: 12400
boot: order=sata0
cores: 8
cpu: host
description: sata0%3A local-zfs%3Avm-500-disk-0,format=raw,size=300G
memory: 16800
meta: creation-qemu=8.1.5,ctime=1718628923
name: MediaServer
net0: virtio=<omissis>,bridge=vmbr1,tag=100
numa: 0
ostype: l26
scsihw: virtio-scsi-single
smbios1: uuid=<omissis>
sockets: 1
tags: media;server
vmgenid: <omissis>

-pvesm status:
Name Type Status Total Used Available %

BackupPool pbs active 3771132928 365795456 3405337472 9.70%

UnraidShare nfs active 7665156096 3444599808 4220556288 44.94%

local dir active 853470336 128 853470208 0.00%

local-zfs zfspool active 940220956 86750684 853470272 9.23%

- cat /etc/pve/storage.cfg
dir: local

path /var/lib/vz

content iso,backup,vztmpl



zfspool: local-zfs

pool rpool/data

content rootdir,images

sparse 1



pbs: BackupPool

datastore BackupPool

server 10.10.99.105

content backup

encryption-key <omissis>

fingerprint <omissis>

prune-backups keep-all=1

username root@pam



nfs: UnraidShare

export /mnt/user/ProxShares

path /mnt/pve/UnraidShare

server 10.10.10.203

content backup,vztmpl,iso,rootdir,snippets,images

prune-backups keep-all=1

- pvecm status:
Cluster information

-------------------

Name: ClusterNet

Config Version: 3

Transport: knet

Secure auth: on



Quorum information

------------------

Date: Thu Jul 4 20:17:22 2024

Quorum provider: corosync_votequorum

Nodes: 2

Node ID: 0x00000002

Ring ID: 1.112

Quorate: Yes



Votequorum information

----------------------

Expected votes: 3

Highest expected: 3

Total votes: 3

Quorum: 2

Flags: Quorate Qdevice



Membership information

----------------------

Nodeid Votes Qdevice Name

0x00000001 1 A,V,NMW 10.10.99.101

0x00000002 1 A,V,NMW 10.10.99.103 (local)

0x00000000 1 Qdevice

- ls -al /etc/pve/nodes/pve-03
total 4

drwxr-xr-x 2 root www-data 0 May 7 00:21 .

drwxr-xr-x 2 root www-data 0 May 4 15:08 ..

-rw-r----- 1 root www-data 130 May 25 13:56 config

-rw-r----- 1 root www-data 83 Jul 4 20:20 lrm_status

drwxr-xr-x 2 root www-data 0 May 7 00:21 lxc

drwxr-xr-x 2 root www-data 0 May 7 00:21 openvz

drwx------ 2 root www-data 0 May 7 00:21 priv

-rw-r----- 1 root www-data 3272 May 25 14:00 pveproxy-ssl.key

-rw-r----- 1 root www-data 3981 May 25 14:00 pveproxy-ssl.pem

-rw-r----- 1 root www-data 1708 May 7 00:21 pve-ssl.key

-rw-r----- 1 root www-data 1797 May 7 00:21 pve-ssl.pem

drwxr-xr-x 2 root www-data 0 May 7 00:21 qemu-server

-rw-r----- 1 root www-data 559 Jul 4 18:43 ssh_known_hosts

- ls -al /etc/pve/nodes/pve-01
total 4

drwxr-xr-x 2 root www-data 0 May 4 15:08 .

drwxr-xr-x 2 root www-data 0 May 4 15:08 ..

-rw-r----- 1 root www-data 130 May 25 13:56 config

-rw-r----- 1 root www-data 22 May 25 02:05 host.fw

-rw-r----- 1 root www-data 156 Jul 4 20:19 lrm_status

drwxr-xr-x 2 root www-data 0 May 4 15:08 lxc

drwxr-xr-x 2 root www-data 0 May 4 15:08 openvz

drwx------ 2 root www-data 0 May 4 15:08 priv

-rw-r----- 1 root www-data 3272 May 25 13:57 pveproxy-ssl.key

-rw-r----- 1 root www-data 3981 May 25 13:57 pveproxy-ssl.pem

-rw-r----- 1 root www-data 1704 May 4 15:08 pve-ssl.key

-rw-r----- 1 root www-data 1797 May 4 15:08 pve-ssl.pem

drwxr-xr-x 2 root www-data 0 May 4 15:08 qemu-server

-rw-r----- 1 root www-data 559 Jul 4 12:34 ssh_known_hosts
 
pve-03 is the node where the VM is and pve-01 is the node where the VM disk is.
No idea why …
All the previous commands are from pve-03
 
At the end I solved it… Was just a matter of copying the VM’s configuration on the node where the VM disk was and everything was back on line