[SOLVED] No solution found for migration error found stale volume copy

reswob

New Member
Apr 13, 2021
4
0
1
55
3 node proxmox cluster, 6.3-6 version


Here's my error:
Code:
2021-04-26 21:14:07 starting migration of VM 104 to node 'DC-LA' (192.168.1.22)
2021-04-26 21:14:08 found local disk 'local-lvm:base-104-disk-0' (in current VM config)
2021-04-26 21:14:08 copying local disk images
2021-04-26 21:14:09 illegal name 'base-104-disk-0' - sould be 'vm-104-*'
2021-04-26 21:14:09 command 'dd 'if=/dev/pve/base-104-disk-0' 'bs=64k'' failed: got signal 13
send/receive failed, cleaning up snapshot(s)..
2021-04-26 21:14:10 ERROR: Failed to sync data - storage migration for 'local-lvm:base-104-disk-0' to storage 'local-lvm' failed - command 'set -o pipefail && pvesm export local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=DC-LA' root@192.168.1.22 -- pvesm import local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 -allow-rename 1' failed: exit code 255
2021-04-26 21:14:10 aborting phase 1 - cleanup resources
2021-04-26 21:14:10 ERROR: found stale volume copy 'local-lvm:base-104-disk-0' on node 'DC-LA'
2021-04-26 21:14:10 ERROR: migration aborted (duration 00:00:03): Failed to sync data - storage migration for 'local-lvm:base-104-disk-0' to storage 'local-lvm' failed - command 'set -o pipefail && pvesm export local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=DC-LA' root@192.168.1.22 -- pvesm import local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 -allow-rename 1' failed: exit code 255
TASK ERROR: migration aborted

I've looked at:

https://techblog.jeppson.org/2018/03/proxmox-vm-migration-failed-found-stale-volume-copy/
https://www.reddit.com/r/Proxmox/comments/ahdo5j/unable_to_migrate_old_vms_after_setting_up_shared/
https://forum.proxmox.com/threads/migration-fails-found-stale-volume-copy.70835/ (Translated solution: Problem solved, there were still remnants of the subvolume in the / rpool directory to "rm -r / rpool / data / subvol-103-disk-0" it was possible to migrate without any problem)

and I'm looking over this: https://forum.proxmox.com/threads/how-i-can-remove-directory-entry-from-gui.50006/

But nothing is working so far.

I can't find where DC-LA thinks this VM exists.

Code:
root@DC-LA:~# qm rescan --vmid 104
rescan volumes...
Configuration file 'nodes/DC-LA/qemu-server/104.conf' does not exist


Code:
root@DC-LA:~#  pvesm list la1-lvm
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list local
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list local-lvm
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list LA-pool3
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list LA-pool4
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list ISO2storage
Volid                                                                                        Format  Type             Size VMID
ISO2storage:100/base-100-disk-0.vmdk                                                         vmdk    images    80530636800 100
ISO2storage:101/base-101-disk-0.qcow2                                                        qcow2   images    80530636800 101
ISO2storage:iso/17763.737.190906-2324.rs5_release_svc_refresh_SERVER_EVAL_x64FRE_en-us_1.iso iso     iso        5296713728
ISO2storage:iso/CentOS-7-x86_64-Minimal-1611.iso                                             iso     iso         713031680
ISO2storage:iso/ubuntu-18.04.4-live-server-amd64.iso                                         iso     iso         912261120
ISO2storage:iso/ubuntu-20.04.2.0-desktop-amd64.iso                                           iso     iso        2877227008
ISO2storage:iso/virtio-win-0.1.190.iso                                                       iso     iso         501745664
root@DC-LA:~# pvesm path 'local-lvm:base-104-disk-0'
/dev/pve/base-104-disk-0

Here is the 104.conf

Code:
boot: order=scsi0;net0
cores: 2
ide2: none,media=cdrom
memory: 8192
name: Splunk1
net0: virtio=22:A5:AA:72:53:A9,bridge=vmbr1,firewall=1
numa: 0
ostype: l26
scsi0: local-lvm:base-104-disk-0,size=75G
scsihw: virtio-scsi-pci
smbios1: uuid=b7d203fd-bb63-4274-86ba-5a39a9775b58
sockets: 4
template: 1
vmgenid: 29c555ec-8706-483f-8e1c-71d11b5e7e88

Here is my storage.cfg

Code:
dir: local
        path /var/lib/vz
        content backup,iso,vztmpl

lvmthin: local-lvm
        thinpool data
        vgname pve
        content rootdir,images

lvmthin: new-lvm
        thinpool tpool1
        vgname vol1
        content images,rootdir
        nodes DC-NYC

lvmthin: la1-lvm
        thinpool tpool2
        vgname vol2
        content images,rootdir
        nodes DC-LA

cifs: ISO2storage
        path /mnt/pve/ISO2storage
        server 192.168.50.200
        share ISO-VMS
        content images,iso
        smbversion 2.0
        username proxmox

zfspool: LA-pool3
        pool LA-pool3
        content rootdir,images
        mountpoint /LA-pool3
        nodes DC-LA
        sparse 1

zfspool: LA-pool4
        pool LA-pool4
        content images,rootdir
        mountpoint /LA-pool4
        nodes DC-LA
        sparse 1

dir: usb
        path /mnt/usb
        content iso,images
        nodes DC-NYC
        prune-backups keep-all=1
        shared 0


Any suggestions? Or pointers to something I missed?
 
Hi,
3 node proxmox cluster, 6.3-6 version


Here's my error:
Code:
2021-04-26 21:14:07 starting migration of VM 104 to node 'DC-LA' (192.168.1.22)
2021-04-26 21:14:08 found local disk 'local-lvm:base-104-disk-0' (in current VM config)
2021-04-26 21:14:08 copying local disk images
2021-04-26 21:14:09 illegal name 'base-104-disk-0' - sould be 'vm-104-*'
2021-04-26 21:14:09 command 'dd 'if=/dev/pve/base-104-disk-0' 'bs=64k'' failed: got signal 13
send/receive failed, cleaning up snapshot(s)..
2021-04-26 21:14:10 ERROR: Failed to sync data - storage migration for 'local-lvm:base-104-disk-0' to storage 'local-lvm' failed - command 'set -o pipefail && pvesm export local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=DC-LA' root@192.168.1.22 -- pvesm import local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 -allow-rename 1' failed: exit code 255
2021-04-26 21:14:10 aborting phase 1 - cleanup resources
2021-04-26 21:14:10 ERROR: found stale volume copy 'local-lvm:base-104-disk-0' on node 'DC-LA'
2021-04-26 21:14:10 ERROR: migration aborted (duration 00:00:03): Failed to sync data - storage migration for 'local-lvm:base-104-disk-0' to storage 'local-lvm' failed - command 'set -o pipefail && pvesm export local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=DC-LA' root@192.168.1.22 -- pvesm import local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 -allow-rename 1' failed: exit code 255
TASK ERROR: migration aborted
storage migration for base volumes is not yet implemented. As a workaround you can fully clone the template, migrate the clone and then convert to template on the other node.

I can't find where DC-LA thinks this VM exists.

Code:
root@DC-LA:~# qm rescan --vmid 104
rescan volumes...
Configuration file 'nodes/DC-LA/qemu-server/104.conf' does not exist
The VM should still be on the old node, because the migration failed.