[SOLVED] No solution found for migration error found stale volume copy

reswob

New Member
Apr 13, 2021
4
0
1
54
3 node proxmox cluster, 6.3-6 version


Here's my error:
Code:
2021-04-26 21:14:07 starting migration of VM 104 to node 'DC-LA' (192.168.1.22)
2021-04-26 21:14:08 found local disk 'local-lvm:base-104-disk-0' (in current VM config)
2021-04-26 21:14:08 copying local disk images
2021-04-26 21:14:09 illegal name 'base-104-disk-0' - sould be 'vm-104-*'
2021-04-26 21:14:09 command 'dd 'if=/dev/pve/base-104-disk-0' 'bs=64k'' failed: got signal 13
send/receive failed, cleaning up snapshot(s)..
2021-04-26 21:14:10 ERROR: Failed to sync data - storage migration for 'local-lvm:base-104-disk-0' to storage 'local-lvm' failed - command 'set -o pipefail && pvesm export local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=DC-LA' root@192.168.1.22 -- pvesm import local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 -allow-rename 1' failed: exit code 255
2021-04-26 21:14:10 aborting phase 1 - cleanup resources
2021-04-26 21:14:10 ERROR: found stale volume copy 'local-lvm:base-104-disk-0' on node 'DC-LA'
2021-04-26 21:14:10 ERROR: migration aborted (duration 00:00:03): Failed to sync data - storage migration for 'local-lvm:base-104-disk-0' to storage 'local-lvm' failed - command 'set -o pipefail && pvesm export local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=DC-LA' root@192.168.1.22 -- pvesm import local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 -allow-rename 1' failed: exit code 255
TASK ERROR: migration aborted

I've looked at:

https://techblog.jeppson.org/2018/03/proxmox-vm-migration-failed-found-stale-volume-copy/
https://www.reddit.com/r/Proxmox/comments/ahdo5j/unable_to_migrate_old_vms_after_setting_up_shared/
https://forum.proxmox.com/threads/migration-fails-found-stale-volume-copy.70835/ (Translated solution: Problem solved, there were still remnants of the subvolume in the / rpool directory to "rm -r / rpool / data / subvol-103-disk-0" it was possible to migrate without any problem)

and I'm looking over this: https://forum.proxmox.com/threads/how-i-can-remove-directory-entry-from-gui.50006/

But nothing is working so far.

I can't find where DC-LA thinks this VM exists.

Code:
root@DC-LA:~# qm rescan --vmid 104
rescan volumes...
Configuration file 'nodes/DC-LA/qemu-server/104.conf' does not exist


Code:
root@DC-LA:~#  pvesm list la1-lvm
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list local
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list local-lvm
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list LA-pool3
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list LA-pool4
Volid Format  Type      Size VMID
root@DC-LA:~#  pvesm list ISO2storage
Volid                                                                                        Format  Type             Size VMID
ISO2storage:100/base-100-disk-0.vmdk                                                         vmdk    images    80530636800 100
ISO2storage:101/base-101-disk-0.qcow2                                                        qcow2   images    80530636800 101
ISO2storage:iso/17763.737.190906-2324.rs5_release_svc_refresh_SERVER_EVAL_x64FRE_en-us_1.iso iso     iso        5296713728
ISO2storage:iso/CentOS-7-x86_64-Minimal-1611.iso                                             iso     iso         713031680
ISO2storage:iso/ubuntu-18.04.4-live-server-amd64.iso                                         iso     iso         912261120
ISO2storage:iso/ubuntu-20.04.2.0-desktop-amd64.iso                                           iso     iso        2877227008
ISO2storage:iso/virtio-win-0.1.190.iso                                                       iso     iso         501745664
root@DC-LA:~# pvesm path 'local-lvm:base-104-disk-0'
/dev/pve/base-104-disk-0

Here is the 104.conf

Code:
boot: order=scsi0;net0
cores: 2
ide2: none,media=cdrom
memory: 8192
name: Splunk1
net0: virtio=22:A5:AA:72:53:A9,bridge=vmbr1,firewall=1
numa: 0
ostype: l26
scsi0: local-lvm:base-104-disk-0,size=75G
scsihw: virtio-scsi-pci
smbios1: uuid=b7d203fd-bb63-4274-86ba-5a39a9775b58
sockets: 4
template: 1
vmgenid: 29c555ec-8706-483f-8e1c-71d11b5e7e88

Here is my storage.cfg

Code:
dir: local
        path /var/lib/vz
        content backup,iso,vztmpl

lvmthin: local-lvm
        thinpool data
        vgname pve
        content rootdir,images

lvmthin: new-lvm
        thinpool tpool1
        vgname vol1
        content images,rootdir
        nodes DC-NYC

lvmthin: la1-lvm
        thinpool tpool2
        vgname vol2
        content images,rootdir
        nodes DC-LA

cifs: ISO2storage
        path /mnt/pve/ISO2storage
        server 192.168.50.200
        share ISO-VMS
        content images,iso
        smbversion 2.0
        username proxmox

zfspool: LA-pool3
        pool LA-pool3
        content rootdir,images
        mountpoint /LA-pool3
        nodes DC-LA
        sparse 1

zfspool: LA-pool4
        pool LA-pool4
        content images,rootdir
        mountpoint /LA-pool4
        nodes DC-LA
        sparse 1

dir: usb
        path /mnt/usb
        content iso,images
        nodes DC-NYC
        prune-backups keep-all=1
        shared 0


Any suggestions? Or pointers to something I missed?
 
Hi,
3 node proxmox cluster, 6.3-6 version


Here's my error:
Code:
2021-04-26 21:14:07 starting migration of VM 104 to node 'DC-LA' (192.168.1.22)
2021-04-26 21:14:08 found local disk 'local-lvm:base-104-disk-0' (in current VM config)
2021-04-26 21:14:08 copying local disk images
2021-04-26 21:14:09 illegal name 'base-104-disk-0' - sould be 'vm-104-*'
2021-04-26 21:14:09 command 'dd 'if=/dev/pve/base-104-disk-0' 'bs=64k'' failed: got signal 13
send/receive failed, cleaning up snapshot(s)..
2021-04-26 21:14:10 ERROR: Failed to sync data - storage migration for 'local-lvm:base-104-disk-0' to storage 'local-lvm' failed - command 'set -o pipefail && pvesm export local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=DC-LA' root@192.168.1.22 -- pvesm import local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 -allow-rename 1' failed: exit code 255
2021-04-26 21:14:10 aborting phase 1 - cleanup resources
2021-04-26 21:14:10 ERROR: found stale volume copy 'local-lvm:base-104-disk-0' on node 'DC-LA'
2021-04-26 21:14:10 ERROR: migration aborted (duration 00:00:03): Failed to sync data - storage migration for 'local-lvm:base-104-disk-0' to storage 'local-lvm' failed - command 'set -o pipefail && pvesm export local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 | /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=DC-LA' root@192.168.1.22 -- pvesm import local-lvm:base-104-disk-0 raw+size - -with-snapshots 0 -allow-rename 1' failed: exit code 255
TASK ERROR: migration aborted
storage migration for base volumes is not yet implemented. As a workaround you can fully clone the template, migrate the clone and then convert to template on the other node.

I can't find where DC-LA thinks this VM exists.

Code:
root@DC-LA:~# qm rescan --vmid 104
rescan volumes...
Configuration file 'nodes/DC-LA/qemu-server/104.conf' does not exist
The VM should still be on the old node, because the migration failed.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!