ERROR: Clone failed

rednet

Member
Jul 21, 2020
3
0
21
Argentina
I have an error when clone a VM

create full clone of drive virtio0 (datos:vm-1001-disk-0)
Logical volume "vm-11002-disk-0" created.
transferred: 0 bytes remaining: 42949672960 bytes total: 42949672960 bytes progression: 0.00 %
transferred: 429496729 bytes remaining: 42520176231 bytes total: 42949672960 bytes progression: 1.00 %
transferred: 858993459 bytes remaining: 42090679501 bytes total: 42949672960 bytes progression: 2.00 %
.....
transferred: 42133629173 bytes remaining: 816043787 bytes total: 42949672960 bytes progression: 98.10 %
transferred: 42563125903 bytes remaining: 386547057 bytes total: 42949672960 bytes progression: 99.10 %
transferred: 42949672960 bytes remaining: 0 bytes total: 42949672960 bytes progression: 100.00 %
transferred: 42949672960 bytes remaining: 0 bytes total: 42949672960 bytes progression: 100.00 %
WARNING: Device /dev/dm-13 not initialized in udev database even after waiting 10000000 microseconds.
WARNING: Device /dev/dm-13 not initialized in udev database even after waiting 10000000 microseconds.
Logical volume "vm-11002-disk-0" successfully removed
WARNING: Device /dev/dm-13 not initialized in udev database even after waiting 10000000 microseconds.
TASK ERROR: clone failed: command '/sbin/lvs --separator : --noheadings --units b --unbuffered --nosuffix --options lv_size /dev/hdd1/vm-11002-disk-0' failed: got timeout

And I see this error on syslog:
Oct 06 21:03:49 pve220 kernel: INFO: task systemd-udevd:5718 blocked for more than 120 seconds.
Oct 06 21:03:49 pve220 kernel: Tainted: P O 5.4.65-1-pve #1
Oct 06 21:03:49 pve220 kernel: "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
Oct 06 21:03:49 pve220 kernel: systemd-udevd D 0 5718 983 0x00000324
Oct 06 21:03:49 pve220 kernel: Call Trace:
Oct 06 21:03:49 pve220 kernel: __schedule+0x2e6/0x6f0
Oct 06 21:03:49 pve220 kernel: ? fsnotify_grab_connector+0x4e/0x90
Oct 06 21:03:49 pve220 kernel: schedule+0x33/0xa0
Oct 06 21:03:49 pve220 kernel: schedule_preempt_disabled+0xe/0x10
Oct 06 21:03:49 pve220 kernel: __mutex_lock.isra.10+0x2c9/0x4c0
Oct 06 21:03:49 pve220 kernel: ? exact_lock+0x11/0x20
Oct 06 21:03:49 pve220 kernel: ? disk_map_sector_rcu+0x70/0x70
Oct 06 21:03:49 pve220 kernel: __mutex_lock_slowpath+0x13/0x20
Oct 06 21:03:49 pve220 kernel: mutex_lock+0x2c/0x30
Oct 06 21:03:49 pve220 kernel: __blkdev_get+0x7a/0x560
Oct 06 21:03:49 pve220 kernel: blkdev_get+0xef/0x150
Oct 06 21:03:49 pve220 kernel: ? blkdev_get_by_dev+0x50/0x50
Oct 06 21:03:49 pve220 kernel: blkdev_open+0x87/0xa0
Oct 06 21:03:49 pve220 kernel: do_dentry_open+0x143/0x3a0
Oct 06 21:03:49 pve220 kernel: vfs_open+0x2d/0x30
Oct 06 21:03:49 pve220 kernel: path_openat+0x2e9/0x16f0
Oct 06 21:03:49 pve220 kernel: ? unlock_page_memcg+0x12/0x20
Oct 06 21:03:49 pve220 kernel: ? page_add_file_rmap+0x131/0x190
Oct 06 21:03:49 pve220 kernel: ? wp_page_copy+0x37b/0x750
Oct 06 21:03:49 pve220 kernel: do_filp_open+0x93/0x100
Oct 06 21:03:49 pve220 kernel: ? __alloc_fd+0x46/0x150
Oct 06 21:03:49 pve220 kernel: do_sys_open+0x177/0x280
Oct 06 21:03:49 pve220 kernel: __x64_sys_openat+0x20/0x30
Oct 06 21:03:49 pve220 kernel: do_syscall_64+0x57/0x190
Oct 06 21:03:49 pve220 kernel: entry_SYSCALL_64_after_hwframe+0x44/0xa9
Oct 06 21:03:49 pve220 kernel: RIP: 0033:0x7feaefdac1ae
Oct 06 21:03:49 pve220 kernel: Code: Bad RIP value.
Oct 06 21:03:49 pve220 kernel: RSP: 002b:00007ffee3a2c980 EFLAGS: 00000246 ORIG_RAX: 0000000000000101
Oct 06 21:03:49 pve220 kernel: RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007feaefdac1ae
Oct 06 21:03:49 pve220 kernel: RDX: 0000000000080000 RSI: 0000562316fc9b60 RDI: 00000000ffffff9c
Oct 06 21:03:49 pve220 kernel: RBP: 00007feaef5cbc60 R08: 000056231533e270 R09: 000000000000000f
Oct 06 21:03:49 pve220 kernel: R10: 0000000000000000 R11: 0000000000000246 R12: 00000000ffffffff
Oct 06 21:03:49 pve220 kernel: R13: 0000000000000000 R14: 0000000000000000 R15: 0000562316fd09b0
Oct 06 21:04:00 pve220 systemd[1]: Starting Proxmox VE replication runner...
Oct 06 21:04:01 pve220 systemd[1]: pvesr.service: Succeeded.

The server is new, not run any VM yet, the disk (LVM) are OK.

Version Running 6.2-12
proxmox-ve: 6.2-2 (running kernel: 5.4.65-1-pve)
pve-manager: 6.2-12 (running version: 6.2-12/b287dd27)
pve-kernel-5.4: 6.2-7
pve-kernel-helper: 6.2-7
pve-kernel-5.4.65-1-pve: 5.4.65-1
pve-kernel-5.4.60-1-pve: 5.4.60-2
pve-kernel-5.4.55-1-pve: 5.4.55-1
pve-kernel-5.4.44-2-pve: 5.4.44-2
pve-kernel-5.4.34-1-pve: 5.4.34-2
ceph-fuse: 12.2.11+dfsg1-2.1+b1
corosync: 3.0.4-pve1
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: not correctly installed
ifupdown2: 3.0.0-1+pve3
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.16-pve1
libproxmox-acme-perl: 1.0.5
libpve-access-control: 6.1-2
libpve-apiclient-perl: 3.0-3
libpve-common-perl: 6.2-2
libpve-guest-common-perl: 3.1-3
libpve-http-server-perl: 3.0-6
libpve-storage-perl: 6.2-8
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve4
lxc-pve: 4.0.3-1
lxcfs: 4.0.3-pve3
novnc-pve: 1.1.0-1
proxmox-backup-client: 0.9.0-2
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.2-12
pve-cluster: 6.1-8
pve-container: 3.2-2
pve-docs: 6.2-6
pve-edk2-firmware: 2.20200531-1
pve-firewall: 4.1-3
pve-firmware: 3.1-3
pve-ha-manager: 3.1-1
pve-i18n: 2.2-1
pve-qemu-kvm: 5.1.0-2
pve-xtermjs: 4.7.0-2
qemu-server: 6.2-14
smartmontools: 7.1-pve2
spiceterm: 3.1-1
vncterm: 1.6-2
zfsutils-linux: 0.8.4-pve1

I need a solution as soon as posible, thanks
 
Please provide some details about the new server. CPU/Mainboard/RAID controller/DIsks.
 
Lenovo ThinkSystem SR570 2 Xeon Silver 4214 12C 85W 2.2GHz with mem 208Gb
Adapter: RAID 530-8i PCIe 12Gb
raid1 with 2 Intel S4610 480GB Mainstream SATA 6Gb Hot Swap SSD
raid10 with 6 Lenovo 1.2TB 10K SAS 12Gb Hot Swap 512n HDD

NOTE: For your information, before the error I had already done a successful clone and do upgrade the system.
 
Last edited:
I have disabled 'obtain_device_list_from_udev' and the problem has been fixed.

But the only thing I see when I do a clone is this:

INFO: task qemu-img:11907 blocked for more than 120 seconds.
Tainted: P O 5.4.65-1-pve #1
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
qemu-img D 0 11907 11555 0x00000000
Call Trace:
__schedule+0x2e6/0x6f0
? bit_wait_timeout+0xa0/0xa0
schedule+0x33/0xa0
io_schedule+0x16/0x40
bit_wait_io+0x11/0x50
__wait_on_bit+0x33/0xa0
out_of_line_wait_on_bit+0x90/0xb0
? var_wake_function+0x30/0x30
__block_write_begin_int+0x22c/0x620
? blkdev_direct_IO+0x490/0x490
? blkdev_direct_IO+0x490/0x490
block_write_begin+0x4d/0xf0
blkdev_write_begin+0x23/0x30
generic_perform_write+0xf2/0x1b0
__generic_file_write_iter+0x101/0x1f0
blkdev_write_iter+0xae/0x140
new_sync_write+0x125/0x1c0
__vfs_write+0x29/0x40
vfs_write+0xab/0x1b0
ksys_pwrite64+0x66/0xa0
__x64_sys_pwrite64+0x1e/0x20
do_syscall_64+0x57/0x190
entry_SYSCALL_64_after_hwframe+0x44/0xa9
RIP: 0033:0x7f3558c1eedf
Code: Bad RIP value.
RSP: 002b:00007f3542bfa740 EFLAGS: 00000293 ORIG_RAX: 0000000000000012
RAX: ffffffffffffffda RBX: 000000000000000a RCX: 00007f3558c1eedf
RDX: 0000000000200000 RSI: 00007f3548de6000 RDI: 000000000000000a
RBP: 00007f3548de6000 R08: 0000000000000000 R09: 00000000ffffffff
R10: 00000004887fee00 R11: 0000000000000293 R12: 0000000000200000
R13: 00000004887fee00 R14: 00007f354c0227a0 R15: 00007f3542c02010

Sorry for my delay, because I did several tests.