proxmox unstable when performing restore

gergogyerek

Member
Jul 7, 2020
10
1
21
46
Dear Community,
I tried to restore a backup (container) and it always fails and the whole proxmox interface gets unresponsive.
Could you please suggest which logs should I check?
thanks
Gergo
 

Attachments

  • proxmox_unresponsive.PNG
    proxmox_unresponsive.PNG
    41.1 KB · Views: 12
Hello,

please post output of pveversion -v and container config that restored as well

the syslog, journalctl and dmesg are always helpful ;)
 
hi,
please find below the requested info.


Bash:
root@proxmox:~# pveversion -v
proxmox-ve: 6.2-1 (running kernel: 5.4.60-1-pve)
pve-manager: 6.2-11 (running version: 6.2-11/22fb4983)
pve-kernel-5.4: 6.2-6
pve-kernel-helper: 6.2-6
pve-kernel-5.4.60-1-pve: 5.4.60-2
pve-kernel-5.4.44-2-pve: 5.4.44-2
pve-kernel-5.4.34-1-pve: 5.4.34-2
ceph-fuse: 12.2.11+dfsg1-2.1+b1
corosync: 3.0.4-pve1
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: 0.8.35+pve1
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.16-pve1
libproxmox-acme-perl: 1.0.5
libpve-access-control: 6.1-2
libpve-apiclient-perl: 3.0-3
libpve-common-perl: 6.2-2
libpve-guest-common-perl: 3.1-3
libpve-http-server-perl: 3.0-6
libpve-storage-perl: 6.2-6
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve4
lxc-pve: 4.0.3-1
lxcfs: 4.0.3-pve3
novnc-pve: 1.1.0-1
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.2-12
pve-cluster: 6.1-8
pve-container: 3.2-1
pve-docs: 6.2-5
pve-edk2-firmware: 2.20200531-1
pve-firewall: 4.1-2
pve-firmware: 3.1-3
pve-ha-manager: 3.1-1
pve-i18n: 2.2-1
pve-qemu-kvm: 5.1.0-1
pve-xtermjs: 4.7.0-2
qemu-server: 6.2-14
smartmontools: 7.1-pve2
spiceterm: 3.1-1
vncterm: 1.6-2
zfsutils-linux: 0.8.4-pve1
root@proxmox:~#

and apologies, it is not a container, it is a vm. sorry.
here is the config

Bash:
root@proxmox:/etc/pve/qemu-server# cat 103.conf
agent: 1
bios: ovmf
bootdisk: sata0
efidisk0: local-lvm:vm-103-disk-0,size=4M
name: hassOS-4.11
net0: virtio=DE:51:4B:22:B2:A1,bridge=vmbr0
onboot: 1
ostype: l26
sata0: local-lvm:vm-103-disk-1,size=6G
scsihw: virtio-scsi-pci
smbios1: uuid=11dc55cf-7076-451f-a99b-a9022ea5d800
vmgenid: 266cefd6-7152-46eb-88e0-10248fc72841
root@proxmox:/etc/pve/qemu-server#
 
Bash:
restore vma archive: zstd -q -d -c /mnt/picture/dump/vzdump-qemu-103-2020_09_05-02_33_16.vma.zst | vma extract -v -r /var/tmp/vzdumptmp51579.fifo - /var/tmp/vzdumptmp51579
CFG: size: 418 name: qemu-server.conf
DEV: dev_id=1 size: 131072 devname: drive-efidisk0
DEV: dev_id=2 size: 6442450944 devname: drive-sata0
CTIME: Sat Sep  5 02:33:17 2020
  Logical volume "vm-103-disk-1" successfully removed
  Logical volume "vm-103-disk-0" successfully removed
  Rounding up size to full physical extent 4.00 MiB
  WARNING: You have not turned on protection against thin pools running out of space.
  WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full.
  Logical volume "vm-103-disk-0" created.
  WARNING: Sum of all thin volume sizes (152.00 GiB) exceeds the size of thin pool pve/data and the amount of free space in volume group (16.00 GiB).
new volume ID is 'local-lvm:vm-103-disk-0'
  WARNING: You have not turned on protection against thin pools running out of space.
  WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full.
  Logical volume "vm-103-disk-1" created.
  WARNING: Sum of all thin volume sizes (158.00 GiB) exceeds the size of thin pool pve/data and the amount of free space in volume group (16.00 GiB).
new volume ID is 'local-lvm:vm-103-disk-1'
map 'drive-efidisk0' to '/dev/pve/vm-103-disk-0' (write zeros = 0)
map 'drive-sata0' to '/dev/pve/vm-103-disk-1' (write zeros = 0)
progress 1% (read 64487424 bytes, duration 1 sec)
progress 2% (read 128909312 bytes, duration 1 sec)
progress 3% (read 193331200 bytes, duration 1 sec)
progress 4% (read 257753088 bytes, duration 1 sec)
progress 5% (read 322174976 bytes, duration 1 sec)
progress 6% (read 386596864 bytes, duration 1 sec)
progress 7% (read 451018752 bytes, duration 2 sec)
progress 8% (read 515440640 bytes, duration 2 sec)
progress 9% (read 579862528 bytes, duration 2 sec)
progress 10% (read 644284416 bytes, duration 2 sec)
progress 11% (read 708706304 bytes, duration 2 sec)
progress 12% (read 773128192 bytes, duration 2 sec)
progress 13% (read 837550080 bytes, duration 2 sec)
progress 14% (read 901971968 bytes, duration 2 sec)
progress 15% (read 966393856 bytes, duration 3 sec)
progress 16% (read 1030815744 bytes, duration 4 sec)
progress 17% (read 1095303168 bytes, duration 4 sec)
progress 18% (read 1159725056 bytes, duration 4 sec)
progress 19% (read 1224146944 bytes, duration 5 sec)
progress 20% (read 1288568832 bytes, duration 5 sec)
progress 21% (read 1352990720 bytes, duration 5 sec)
progress 22% (read 1417412608 bytes, duration 6 sec)
progress 23% (read 1481834496 bytes, duration 6 sec)
progress 24% (read 1546256384 bytes, duration 6 sec)
progress 25% (read 1610678272 bytes, duration 7 sec)
progress 26% (read 1675100160 bytes, duration 7 sec)
progress 27% (read 1739522048 bytes, duration 7 sec)
progress 28% (read 1803943936 bytes, duration 7 sec)
progress 29% (read 1868365824 bytes, duration 8 sec)
progress 30% (read 1932787712 bytes, duration 8 sec)
progress 31% (read 1997209600 bytes, duration 9 sec)
progress 32% (read 2061631488 bytes, duration 9 sec)
progress 33% (read 2126053376 bytes, duration 9 sec)
progress 34% (read 2190540800 bytes, duration 10 sec)
progress 35% (read 2254962688 bytes, duration 10 sec)
progress 36% (read 2319384576 bytes, duration 10 sec)
progress 37% (read 2383806464 bytes, duration 10 sec)
progress 38% (read 2448228352 bytes, duration 11 sec)
progress 39% (read 2512650240 bytes, duration 11 sec)
progress 40% (read 2577072128 bytes, duration 11 sec)
progress 41% (read 2641494016 bytes, duration 11 sec)
progress 42% (read 2705915904 bytes, duration 12 sec)
progress 43% (read 2770337792 bytes, duration 12 sec)
progress 44% (read 2834759680 bytes, duration 12 sec)
progress 45% (read 2899181568 bytes, duration 13 sec)
progress 46% (read 2963603456 bytes, duration 13 sec)
progress 47% (read 3028025344 bytes, duration 13 sec)
progress 48% (read 3092447232 bytes, duration 14 sec)
progress 49% (read 3156869120 bytes, duration 14 sec)
progress 50% (read 3221291008 bytes, duration 15 sec)
progress 51% (read 3285778432 bytes, duration 16 sec)
progress 52% (read 3350200320 bytes, duration 18 sec)
progress 53% (read 3414622208 bytes, duration 20 sec)
progress 54% (read 3479044096 bytes, duration 22 sec)
progress 55% (read 3543465984 bytes, duration 24 sec)
progress 56% (read 3607887872 bytes, duration 27 sec)
 
Hi again,

thank you for output :)

- did you upgrade your PVE node without reboot? (if yes - please try reboot your PVE)

- could you please post output of cat /etc/pve/storage.cfg

- do you have enough space in your storage?
 
Honestly I do not remember if I rebooted.. let me do it now.


Bash:
root@proxmox:~# cat /etc/pve/storage.cfg
dir: local
        path /var/lib/vz
        content vztmpl,iso,backup,rootdir
        maxfiles 2
        shared 0

lvmthin: local-lvm
        thinpool data
        vgname pve
        content rootdir,images

dir: picture
        path /mnt/picture
        content backup,images,vztmpl,rootdir,iso,snippets
        maxfiles 5
        shared 0

root@proxmox:~#