I'm having issues moving CT root fs from local storage (zfs) to ceph rbd.
The problem only occurs on existing CT, created some time ago. At that time I was using proxmox 4.x
The error does not occur on new CT created with proxmox 6.2, using ubuntu 20.x template.
EDIT: It seems the issue is due to subvol-based CT where the size of the disk is determined as being 0 (see post #2 and #3). The raw-based CT work ok.
My setup is a 6 node proxmox cluster, on which 3 nodes are ceph hyper-converged nodes. The 4th node has connection to the ceph storage.
The problem seems to be an error on creating the rbd device, so it cannot be formatted:
Here's is the only line in kernel log I get:
I tried also migrating the rootfs of the CT(s) from one local storage (ZFS) to the classic LVM storage (/var/lib/vz). This works. Then migrating from the LVM storage to the rbd device gives the same error. This is true for multiple CTs.
Backup-restore on top of the rbd device works normally. It's only moving the rootfs that has issues. Not sure if this is a bug or something wrong with my installation.
I installed the ceph cluster from the proxmox gui so I'm not sure what to check. Is there a way to get a detailed debug log from the operations performed by proxmox so I can try to understand what the error is?
thank you.
The problem only occurs on existing CT, created some time ago. At that time I was using proxmox 4.x
The error does not occur on new CT created with proxmox 6.2, using ubuntu 20.x template.
EDIT: It seems the issue is due to subvol-based CT where the size of the disk is determined as being 0 (see post #2 and #3). The raw-based CT work ok.
My setup is a 6 node proxmox cluster, on which 3 nodes are ceph hyper-converged nodes. The 4th node has connection to the ceph storage.
The problem seems to be an error on creating the rbd device, so it cannot be formatted:
Code:
/dev/rbd0
mke2fs 1.44.5 (15-Dec-2018)
mkfs.ext4: Device size reported to be zero. Invalid partition specified, or
partition table wasn't reread after running fdisk, due to
a modified partition being busy and in use. You may need to reboot
to re-read your partition table.
Removing image: 100% complete...done.
TASK ERROR: command 'mkfs.ext4 -O mmp -E 'root_owner=0:0' /dev/rbd/vmd/vm-107-disk-0' failed: exit code 1
Here's is the only line in kernel log I get:
Code:
Aug 8 10:57:59 node4 kernel: [321974.128498] rbd: rbd1: capacity 0 features 0x3d
Backup-restore on top of the rbd device works normally. It's only moving the rootfs that has issues. Not sure if this is a bug or something wrong with my installation.
I installed the ceph cluster from the proxmox gui so I'm not sure what to check. Is there a way to get a detailed debug log from the operations performed by proxmox so I can try to understand what the error is?
thank you.
Last edited: