problem with starting pve with root on zfs pool

Gladio273

New Member
Jan 24, 2024
3
0
1
the proxmox server has the following configuration:
1 * SSD - Pool: rpool
2 * HDD - Pool: storage

The boot is stopping in initramfs while importing rpool.

zpool import shows:
pool: rpool
id: 4391380091351698291
state: ONLINE
status: The pool was last accessed by another system.
action: The pool can be imported using its name or numeric identifier and
the '-f' flag.
see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-EY
config:

rpool ONLINE
nvme-eui.002538b811c143a4-part3 ONLINE


zpool list: (storage already imported)

NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT
storage 3.62T 93.1G 3.53T - - 0% 2% 1.00x ONLINE -

zpool import rpool shows:
cannot import 'rpool': pool was previously in use from another system.
Last accessed by csosrv12 (hostid=86229ab9) at Mon Jan 13 16:05:08 2025
The pool can be imported, use 'zpool import -f' to import the pool.

(This is correct - I am on a rescue system)
zpool import -f rpool shows:
cannot import 'rpool': I/O error
Destroy and re-create the pool from
a backup source.

fdisk -l /dev/nvme0n1 shows

Disk /dev/nvme0n1: 476.94 GiB, 512110190592 bytes, 1000215216 sectors
Disk model: SAMSUNG MZVL2512HCJQ-00B00
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: gpt
Disk identifier: B9883CBB-BBFE-4BD0-9916-07429260538C

Device Start End Sectors Size Type
/dev/nvme0n1p1 34 2047 2014 1007K BIOS boot
/dev/nvme0n1p2 2048 2099199 2097152 1G EFI System
/dev/nvme0n1p3 2099200 1000215182 998115983 475.9G Solaris /usr & Apple ZFS


Partitions are corect size and location. (100%)

zpool -f -F -X rpool (dauert sehr lange - volle IO-Last)
cannot import 'rpool': one or more devices is currently unavailable

cd /backup (mounted by zpool storage)
dd if=/dv/nvme0n1 of=nvme0n1.dd bs=20M

root@rescue /backup # dd if=/dev/nvme0n1 of=nvme0n1.dd bs=20M
24419+1 records in
24419+1 records out
512110190592 bytes (512 GB, 477 GiB) copied, 2285.81 s, 224 MB/s

works without any error.

Is there a write error on a bad sector, while importing the pool?

What can I do going further on my proxmox server.

Greetings
 
Last edited:
  • Like
Reactions: waltar
I'll start by saying that I imagine the physical drive device itself is probably OK based on the dd read (this obviously does not include the write part) - however you probably have data-corruption. If you experienced a power-loss this is probably the cause - & the device is probably still good even for write.

You may want to take a look at this product, I've never used it myself, but based on other users it does seem effective if expensive. Probably depends on how much your data is worth to you. There may be other choices - I don't really use ZFS.

Anyway, good luck.
 
  • Like
Reactions: Kingneutron

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!