Lvm group shows high storage use

arrcy

Member
Mar 21, 2020
17
0
6
24
Today i got the following error/warning when making a snapshot of my vm:

Code:
WARNING: You have not turned on protection against thin pools running out of space.
  WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full.
  Logical volume "vm-100-state-test" created.
  WARNING: Sum of all thin volume sizes (<227.48 GiB) exceeds the size of thin pool pve/data and the size of whole volume group (<223.07 GiB).
saving VM state and RAM using storage 'local-lvm'
2.01 MiB in 0s
404.63 MiB in 1s
435.62 MiB in 2s
629.49 MiB in 3s
762.66 MiB in 4s
completed saving the VM state in 4s, saved 783.73 MiB
snapshotting 'drive-scsi0' (local-lvm:vm-100-disk-0)
  WARNING: You have not turned on protection against thin pools running out of space.
  WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full.
  Logical volume "snap_vm-100-disk-0_test" created.
  WARNING: Sum of all thin volume sizes (<259.48 GiB) exceeds the size of thin pool pve/data and the size of whole volume group (<223.07 GiB).
snapshotting 'drive-efidisk0' (local-lvm:vm-100-disk-1)
  WARNING: You have not turned on protection against thin pools running out of space.
  WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full.
  Logical volume "snap_vm-100-disk-1_test" created.
  WARNING: Sum of all thin volume sizes (259.48 GiB) exceeds the size of thin pool pve/data and the size of whole volume group (<223.07 GiB).
TASK OK

This got me worried, and after inspection i found that my lvm pool was almost full under UI > Disk> LVM
Lvm storage.PNG
I don't understand why it shows that the disk is almost full because i have not used any storage, this claim also gets backed up when i look at my disks:storage2.PNG.
For those wondering why i even created a volume group,
I had a choice (remaining root storage) between directory or LVM, I choose Lvm because it had the possibility for snapshots.

see my disk configuration:
storage.PNG
DF --o gives:
Code:
Filesystem                     Type        Inodes  IUsed     IFree IUse% 1K-blocks     Used     Avail Use% File Mounted on
udev                           devtmpfs   2028553    744   2027809    1%   8114212        0   8114212   0% -    /dev
tmpfs                          tmpfs      2036573   1281   2035292    1%   1629260     1320   1627940   1% -    /run
/dev/mapper/pve-root           ext4       3653632  86616   3567016    3%  57278576 31226164  23113124  58% -    /
tmpfs                          tmpfs      2036573     99   2036474    1%   8146292    43680   8102612   1% -    /dev/shm
tmpfs                          tmpfs      2036573     23   2036550    1%      5120        0      5120   0% -    /run/lock
/dev/sdb2                      vfat             0      0         0     -    523248      312    522936   1% -    /boot/efi
storage-nvme                   zfs      587378143     15 587378128    1% 293689088      128 293688960   1% -    /storage-nvme
storage-nvme/subvol-102-disk-0 zfs       15595898  26218  15569680    1%   8388608   603776   7784832   8% -    /storage-nvme/subvol-102-disk-0
storage-nvme/subvol-101-disk-0 zfs       20747147 146347  20600800    1%  25165824 14865536  10300288  60% -    /storage-nvme/subvol-101-disk-0
storage-nvme/subvol-116-disk-0 zfs      351448086  49494 351398592    1% 188743680 13044480 175699200   7% -    /storage-nvme/subvol-116-disk-0
storage-nvme/subvol-107-disk-0 zfs      158813263  41087 158772176    1%  83886080  4500096  79385984   6% -    /storage-nvme/subvol-107-disk-0
storage-nvme/subvol-114-disk-0 zfs       47666900  41460  47625440    1%  25165824  1353216  23812608   6% -    /storage-nvme/subvol-114-disk-0
storage-nvme/subvol-117-disk-0 zfs       48828352  29096  48799256    1%  25165824   766208  24399616   4% -    /storage-nvme/subvol-117-disk-0
storage-nvme/subvol-119-disk-0 zfs       39479584  27128  39452456    1%  26214400  6488192  19726208  25% -    /storage-nvme/subvol-119-disk-0
storage-nvme/subvol-118-disk-0 zfs      161522148  40524 161481624    1%  83886080  3145344  80740736   4% -    /storage-nvme/subvol-118-disk-0
storage-nvme/subvol-103-disk-0 zfs       45964936  34680  45930256    1%  25165824  2200704  22965120   9% -    /storage-nvme/subvol-103-disk-0
storage/bkup                   zfs      438070474     18 438070456    1% 220059136  1024000 219035136   1% -    /bkup
storage                        zfs      438070463      7 438070456    1% 219035264      128 219035136   1% -    /storage
storage/subvol-109-disk-0      zfs       81770504  50632  81719872    1%  41943040  1083136  40859904   3% -    /storage/subvol-109-disk-0
/dev/fuse                      fuse        262144     56    262088    1%    131072       32    131040   1% -    /etc/pve


Now the question,
How do i remove something from LVM, so it's not almost full. Or make it show the correct storage use.

Best Regards,
 
Hi,
Today i got the following error/warning when making a snapshot of my vm:

Code:
WARNING: You have not turned on protection against thin pools running out of space.
  WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full.
  Logical volume "vm-100-state-test" created.
  WARNING: Sum of all thin volume sizes (<227.48 GiB) exceeds the size of thin pool pve/data and the size of whole volume group (<223.07 GiB).
saving VM state and RAM using storage 'local-lvm'
2.01 MiB in 0s
404.63 MiB in 1s
435.62 MiB in 2s
629.49 MiB in 3s
762.66 MiB in 4s
completed saving the VM state in 4s, saved 783.73 MiB
snapshotting 'drive-scsi0' (local-lvm:vm-100-disk-0)
  WARNING: You have not turned on protection against thin pools running out of space.
  WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full.
  Logical volume "snap_vm-100-disk-0_test" created.
  WARNING: Sum of all thin volume sizes (<259.48 GiB) exceeds the size of thin pool pve/data and the size of whole volume group (<223.07 GiB).
snapshotting 'drive-efidisk0' (local-lvm:vm-100-disk-1)
  WARNING: You have not turned on protection against thin pools running out of space.
  WARNING: Set activation/thin_pool_autoextend_threshold below 100 to trigger automatic extension of thin pools before they get full.
  Logical volume "snap_vm-100-disk-1_test" created.
  WARNING: Sum of all thin volume sizes (259.48 GiB) exceeds the size of thin pool pve/data and the size of whole volume group (<223.07 GiB).
TASK OK

This got me worried, and after inspection i found that my lvm pool was almost full under UI > Disk> LVM
View attachment 32054
This only means that so much of the disk capacity is assigned to LVM, not that it is already filled within LVM. See the LVM-Thin panel for usage information within the thin pool, or use lvs and look at the Data% column to see all the details.

Provisioning more space than available (what the warning is about) is not a problem in itself, but you should keep an eye on the actual usage within the pool, to not have it run full.

I don't understand why it shows that the disk is almost full because i have not used any storage, this claim also gets backed up when i look at my disks:View attachment 32056.

For those wondering why i even created a volume group,
I had a choice (remaining root storage) between directory or LVM, I choose Lvm because it had the possibility for snapshots.

see my disk configuration:
View attachment 32055
DF --o gives:
Code:
Filesystem                     Type        Inodes  IUsed     IFree IUse% 1K-blocks     Used     Avail Use% File Mounted on
udev                           devtmpfs   2028553    744   2027809    1%   8114212        0   8114212   0% -    /dev
tmpfs                          tmpfs      2036573   1281   2035292    1%   1629260     1320   1627940   1% -    /run
/dev/mapper/pve-root           ext4       3653632  86616   3567016    3%  57278576 31226164  23113124  58% -    /
tmpfs                          tmpfs      2036573     99   2036474    1%   8146292    43680   8102612   1% -    /dev/shm
tmpfs                          tmpfs      2036573     23   2036550    1%      5120        0      5120   0% -    /run/lock
/dev/sdb2                      vfat             0      0         0     -    523248      312    522936   1% -    /boot/efi
storage-nvme                   zfs      587378143     15 587378128    1% 293689088      128 293688960   1% -    /storage-nvme
storage-nvme/subvol-102-disk-0 zfs       15595898  26218  15569680    1%   8388608   603776   7784832   8% -    /storage-nvme/subvol-102-disk-0
storage-nvme/subvol-101-disk-0 zfs       20747147 146347  20600800    1%  25165824 14865536  10300288  60% -    /storage-nvme/subvol-101-disk-0
storage-nvme/subvol-116-disk-0 zfs      351448086  49494 351398592    1% 188743680 13044480 175699200   7% -    /storage-nvme/subvol-116-disk-0
storage-nvme/subvol-107-disk-0 zfs      158813263  41087 158772176    1%  83886080  4500096  79385984   6% -    /storage-nvme/subvol-107-disk-0
storage-nvme/subvol-114-disk-0 zfs       47666900  41460  47625440    1%  25165824  1353216  23812608   6% -    /storage-nvme/subvol-114-disk-0
storage-nvme/subvol-117-disk-0 zfs       48828352  29096  48799256    1%  25165824   766208  24399616   4% -    /storage-nvme/subvol-117-disk-0
storage-nvme/subvol-119-disk-0 zfs       39479584  27128  39452456    1%  26214400  6488192  19726208  25% -    /storage-nvme/subvol-119-disk-0
storage-nvme/subvol-118-disk-0 zfs      161522148  40524 161481624    1%  83886080  3145344  80740736   4% -    /storage-nvme/subvol-118-disk-0
storage-nvme/subvol-103-disk-0 zfs       45964936  34680  45930256    1%  25165824  2200704  22965120   9% -    /storage-nvme/subvol-103-disk-0
storage/bkup                   zfs      438070474     18 438070456    1% 220059136  1024000 219035136   1% -    /bkup
storage                        zfs      438070463      7 438070456    1% 219035264      128 219035136   1% -    /storage
storage/subvol-109-disk-0      zfs       81770504  50632  81719872    1%  41943040  1083136  40859904   3% -    /storage/subvol-109-disk-0
/dev/fuse                      fuse        262144     56    262088    1%    131072       32    131040   1% -    /etc/pve


Now the question,
How do i remove something from LVM, so it's not almost full. Or make it show the correct storage use.

Best Regards,
 
Hi ive got this issue also...spent hours researching it but cant find an answer that i understand.....one view is showing as 98% full and another view is showing 32% full.... of course i want to assign all of the disc to possible vm usage but, i the other view im only 32% which means ive got room but i cant rebuild my HomeAssistant vm....totally stuck at the moment!
 
Hi,
Hi ive got this issue also...spent hours researching it but cant find an answer that i understand.....one view is showing as 98% full and another view is showing 32% full.... of course i want to assign all of the disc to possible vm usage but, i the other view im only 32% which means ive got room but i cant rebuild my HomeAssistant vm....totally stuck at the moment!
please share the output of vgs and lvs. What exactly is the error you get when you try to create the volume for the new VM?
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!