Recent content by Denny Fuchs

  1. Denny Fuchs

    [SOLVED] LVM (over iscsi) volume stuck and cannot be created

    hi @fweber gladly .. I wanted to create a new developer VM and in this case .. exactly this happens, what you told about your issue. An existing device will be reused. Maybe .. you can still use this info: root@fra-corp-pmox-03:[~]: pveversion pve-manager/7.4-3/9002ab8a (running kernel...
  2. Denny Fuchs

    [SOLVED] LVM (over iscsi) volume stuck and cannot be created

    hi @fweber don't think so. It does not happen anymore, if you add a filter in lvm.conf like this: global_filter=["r|/dev/zd.*|","r|/dev/mapper/.*-vm--[0-9]+--disk--[0-9]+|"] Most of the issues we had, for create new VMs where no ID was reused. cu denny
  3. Denny Fuchs

    PVE7: VMs stopping, because Host ZFS is full (rpool/ROOT/pve-1)

    Here comes the fancy stuff: Syslog VM: Filesystem Size Used Avail Use% Mounted on /dev/mapper/vg2-data 2.7T 927G 1.7T 37% /data/syslog Debian VM df -h /opt/ Filesystem Size Used Avail Use% Mounted on /dev/mapper/vg--data-optfs 777G 625G 121G 84% /opt...
  4. Denny Fuchs

    PVE7: VMs stopping, because Host ZFS is full (rpool/ROOT/pve-1)

    Hi, we don't have snapshots. We delete them, after dangerous work was done: root@fc-r02-pmox-06:[~]: zfs get all rpool/pve-container/vm-100-disk-2 NAME PROPERTY VALUE SOURCE rpool/pve-container/vm-100-disk-2 type...
  5. Denny Fuchs

    PVE7: VMs stopping, because Host ZFS is full (rpool/ROOT/pve-1)

    We had issues in the past .. that the syslog host killed all VMs, with I/O .. so we limited: ostype: l26 scsi0: pve-container:vm-105-disk-0,size=25G,ssd=1 scsi1: pve-container:vm-105-disk-1,discard=on,mbps_wr=40,size=2789377M,ssd=1 we had no issues anymore .. since years .. but the OS was...
  6. Denny Fuchs

    PVE7: VMs stopping, because Host ZFS is full (rpool/ROOT/pve-1)

    hi @aaron I've checked our syslog host .. in the timerange .. a log cleaner started, which cleaned up 300GB. Exactly the same 300GB .. which are "lost" on the rpool and for rpool/ROOT/pve-1. For the syslog host: zfs get volsize,refreservation,used rpool/pve-container/vm-105-disk-1 NAME...
  7. Denny Fuchs

    PVE7: VMs stopping, because Host ZFS is full (rpool/ROOT/pve-1)

    Hello, since a few days, we have on a PVE7 - standalone - host the strange issue, that VMs stuck and not responding (host unreachable) anymore. The root cause seems to be, that the host "/" is getting full: Some envs: Debian Bullseye pve-manager/7.4-17/513c62be Filesystem Size...
  8. Denny Fuchs

    [SOLVED] LVM over iscsi: Delete VM1 with disks. Create new VM2 .. old data from VM1 are available

    hi, is there a way / hook to make sure, if a VM is purged, that PVE destroys data on the assigned disks first ? I have very often the issue, that data from purged VMs gets available often again, on a new VM. Yesterday I had this issue again, as I've added a 2nd disk to the new VM and create a...
  9. Denny Fuchs

    [SOLVED] LVM (over iscsi) volume stuck and cannot be created

    hi, I have this issue also with PVE7 on two different PVE7 clusters, which has ISCSI as shared device. I know this workaround with dmesetup remove .. but I do not really understand, the source of this issue. I can be only a race condition, as it happens not always .. but pretty often for...
  10. Denny Fuchs

    Ceph Quincy: "ceph osd crush move ... root=default" -> rbd 100%

    Yepp, thanks .. what do you think .. can we let the VMs powered on, or should we power them off ? cu denny
  11. Denny Fuchs

    Ceph Quincy: "ceph osd crush move ... root=default" -> rbd 100%

    Hi, I've posted also to the Ceph Mailinglist and one answered me .. that: ceph osd pool set db-pool crush_rule fc-r02-ssd ceph osd pool set ssd-pool crush_rule fc-r02-ssd will do the work .. because of "item_name": "default~ssd" .. We will take most parts offline .. as I have no idea ...
  12. Denny Fuchs

    Ceph Quincy: "ceph osd crush move ... root=default" -> rbd 100%

    Hello, we upgraded our Nautilus till Quincy two weeks ago and wanted to get rid of old settings, which we have since Luminous or older: Splitted SSD / HDD, before Ceph had device-classes: https://www.sebastien-han.fr/blog/2014/08/25/ceph-mix-sata-and-ssd-within-the-same-box/ We executed...
  13. Denny Fuchs

    PVE6 | pvemailforward | pveupdate: node-03 sends mail to node-01 instead

    Hello, I have since update from 5 to 6 the problem, that the node-03 wants to send mails to node-01 .. instead of, what is configured on Postfix. We manage the whole Postfix configuration with Puppet, so every node is identical ... except myorigin = fra-corp-pmox-03.example.local Every node...
  14. Denny Fuchs

    Proxmox: Irgendwo wird Debian Mirror(?) aufgerufen (199.232.138.132)

    hi, sorry völlig vergessen ... wir haben ausschließlich eigene Quellen .. daher wundere ich mich ja. cu denny

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!