Search results

  1. R

    INFO: task kworker blocked for more than 122 seconds

    using pve 9.0.5 . 5 node ceph cluster. nodes have a mix of zfs and non zfs root/boot disks along with one large nvme formatted ext4 for vzdumps. we also use pbs . i have a cronscript which we have used for years that checks this: dmesg -T | grep hung | grep -v vethXChung ## **URGENT**...
  2. R

    [SOLVED] ceph - how to remove a monitor for a deleted node

    we had a crashed node. it is deleted from cluster. we did not have a chance to remove it as a ceph monitor. [ we did replace the monitor ]. ceph -s shows: health: HEALTH_WARN .. mon: 4 daemons, quorum pve11,pve2,pve5 (age 7m), out of quorum: pve4 how can I delete a monitor assigned...
  3. R

    osd move to new server

    generally I destroy the OSD on old node , move it to new node then create a new osd. Or is there a stable way to do the osd ? A few months ago I checked threads and had no luck. Also I did not see an osd move in manual.
  4. R

    [SOLVED] after upgrade to 9, one node has ceph fail

    I noticed osd's down at pve web page. tried to start , fail . at a not where ceph is up: # ceph -s cluster: id: 220b9a53-4556-48e3-a73c-28deff665e45 health: HEALTH_WARN noout flag(s) set 10 osds down 1 host (10 osds) down Degraded...
  5. R

    [SOLVED] ifupdown2 and /etc/network/interfaces issue

    I upgraded a KVM running Debian bookworm to trixie. pve runs v8 after upgrade and reboot the network was down. tried to bring up network: ifup ens19 main exception 'RawConfigParser' object has no attribute 'readfp' my interfaces file is below. the solution from grok ai: Edit...
  6. R

    [SOLVED] Trixie PBS client

    I am upgrading a few systems to Trixie. Is there a pbs client sources list for trixie?
  7. R

    How to find which disk is used by an rbd

    Hello I have a disk which has an issue. per zabbix: rbd15 : Disk read/write request responses are too high (read > 20 ms for 15m or write > 20 ms for 15m) I spent an hour trying to find which disk has the issue. the regular cli tools I use , and could not find it. so I checked dmsg and...
  8. R

    lxc reboot issue

    'reboot' error from cli and pve run_buffer: 571 Script exited with status 32 lxc_init: 845 Failed to run lxc.hook.pre-start for container "501" __lxc_start: 2034 Failed to initialize container "501" TASK ERROR: startup for container '501' failed more info: # pct config 501 arch: amd64 cores...
  9. R

    [SOLVED] dhcp issue with KVM, LXC does not have the issue

    Hello we use pve enterprise repos all up to date. dhcp server is isc-kea we have fixed reservations set up at dhcp server based on mac address for some kvm's which are configured as dhcp clients in /etc/network/interfaces . those kvms are not getting the reserved IP, instead they get an IP...
  10. R

    how can I access PBS backups on USB

    yearly we rsync pbs backups to a large usb. For a court case I need to access emails from one of the usb's . I tried adding a new datastore using the directory . but got this error: () unable to create chunk store 'usb12' subdir "/media/usb12/pbs-backups/daily/.chunks" - unable to create...
  11. R

    [SOLVED] 5.8 kernel , vzdump to local storage results in node fenced

    Hello we have a 5 node cluster. a couple of months ago we had this issue, to solve I pinned a 6.5 kernel. Last night I unpinned and booted the 5 nodes to use 6.8.4-2-pve. at 2AM shortly after a vzdump backup to this storage : dir: z-local-nvme path /nvme-ext4 content...
  12. R

    [SOLVED] TASK ERROR: permission check failed.

    Hello I have two remote PBS servers which are used as re mote sync for our main pbs system. Those had been working rock solid for a few years. However I just checked both and sync jobs have been failing since 11/11/2023 . Questions: 1- is there a way of getting notified when a sync job...
  13. R

    [SOLVED] dirty-bitmap status: existing bitmap was invalid and has been cleared

    We backup a particular KVM to two different datastores. to one datastore the backup takes 10-11 minutes each time . to the other datastore the backup time is less then 20 seconds. backup to the slow data store has the following messages in log: INFO: scsi0: dirty-bitmap status...
  14. R

    [SOLVED] tags - what are they and where to get more info.

    Hello Where can I get information on what tags are and how to use? I tried to search tags on forum and did not see an answer. Also is there a way to search documentation for something like this? best regards, Rob Fantini
  15. R

    migrations on a node restart

    Hello I think a nice change would be for migrations to occur before the non high availability vm;s are shutdown.
  16. R

    container backup fail

    Hello We have around 15 containers. Just one has backup failures - 2 times in the past 4 days here is more info: dmesg [Sat Oct 14 08:44:46 2023] rbd: rbd1: capacity 15032385536 features 0x1d [Sat Oct 14 08:44:46 2023]...
  17. R

    re installing pve on a system with osd's

    we've 5 pve hosts with 7 OSD's each. If for some reason I had to reinstall pve to one of the nodes is there a way to preserve the osd's ? the reinstall would be fast and noout set beforehand. PS: I assume this: these days with very reliable ssd or nvme [ having good DWPD ] available I do...
  18. R

    [SOLVED] replacing an rpool disk , question on grub-install

    Hello I replaced a disk in an rpool. Per my notes the last step is to run this on the new disk:: grub-install /dev/nvme0n1 However that returned: grub-install is disabled because this system is booted via proxmox-boot-tool, if you really need to run it, run /usr/sbin/grub-install.real Is...
  19. R

    [SOLVED] 'ceph pg 55.0 query' not working

    hello , today I upgraded to the latest ceph. after the first note was upgraded i noticed an inactive pg warning. I continued and finished the upgrade hoping the inactive pg would be fixed with a complete upgrade. But that was not the case. following info from...
  20. R

    log rotation

    For some reason on PVE and Nextcloud and other systems log rotation came to a halt in November. /var/log had over 7G . /var/log/journal had 4G i think it has something to do with a conflict with rsyslog and systemd journal stuff. will check later to fix edit /etc/cron.daily/logrotate #...