Search results

  1. C

    Ceph show "slow requests are blocked" when creating / modifying CephFS

    Here the output of ceph balancer status (works again after health check is green). root@ld3955:~# ceph balancer status { "active": true, "plans": [], "mode": "upmap" }
  2. C

    Ceph show "slow requests are blocked" when creating / modifying CephFS

    Hi, please focus on the issue slow requests are blocked in this ticket only. The other issues are a) under control (+35GB free disk space on monitor nodes) or b) addressed in ceph-user-list. Unfortunately ceph balancer status is not responding... could be related to slow requests are blocked.
  3. C

    Ceph show "slow requests are blocked" when creating / modifying CephFS

    Indeed I activated the balancer mode "upmap" when Ceph health status was green, means there was no relevant activity. Can you please advise which logs should be checked? Based on the output of ceph health detail I can see which OSDs are affected, and to my best knowledge the OSDs are always...
  4. C

    Ceph show "slow requests are blocked" when creating / modifying CephFS

    Hi Alwin, after some time the number of slow requests are blocked decreased, but very slowly. In my opinion there's a correlation between the number of PGs inactive and the number of slow requests are blocked. I must understand what is causing the slow requests are blocked on the pools that...
  5. C

    Ceph show "slow requests are blocked" when creating / modifying CephFS

    Hi, I have noticed in Ceph log (ceph -w) an increase of "slow requests are blocked" when I create CephFS, e.g. 2019-10-14 16:41:32.083294 mon.ld5505 [INF] daemon mds.ld4465 assigned to filesystem cephfs as rank 0 2019-10-14 16:41:32.121895 mon.ld5505 [INF] daemon mds.ld4465 is now active in...
  6. C

    [SOLVED] pveceph osd destroy is not cleaning device properly

    Hm... the manpage of pveceph shows --cleanup <boolean> (default = 0) If set, we remove partition table entries. My understanding of this is not that LVM, means volume group and logical volume, will be removed. Can you confirm that this option will remove volume group and logical...
  7. C

    [SOLVED] pveceph osd destroy is not cleaning device properly

    Hi, to remove an OSD I run this command: root@ld5506:/var/lib/vz# pveceph osd destroy 2 destroy OSD osd.2 Remove osd.2 from the CRUSH map Remove the osd.2 authentication key. Remove OSD osd.2 --> Zapping...
  8. C

    [SOLVED] Cannot move EFI disk to another storage

    Hi, my VMs use an EFI disk (in addition to the standard disk). Now I want to move all disks to another storage. There's no issue with the standard disk. However, the option to move the EFI disk is not available in WebUI. This means, I need drop the EFI disk and re-create it in new storage. But...
  9. C

    Ceph health status reports: Health error: 1 MDSs report slow metadata IOs, 1 MDSs report slow requests

    Hi, thanks for your reply. I have been pointed many time now to the backfill_toofull status as a root cause for the issue with slow MDS. However I'm not sure if this explanation still applies if you consider this: All OSDs that are affected by backfill_toofull reside on dedicated drives, and...
  10. C

    Ceph health status reports: Health error: 1 MDSs report slow metadata IOs, 1 MDSs report slow requests

    Hi, ceph health reports 1 MDSs report slow metadata IOs 1 MDSs report slow requests This is the complete output of ceph -s: root@ld3955:~# ceph -s cluster: id: 6b1b5117-6e08-4843-93d6-2da3cf8a6bae health: HEALTH_ERR 1 MDSs report slow metadata IOs 1 MDSs...
  11. C

    Cannot start any KVM / LXC

    Update: I think that this issue is related to other issues reported here and here. Furthermore I found out that I cannot copy data from the affected pool to local disk. I started copying a LXC dump file and this hangs after transferring Source root@ld3955:~# ls -l /mnt/pve/pve_cephfs/dump/...
  12. C

    Cannot start any KVM / LXC

    Hi, here I describe 1 of the 2 major issues I'm currently facing in my 8 node ceph cluster (2x MDS, 6x ODS). The issue is that I cannot start any virtual machine KVM or container LXC; the boot process just hangs after a few seconds. All these KVMs and LXCs have in common that their virtual...
  13. C

    [SOLVED] Kernel parameters not loaded during boot from /etc/sysctl.d/*.conf

    Hi, I was getting this error in syslog: nf_conntrack: nf_conntrack: table full, dropping packet To solve this issue I found this: CONNTRACK_MAX = RAMSIZE (in bytes) / 16384 / (ARCH / 32) Having Mellanox NIC installed on my server I followed the recommendation to improve performance. This...
  14. C

    [SOLVED] WebUI shows utilization = 62.91% of local storage, but /var/lib/vz is empty

    Hello! On any node belonging to PVE cluster I can see much more local storage utilization compared to what is stored on disk. The screenshot attached here is from my MGR node that shows utilization of 62.91%, but /var/lib/vz is empty! root@ld3955:~# ls -lR /var/lib/vz/ /var/lib/vz/: insgesamt...
  15. C

    Error in syslog: probably no mds server is up

    Hi, after rebooting 1 node serving MDS I get this error message in this node's syslog: root@ld3955:~# tail /var/log/syslog Sep 17 12:21:18 ld3955 kernel: [ 3141.167834] ceph: probably no mds server is up Sep 17 12:21:18 ld3955 pvestatd[2482]: mount error: exit code 2 Sep 17 12:21:28 ld3955...
  16. C

    [SOLVED] Howto define OSD weight in Crush map

    OK. I created a new OSD (from scratch), but there's no relevant entry in crush map except for device 8 osd.8 class hdd in section "devices. root@ld5505:~# pveceph osd create /dev/sdbm --db_dev /dev/sdbk --db_size 10 create OSD on /dev/sdbm (bluestore) creating block.db on '/dev/sdbk' Physical...
  17. C

    [SOLVED] MDS fails to start: unable to find a keyring on /var/lib/ceph/mds/ceph-admin/keyring

    The file /var/lib/ceph/mds/ceph-<ID>/keyring already exists. Therefore I simply modified the config in /etc/ceph/ceph.conf and now MDS starts w/o errors.
  18. C

    [SOLVED] MDS fails to start: unable to find a keyring on /var/lib/ceph/mds/ceph-admin/keyring

    root@ld3955:~# more /etc/pve/ceph.conf [global] auth client required = cephx auth cluster required = cephx auth service required = cephx cluster network = 192.168.1.0/27 debug ms = 0/0 fsid = 6b1b5117-6e08-4843-93d6-2da3cf8a6bae mon allow...
  19. C

    [SOLVED] MDS fails to start: unable to find a keyring on /var/lib/ceph/mds/ceph-admin/keyring

    Hi, I cannot start MDS services on active/standby node: root@ld3955:/var/log# systemctl status ceph-mds@ld3955 ● ceph-mds@ld3955.service - Ceph metadata server daemon Loaded: loaded (/lib/systemd/system/ceph-mds@.service; enabled; vendor preset: enabled) Drop-In...
  20. C

    [SOLVED] Howto define OSD weight in Crush map

    Hi, after adding an OSD to Ceph it is adviseable to create a relevant entry in Crush map using a weight size depending on disk size. Example: ceph osd crush set osd.<id> <weight> root=default host=<hostname> Question: How is the weight defined depending on disk size? Which algorithm can be...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!