good evening,
i posted in another thread (https://forum.proxmox.com/threads/proxmox-6-ceph-mds-stuck-on-creating.57524/#post-268549) that was created on the same topic and just hopped on to it, but that thread seems to be dead. so i am trying my luck here to see if this is a general problem with the proxmox ceph implementation or something else. i am aware that i dont have a subscription and that this post might be lower priority because of it.
fresh ceph node, no clustering, getting all updates after install (non-subscription-repo) and then used the webgui to install ceph on the node.
ceph installs fine, i create OSDs and they show up. i can create a pool, all good.
i create an MDS in order to create a cephfs and the MDS shows up as standby. i then create the cephfs which hangs during creation with the reason that no MDS has responded.
restarting the created MDS gives this syslog (gotten via the webgui):
i took some screenshots to illustrate: https://imgur.com/a/WpmR7ZW
please let me know what i am doing wrong or if ceph is broken on the current proxmox release.
i posted in another thread (https://forum.proxmox.com/threads/proxmox-6-ceph-mds-stuck-on-creating.57524/#post-268549) that was created on the same topic and just hopped on to it, but that thread seems to be dead. so i am trying my luck here to see if this is a general problem with the proxmox ceph implementation or something else. i am aware that i dont have a subscription and that this post might be lower priority because of it.
fresh ceph node, no clustering, getting all updates after install (non-subscription-repo) and then used the webgui to install ceph on the node.
ceph installs fine, i create OSDs and they show up. i can create a pool, all good.
i create an MDS in order to create a cephfs and the MDS shows up as standby. i then create the cephfs which hangs during creation with the reason that no MDS has responded.
restarting the created MDS gives this syslog (gotten via the webgui):
-- Logs begin at Sat 2019-11-02 23:33:41 CET, end at Sat 2019-11-02 23:47:46 CET. --
Nov 02 23:45:29 mond systemd[1]: Started Ceph metadata server daemon.
Nov 02 23:45:29 mond ceph-mds[9643]: starting mds.mond at
Nov 02 23:47:13 mond ceph-mds[9643]: 2019-11-02 23:47:13.661 7f1eef9df700 -1 received signal: Terminated from /sbin/init (PID: 1) UID: 0
Nov 02 23:47:13 mond ceph-mds[9643]: 2019-11-02 23:47:13.661 7f1eef9df700 -1 mds.mond *** got signal Terminated ***
Nov 02 23:47:13 mond systemd[1]: Stopping Ceph metadata server daemon...
Nov 02 23:47:17 mond ceph-mds[9643]: /mnt/pve/ceph-dev/ceph/ceph-14.2.4/src/include/elist.h: In function 'elist<T>::~elist() [with T = MDSIOContextBase*]' thread 7f1ef2c30340 time 2019-11-02 23:47:17.624970
Nov 02 23:47:17 mond ceph-mds[9643]: /mnt/pve/ceph-dev/ceph/ceph-14.2.4/src/include/elist.h: 91: FAILED ceph_assert(_head.empty())
Nov 02 23:47:17 mond ceph-mds[9643]: ceph version 14.2.4 (65249672c6e6d843510e7e01f8a4b976dcac3db1) nautilus (stable)
Nov 02 23:47:17 mond ceph-mds[9643]: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x152) [0x7f1ef44d460e]
Nov 02 23:47:17 mond ceph-mds[9643]: 2: (()+0x2807e6) [0x7f1ef44d47e6]
Nov 02 23:47:17 mond ceph-mds[9643]: 3: (()+0x3dd983) [0x55732615f983]
Nov 02 23:47:17 mond ceph-mds[9643]: 4: (()+0x39d8c) [0x7f1ef3165d8c]
Nov 02 23:47:17 mond ceph-mds[9643]: 5: (()+0x39eba) [0x7f1ef3165eba]
Nov 02 23:47:17 mond ceph-mds[9643]: 6: (__libc_start_main()+0xf2) [0x7f1ef31500a2]
Nov 02 23:47:17 mond ceph-mds[9643]: 7: (_start()+0x2a) [0x557325ebe98a]
Nov 02 23:47:17 mond ceph-mds[9643]: *** Caught signal (Segmentation fault) **
Nov 02 23:47:17 mond ceph-mds[9643]: in thread 7f1ef2c30340 thread_name:ceph-mds
Nov 02 23:47:17 mond ceph-mds[9643]: ceph version 14.2.4 (65249672c6e6d843510e7e01f8a4b976dcac3db1) nautilus (stable)
Nov 02 23:47:17 mond ceph-mds[9643]: 1: (()+0x12730) [0x7f1ef389e730]
Nov 02 23:47:17 mond ceph-mds[9643]: 2: (__pthread_mutex_lock()+0) [0x7f1ef38966c0]
Nov 02 23:47:17 mond ceph-mds[9643]: 3: (ceph::logging::Log::submit_entry(ceph::logging::Entry&&)+0x41) [0x7f1ef4848371]
Nov 02 23:47:17 mond ceph-mds[9643]: 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x24c) [0x7f1ef44d4708]
Nov 02 23:47:17 mond ceph-mds[9643]: 5: (()+0x2807e6) [0x7f1ef44d47e6]
Nov 02 23:47:17 mond ceph-mds[9643]: 6: (()+0x3dd983) [0x55732615f983]
Nov 02 23:47:17 mond ceph-mds[9643]: 7: (()+0x39d8c) [0x7f1ef3165d8c]
Nov 02 23:47:17 mond ceph-mds[9643]: 8: (()+0x39eba) [0x7f1ef3165eba]
Nov 02 23:47:17 mond ceph-mds[9643]: 9: (__libc_start_main()+0xf2) [0x7f1ef31500a2]
Nov 02 23:47:17 mond ceph-mds[9643]: 10: (_start()+0x2a) [0x557325ebe98a]
Nov 02 23:47:17 mond systemd[1]: ceph-mds@mond.service: Main process exited, code=killed, status=11/SEGV
Nov 02 23:47:17 mond systemd[1]: ceph-mds@mond.service: Failed with result 'signal'.
Nov 02 23:47:17 mond systemd[1]: Stopped Ceph metadata server daemon.
Nov 02 23:47:45 mond systemd[1]: Started Ceph metadata server daemon.
Nov 02 23:47:46 mond ceph-mds[11092]: starting mds.mond at
i took some screenshots to illustrate: https://imgur.com/a/WpmR7ZW
please let me know what i am doing wrong or if ceph is broken on the current proxmox release.