yeah I am trying to reduce down to just one monitor and manager and then add back but nothign seems to be helping... so much frustration over last 3 months now since upgrade... I did have some vm I would like to restore and save so only reason I am trying so hard not to lose everything.
All the proxmox nodes seem connected fine and working - but ceph is giving me fits
every node shows timeout when using gui to see ceph status... in fact every node is not responding to ceph commands at cli except for node 2 - which is the node I decided to keep for the last known good monitor...
Can anyone point me in the right direction to fix this?
ceph-osd --check-wants-journal
root@node2:/var/lib/ceph/osd/ceph-1# ceph-osd --check-wants-journal
2022-04-14T20:55:43.210-0500 7f71cd669f00 -1 auth: unable to find a keyring on /var/lib/ceph/osd/ceph-admin/keyring: (2) No such file or...
I wonder if this is similar to my problem when I updated... now I get
ceph-osd --check-wants-journal
root@node2:/var/lib/ceph/osd/ceph-1# ceph-osd --check-wants-journal
2022-04-14T20:55:43.210-0500 7f71cd669f00 -1 auth: unable to find a keyring on /var/lib/ceph/osd/ceph-admin/keyring: (2) No...
https://forum.proxmox.com/threads/zippys-noob-log-of-how-to-for-proxmox-and-linux-admin-noob.107537/#post-462320
just started adding stuff there - feel free to correct me and I will update as I go along...
Would it be OK if I start logging all the "Ah Ha" moments for me as I learn basics of PM and Ceph and even basic admin stuff?
How to resize root and move swap space on default Proxmox install.
Today I finally READ the man pages after much fubling through forum after forum post of frustrated...
https://askubuntu.com/questions/1040611/will-multiple-swap-spaces-be-effective
suggests that there may be some benefit to linux default round-robin behavior if I add swap on 2 equal ssd "like raid0 with the performance boost" by assigning both swap partitions equal priority...
this is all new...
I guess I will do as suggested and boot to usb and move everything - was hoping for a cli way to just kill all processes and copy then unmount the /var/lib/ceph which I am still not sure is the problem - but anyhow - and then change fstab to use new drive - thinking I should also move swap over...
confirmed with my 9 node cluster - took all offline for cleaning, adding ssd and updates - then restarted one at a time to confirm changes and cannot use gui (loads webpage for login but "Login failed. Please try again" when login with main root user and correct password - whereas it works fine...
Ok so what's best way to stop all services mounted there and validate all are stopped? I can install new 80gb ssd just for ceph mon.. just need to know best practices for format type, best way to copy and then make sure ceph and everything pm knows where it lives correctly... suppose whatever I...
Anyone have any specific instructions on how to move ceph-mon directory to new partition post-proxmox install in default?
I have 80GB SSD for OS and PM and it seems to fill up with basic OS and PM updates and causes ceph to go into warn/critical shutdown even with space left on the drive due to...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.