Search results

  1. K

    Failed to start Ceph disk activation: /dev/sd* and OSD's down after Proxmox upgrade to v6

    The OSD's are down on this node, but Ceph is still running. I also can't run any 'Ceph' commands on this node. On other node 'ceph health': HEALTH_WARN 1/3 mons down, quorum node003,node004 On other node 'ceph -s': cluster: id: 09935430-cfe7-48d4-ac66-c02e0455d95de health...
  2. K

    Failed to start Ceph disk activation: /dev/sd* and OSD's down after Proxmox upgrade to v6

    Configuration: 6 nodes with Ceph and Proxmox 5. I am currently upgrading Proxmox to version 6 (running corosync version 3 now). I did NOT update Ceph yet. But after upgrading the first node I get this Ceph error, but everything is still up except these osd's: systemctl status...
  3. K

    Corosync update without Proxmox/Ceph latest version?

    Thanks Alwin, but just to be sure... everything will be up and running after the corosync update or do you directly have to upgrade to the latest proxmox and ceph?
  4. K

    Corosync update without Proxmox/Ceph latest version?

    Goodmorning, I am running a cluster with proxmox-ve 5.4-2, ceph 12.2.12 and corosync 2.4.4. Is it possible to upgrade corosync to version 3.0.2. without updating proxmox and ceph? Thanks!
  5. K

    Mixed Subscriptions

    I ran latest updates today and everything seems to work again.
  6. K

    Mixed Subscriptions

    Good afternoon! I have 3 nodes in Proxmox, but suddenly 1 node is currently offline in Proxmox panel (all VM's are still running on that node). Last week it was another node with the same problem and had to reboot the node. I have this error in the Proxmox panel: "Mixed Subscriptions - Warning...
  7. K

    'Error' under guests in Proxmox panel

    Sorry I found it already, There is a state error under HA.
  8. K

    'Error' under guests in Proxmox panel

    My cluster is running healthy and everything is up and running, but in my Proxmox panel on the mainpage under 'Guests' there is an 'Error 1' but doesn't show any more information (Everything is running in green and nothing is stopped). Does anyone knows in which log I can check this Error? Thanks!
  9. K

    Deleted ceph on node, stupid!

    Thanks all for your help, and if I missed something I am happy to hear this.
  10. K

    Deleted ceph on node, stupid!

    In /etc/pve/priv there is a file ceph.client.admin.keyring. I changed the key in this file into the key from this file /etc/pve/priv/ceph/ceph_ssd.keyring and Ceph under the node is working again. ceph -s cluster: health: HEALTH_OK
  11. K

    Deleted ceph on node, stupid!

    Some extra info: root@node002:~# systemctl status ceph-mon.target ● ceph-mon.target - ceph target allowing to start/stop all ceph-mon@.service instances at once Loaded: loaded (/lib/systemd/system/ceph-mon.target; enabled; vendor preset: enabled) Active: active since Thu 2019-09-05...
  12. K

    Deleted ceph on node, stupid!

    ceph -s 2019-09-23 20:35:16.098552 7f5101076700 0 librados: client.admin authentication error (1) Operation not permitted [errno 1] error connecting to the cluster
  13. K

    Deleted ceph on node, stupid!

    One big step further now THANKS sg90 and paradox55! I added the wrong name for the mon in ceph.conf. I have now added the IP address and ceph_ssd is working again. Ceph under the node still gives rados_connect failed.
  14. K

    Deleted ceph on node, stupid!

    'ceph -s' takes very long now and no report. I can see more graphics from Ceph now in Proxmox but without any data.
  15. K

    Deleted ceph on node, stupid!

    /etc/init.d/ceph status ● ceph.service - PVE activate Ceph OSD disks Loaded: loaded (/etc/systemd/system/ceph.service; enabled; vendor preset: enabled) Active: inactive (dead) since Mon 2019-09-23 18:05:11 CEST; 1h 13min ago Main PID: 2266631 (code=exited, status=0/SUCCESS) Sep 23...
  16. K

    Deleted ceph on node, stupid!

    ceph -s reports: unable to get monitor info from DNS SRV with service name: ceph-mon no monitors specified to connect to. 2019-09-23 19:13:05.896532 7f91e7c95500 -1 failed for service _ceph-mon._tcp [errno 2] error connecting to the cluster
  17. K

    Deleted ceph on node, stupid!

    I found the fsid and recreated the ceph.conf file and restarted, but I think I am missing some configs in this file (still everything up).
  18. K

    Deleted ceph on node, stupid!

    Yes everything is still there on node1, node2 and node3. It looks like 'only' ceph.conf has been deleted when I ran 'pveceph' purge on node4 (on node4 there are no containers nor vm's running).
  19. K

    Deleted ceph on node, stupid!

    Do you mean with FSID all the keyfiles in /etv/pve/priv ? Those are still there.
  20. K

    Deleted ceph on node, stupid!

    I don't have a ceph.conf backup, but have vzdumps locally (not on Ceph). But before I go the reinstall way I hope there is another faster way to get everything running. It looks like all other files are still in place.