ceph

  1. Waschbüsch

    Unable to access ceph-based disks with krbd option enabled

    Hi all, I just upgraded my cluster to Proxmox VE 6.1 and wanted to give the updated krbd integration a spin. So, I set my rbd storage to use krbd and then tried to migrate a vm to a different node. This is what I got: 2019-12-08 11:49:22 starting migration of VM 211 to node 'srv01'...
  2. ssaman

    slow requests are blocked - very slow VMs

    Hi all, since today, we have an issue with our proxmox / ceph. We already activated ceph balancer I hope someone can help us.
  3. replace sfp+nic in a ceph cluster

    I need to replace a 10gb sfp+ 2ports nic with a similar nic that provides 4 ports instead of 2. This particular nic is serving the internodal ceph network in a meshed network configuration, so no switches inside the ring. I'm in a production 3 node cluster with ceph and latest proxmox. replica...
  4. [SOLVED] Ceph HEALTH_WARN,1 mons down

    Hi, I'm running a proxmox cluster 4.4-12 for a while with 3 nodes, each having 2 osd and running a monitor. Since a few days I have one monitor down on one node and I do not understand how to track the problem. Nothing obvious for me in the /var/log/ceph . I've no recent informations in...
  5. After update of Nov 14 - monitors fail to start

    I ran the updates which installed a new kernel. after the reboot the monitor did not start. Attempted to start from command line: systemctl status ceph-mon@proxp01.service ● ceph-mon@proxp01.service - Ceph cluster monitor daemon Loaded: loaded (/lib/systemd/system/ceph-mon@.service; enabled...
  6. [Duplicate] MGR log spoiled with auth: could not find secret_id

    Hi, my cluster is not healthy, means there were many slow request and unknown pgs. Then I noticed an error message that spoiled the MGR log heavily: 2019-11-13 12:09:34.684 7fc46549b700 0 auth: could not find secret_id=4027 2019-11-13 12:09:34.684 7fc46549b700 0 cephx: verify_authorizer could...
  7. Proxmox VE 6 CEPH renew failure

    We have a cluster of four almost identical servers: - CPU E5-2620v4@2.10GHz (2x16core) - RAM 128, 128, 256, 256 GB - 4x NIC 10Gb, 2x1Gb - 2x SSD 400GB (PVE) - 6x SSD 1TB (VM's on CEPH) - 8x HDD 1TB (backup, misc) The first server has PVE5 (will get a fresh PVE6 soon), the others have PVE6...
  8. Error message spoils MGR log: connect got BADAUTHORIZER

    Hi, my cluster is not healthy, means there were many slow request and unknown pgs. Then I noticed an error message that spoiled the MGR log heavily: 2019-11-06 11:37:39.977 7f90028d7700 0 --1- 10.97.206.96:0/3948014004 >> v1:10.97.206.93:6918/101424 conn(0x56480ee7f600 0x56480eece000 :-1...
  9. Attention: (Potential) bug in Ceph identified

    Hi, I'm affected by a severe issue with Ceph. Once there's an incident that leads to a situation that Ceph is not healthy and starts recovering from this state, the MGR is not doing it's job although the service is running. In my case the relevant MGR log is spoiled with these error messages...
  10. [SOLVED] Help Please - Ceph Public/Cluster Networks

    Background: 4 nodes 3 nodes have 2x10gbe and 2x1gbe 1 node has 1x10gbe and 2x1gbe 1 x unfi 16xg NEW 1x mikrotik 4 port 10gbe switch I had everything setup and working: 10.0.90.0/28 Proxmox corosync on 1gbe interfaces 10.095.0/28 Ceph cluster AND public network on 10gbe interfaces 10.0.50.0/24...
  11. [SOLVED] cannot remove MDS

    Hello, I wanted to use the Cephfs (I'm still testing proxmox for migrating from XCP-ng). I then wanted to remove everything, but I ended up with a standby MDS (only) and I cannot remove it. It seems to be that the Host is 'Unknown' and when I delete/stop it, I have : "entry has no host" I...
  12. ceph tooling fault when creating MDS

    good evening, i posted in another thread (https://forum.proxmox.com/threads/proxmox-6-ceph-mds-stuck-on-creating.57524/#post-268549) that was created on the same topic and just hopped on to it, but that thread seems to be dead. so i am trying my luck here to see if this is a general problem...
  13. Not able to use pveceph purge to completely remove ceph

    I'm new to Proxmox... and Linux. Trying to completely remove ceph and I cannot. Tried pveceph purge after stopping all services and got this message below. I originally installed, uninstalled, then reinstalled Ceph because I wasn't able to add a second NVMe drive from each of the three...
  14. Always On VMs / Containers

    I'm new to Proxmox and I'm a little lost. I have the system setup, Client side resources (VMs / Containers) fail back n fourth and i can manually migrate. The problem I have is when a node is down, a resources sessions are terminated and that resource is shutdown, migrated to another node and...
  15. Ceph - question before first setup, one pool or two pools

    hi, i just migrated our proxmox cluster to v6 and now i am starting to plan and repurchase hardware to replace\upgrade our storage to ceph till now we had qnap filer servers. and it is time to go ceph we have two major pools: lxc\vm\db containers (total size is small, few TB and growing slowly...
  16. Ceph OSD db and wal size

    Hello Guys! I have a big question for the ceph cluster and I need your help or your opinion. I installed a simple 3 nodes setup with Ceph. In one node has 2x146 GB HW RAID 1 + 18x 600 GB 10k SAS without RAID. (Summary we have 54 OSD device and we have to buy 3 SSD for journal) And my big...
  17. Ceph with Open VSwitch

    We want to setup a Cluster with 6 Ceph Nodes splitet in 2 Racks. In the pased we have used Linux Bridge and Bond to build the Network for Ceph 2x 10 GB each Node with 2 VLAN for Access and Sync for Ceph. the Question is: is it Recommended to built this hypervisor with Ceph Network used Open...
  18. Ceph show "slow requests are blocked" when creating / modifying CephFS

    Hi, I have noticed in Ceph log (ceph -w) an increase of "slow requests are blocked" when I create CephFS, e.g. 2019-10-14 16:41:32.083294 mon.ld5505 [INF] daemon mds.ld4465 assigned to filesystem cephfs as rank 0 2019-10-14 16:41:32.121895 mon.ld5505 [INF] daemon mds.ld4465 is now active in...
  19. Upgrade to PVE6 and Ceph Natuilus Failed

    Hi All, I'm hoping I can get some assistance here. I have been reading forums and guides to try and resolve this issue to no avail. Last night I upgraded my Proxmox VE to v6 and my Ceph to Nautilus (I followed the upgrade guide on Proxmox's website.) I assume at some point I did something wrong...
  20. [SOLVED] cluster ceph very slow when one node is offline

    Hi, We have a cluster of 4 servers (all of them are up-to-date 6.0.7 version) We have configured ceph on both, with 3 monitors. All server work well when thez are all online but when one host is down , the other are very very slow. Have you ever seen this? Regards,

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE and Proxmox Mail Gateway. We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get your own in 60 seconds.

Buy now!