In a 5 node cluster, I had to replace some failed SSD's and now the CEPH cluster is stuck with "Reduced data availability: 40 pgs inactive, 42 pgs incomplete"
Reduced data availability: 40 pgs inactive, 42 pgs incomplete
pg 2.57 is incomplete, acting [1,35,14] (reducing pool CephFS_data...
Is it possible to share a CEPH WAL between all the OSD's, instead of having to partition the WAL?
If I have 12 drives, I have to create 12 equal partition on the WAL, and assign each partition to an OSD. Is there a better way to assign the WAL?
I have a 3 host cluster and changed one of the hardware nodes. The new machine didn't want to boot from the proxmox OS drive, so I decided to Install Proxmox onto another drive.
Before I did this, however, I removed the 3rd node (SRV3) from the cluster.
Then i wanted to re-add the node, but...
Is it possible to setup a ZFS multi node cluster, like one would do with CEPH?
Let's say I have 20x servers, each with 24 SSD's, and I want to use ZFS on each server. Can I have replicated storage across the cluster, so that in the event that a node fails, or I need to move some VM's off one...
Please help.
I tried to change my CEPH IP's from 192.168.10.0/24 to 192.168.11.0/24, but it wen't horribly wrong. My initital setup is / was as follows:
3 servers:
SRV1 - 192.168.10.241
SRV2 - 192.168.10.242
SRV3 - 192.168.10.243
I wanted to move CEPH to a 2nd IP subnet, with different network...
How do I see which /dev/sdx an OSD is running on?
I need to remove some drives from my cluster, but don't know which OSD to stop and remove, to remove /dev/sdh (for example).
Hi,
Is it possible to rename a CEPH OSD?
^Croot@virt3:~# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 87.33948 root default
-3 29.11316 host virt1
1 hdd 7.27829 osd.1 up 1.00000 1.00000
2 hdd 7.27829 osd.2 up...
Hi,
We have a 3 node cluster with CEPH, using the following hardware:
3x Supermicro server with following features per server:
128GB RAM
3x 8TB SATA HDD
2x SSD drives (intel_ssdsc2ba400g4 - 400GB DC S3710)
2x 12 core CPU (Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz
Quad port 10Gbe Inter NIC
2x...
Hi,
I am trying to stop a CEPH OSD from the command line but it doesn't want to stop:
root@virt2:~# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 65.50461 root default
-3 21.83487 host virt1
0 hdd 7.27829 osd.0 up 1.00000 1.00000...
I see the CEPH journal is 1GB.
What is the correct procedure to make it bigger?
Do I purely delete the 1G partition in fdisk, create a new one, and then specify what type?
Or is there a preferred way to do it?
Hi,
I wonder if someone can shed some light on this matter.
The environment:
3x Supermicro server with following features per server:
128GB RAM
3x 8TB SATA HDD
2x SSD drives (intel_ssdsc2ba400g4 - 400GB DC S3710)
2x 12 core CPU (Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz
Quad port 10Gbe Inter...
Hi,
I am considering adding some NVME devices to a new installation of Proxmox, but cannot really see on the forums which devices are compatible, or even which are incompatible. Does anyone know where I can find a list of compatible and working NVME's?
I want to use an NVME as ZFS or even LVM...
Hi,
I have read just about every post on this forum on this topic. Some have some useful information, but I can't get all the answers that I have questions for, hence this post.
I need to setup a 3 node high available cluster to host 14 virtual machines, Linux and Windows based. This is to...
Hi
I wonder if this is at all possible: If I setup a 3 node CEPH or GlusterFS cluster and I want to dedicated a 4port 10Gbe NIC in each server for storage network. Can I set it up in such a way that I don't need a switch, but still achieve high availability?
i.e. run cables from:
From...
Hi all,
I'm new to Proxmox, though not new to Linux or virtualization.
Can someone please tell me (I couldn't get a definite answer while searching through the forums) What is the best configuration for Proxmox on 3 or 4 nodes?
I need to virtualize about 12x physical servers and will be...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.