Search results

  1. M

    cluster replication between two locations

    I am looking for the same kind of solution, though I have a faster line to the remote site. Has anyone actually done anything like this, with CEPH?
  2. M

    VMs with dedicated NIC

    I did think about doing it like that. Is that the only way?
  3. M

    3 node cluster setup recommendations

    That is the setup I have already but I feel I could get better performance, somehow? The way I see it now, each SATA drive acts on it's own. OR does CEPH somehow stripe the data acess the spindles to improve performance? The 4th drive will be used for backup and to host ISO's. As I couldn't...
  4. M

    3 node cluster setup recommendations

    Ok, so with 3x 8TB HDD's, how would you setup the OSD's for optimium performance? I have 4x identical servers as folllows: Dual 12core XEON, 128GB RAM, 2x 300GB SSD for cache, 4x 8TB SATA for storage, 4 port 10GBe NIC.
  5. M

    VMs with dedicated NIC

    Did you ever figure this out? I also want to dedicate a specific port from a Quad port NIC to one of my VM's but can't seem to get it working
  6. M

    3 node cluster setup recommendations

    Hi, I have setup Ceph as per the Wiki, but it's still a bit unclear to me. 1. How do I specify how large an OSD is? 2. How do I combine the 4 single drives into a single large drive / partition? 3. So I understand that Ceph does a form of RAID over the network, is that correct? If so, is it...
  7. M

    New Proxmox VE 5.0 with Ceph production cluster and upgrade from 3 to 5 nodes

    Hi, I see CEPH v12.2.0 Luminous was released on the 29 August 2017 and "v12.2.1 Luminous" was released on 28 September 2017. Have you been able to incorporate it into Proxmox yet? Is it stable enough to use in Production?
  8. M

    Which NVME's are compatible?

    Thanx. The motherboard I am considering is the SuperMicro X10DRW-i How / where can I see if it would be compatible?
  9. M

    10Gbe high availability without switches?

    Interesting. I didn't see it earlier, will go check the wiki
  10. M

    3 node cluster setup recommendations

    Thanx. Have you compared this to a hardware RAID card + SSD cache, and ZFS + L2Arc cache? I'd like to see what the performance differences are. I don't have the hardware yet, so can't do this myself at this stage. Just getting as much info as I can before I buy.
  11. M

    3 node cluster setup recommendations

    Ok so I just watched the video on installing CEPH on a 3 node cluster with 10GBe NIC's - something similar to what I want to do. But, what is unclear in the video, seems like the 2x storage (or OSD?) drives are unpartitioned? Does that mean CEPH creates it's own partition? My question is this...
  12. M

    upgrade zfs-0.7.0

    Thank you. I couldn't find any notes about this in the ISO being released, and didn't want to upgrade if it's not there.
  13. M

    Can we use lvm cache in Proxmox 4.x?

    How well is this working, more than a year later?
  14. M

    3 node cluster setup recommendations

    Which filesystem are you using on the drives, before you setup CEPH?
  15. M

    10Gbe high availability without switches?

    That's what I don't know. I don't have the hardware yet, still researching to see if it can be done like this
  16. M

    [SOLVED] VE v5.0 Beta2 Install Freezes @ "create partitions 0%"

    Thanx. As boot devices, or as storage devices?
  17. M

    upgrade zfs-0.7.0

    That's why I asked Fabian if it was fixed in the new release!! Maybe in your country fix mean something else. but please stop trolling.