Search results

  1. 1

    HA cluster completely broken after server maintenence

    Is there any workaround for that init-script issue? Because if I had a machine unexpectedly go down-- as I was simulating with this maintenence by shutting down PX1 "unexpectedly"-- I appear to lose quorum on the remaining two nodes.
  2. 1

    HA cluster completely broken after server maintenence

    So if I have four+ systems in my cluster I will be able to take down a system for maintenance and the cluster will remain functional? Yes, @udo, all is well now that I restarted the osds. I'm still learning the fundamentals of how these clusters are supposed to operate and didn't realize they...
  3. 1

    HA cluster completely broken after server maintenence

    osd.0 and osd.1 on PX1 and PX2, respectively, weren't coming up. They come up fine if I start them manually but why do I have to do the manual start? It seems to defeat the purpose of a high-availability cluster if I have to manually re-start the OSDs after a server goes off-line.
  4. 1

    HA cluster completely broken after server maintenence

    Jan 23 10:33:45 px1 pmxcfs[1356]: [status] notice: RRD update error /var/lib/rrdcached/db/pve2-storage/px1/local: /var/lib/rrdcached/db/pve2-storage/px1/local: illegal attempt to update using time 1485189225 when last update time is 1485189225 (minimum one second step) This is... odd. The times...
  5. 1

    HA cluster completely broken after server maintenence

    I ran pveceph start on px1 and px2 and everything came back into sync, my VMs show as running, and all the nodes are working together. But I still don't understand at all how this happened in the first place or how I can fix it. root@px2:~# pveceph start === osd.1 === create-or-move updated...
  6. 1

    HA cluster completely broken after server maintenence

    Just checked and the clock on all three systems is in sync. Nothing has been changed in any config files since it was last working. This is the result of pvecm, it is the same on all nodes other than the membership information as to which one is local: Quorum information ------------------...
  7. 1

    HA cluster completely broken after server maintenence

    My PX/Ceph HA cluster with three nodes (PX1 PX2 PX3) needed a RAM upgrade so I shut down PX1, did my work, and powered it back up. Back at my desk I took a look at the management GUI and saw "Nodes online: 3" despite PX2 and PX3's icons in the left-hand panel showing "offline". That's odd, I...
  8. 1

    Ceph nodes showing degraded until OSD Start

    Hi everyone, Got a 3-node basic Ceph+Proxmox HA cluster set up and it worked great for roughly a day. Each node has one OSD on it taking up the space of the available RAID array, it was all working great and I spun up a couple machines to test it. Come in a day later and nodes two and three are...
  9. 1

    Totally lost with Ceph

    I figured it out. I had not capitalized "Ceph.keyring" as it is in /etc/pve/storage.cfg . I suppose coming from the lawless capitalization-ignorant land of Windows has made me sloppy. Thanks for your time everyone. I just spun up a Debian test server and it looks to be working brilliantly but...
  10. 1

    Totally lost with Ceph

    Yes, I have been following that wiki page to set this cluster up. Keyring is copied as "ceph.keyring" in accordance with the name of my RBD as dictated in /etc/pve/storage.cfg. On any node I can go to Ceph->OSD in the GUI and see the OSD on each node and how big it is, but the RBD is showing 0...
  11. 1

    Totally lost with Ceph

    I have the default pool, called 'rbd', on all of my nodes. I went to Datacenter and then created a new RBD called "ceph" with the IPs of the Monitors (my three cluster nodes each have a monitor daemon), the default 'admin' user name, and "Disk Image" and "Container", "enable", etc. but when I go...
  12. 1

    Totally lost with Ceph

    So I got ceph running, got OSDs on each node (three nodes), but I'm totally lost on how to properly pool them together so that I can spin up a virtual machine and have high availability. Can anyone shine some light on how to go about just spinning up VMs and having them mirrored/checksummed...
  13. 1

    Question about Ceph and partitioning host disks

    I still don't understand what getting rid of the hot spare does for me. Sure, then I'd have a physical drive to put an OSD on, but what's the point of doing that? I've got a whole RAID array I'm trying to use here. Is there any way to get ceph to play nice with a partition instead of an entire...
  14. 1

    Question about Ceph and partitioning host disks

    My problem is that I can't create any OSDs on my /dev/cciss/c0d0p5 partition. I don't know if it needs to be mounted in a directory for Ceph to use it (tried this to the best of my ability and it didn't work since it appears Ceph expects a device) or what but "pveceph createosd...
  15. 1

    Question about Ceph and partitioning host disks

    So I finally figured out how to make the partitions I wanted by pointing cfdisk at /dev/cciss/c0d0 but now that I have the partitions I want no amount of buggery will make them show up in the Ceph GUI like they are supposed do. root@px2:~# ceph-disk prepare /dev/cciss/c0d0p5 2017-01-11...
  16. 1

    Question about Ceph and partitioning host disks

    I have an incredibly horrible not-at-all-optimal cluster going on with some older HP hardware. All three have RAID 10 with a hot spare and all have Proxmox 4.4 running on them. Proxmox was installed on each computer with a 10GB limit, thus leaving the rest of each logical RAID drive unformatted...
  17. 1

    Fresh 4.4 install -- Can SSH, no Web interface (

    Yeah the whole rack needs rewired. There are lots of switches and cruft that no longer need to be there and two in-production machines that need to be moved. I'm just glad I can finally access and test this cluster. Thanks for all the help!
  18. 1

    Fresh 4.4 install -- Can SSH, no Web interface (

    Now that I have a Windows 10 workstation I installed the Ubuntu subsystem and ran that ssh command in bash. Worked brilliantly and I'm now able to configure my cluster, which also appears to be working just fine.
  19. 1

    Fresh 4.4 install -- Can SSH, no Web interface (

    @Denny thanks for the reply. Is that ssh command meant to be run on the server itself or from a linux CLI on my workstation? The output from ss -lt|grep 8006 is: LISTEN 0 128 *:8006 *:* @dmora nmap shows Starting Nmap 7.40 ( https://nmap.org ) at...
  20. 1

    Fresh 4.4 install -- Can SSH, no Web interface (

    I'm setting up a cluster using some old Gen 5 Intel and Gen 2 AMD HPE servers (yes, I know, old stuff) running the latest version of Proxmox installed via CD. All have working RAID, network, etc. They're all in a VLAN managed by a cisco switch. I set the IP addresses, domain, etc. in the...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!