Search results

  1. E

    Small Issue Leading into Something More

    Oh, also, I am fairly certain this issue has something to do with corosync. An error from syslog: Feb 10 23:55:49 t1n4 pveproxy[1715]: Cluster not quorate - extending auth key lifetime!
  2. E

    Small Issue Leading into Something More

    Hey all, My predicament started with emails that had: - subject: Cron <root@HOSTNAME> test -x /usr/sbin/anacron || ( cd / && run-parts --report /etc/cron.daily ) - body: /usr/bin/mandb: can't set the locale; make sure $LC_* and $LANG are correct I have 9 nodes in a cluster and only 3 of them...
  3. E

    [SOLVED] Questions about ZFS and Migration/Backups

    @narrateourale Correct, I tested HA with ceph when I initially got it set up. It worked great! As for ZFS two node limitation, that's what I've read as well. You can only replicate data from the original host to only one other node. I just tested this. I was able to migrate only another node...
  4. E

    [SOLVED] Questions about ZFS and Migration/Backups

    But it still stands - how would I restore from a different cluster?
  5. E

    [SOLVED] Questions about ZFS and Migration/Backups

    Okay, I see now - I replicated the data to a node, then used this to "move" the VM onto it: mv /etc/pve/nodes/<old-node>/qemu-server/<vm-id>.conf /etc/pve/nodes/<new-node>/qemu-server/ From this link: https://forum.proxmox.com/threads/start-vm-from-replicated-zvol.55401/
  6. E

    [SOLVED] Questions about ZFS and Migration/Backups

    Also, even for HA to work - I have to replicate the data first before it works? I think maybe I'm just thinking about this all wrong. I see mentions of restoring VMs from backups but not from replications. Maybe this is my issue
  7. E

    [SOLVED] Questions about ZFS and Migration/Backups

    @narrateourale We moved from this as was, but I believe we left corosync in its default. We used 10Gb for ceph and migration, that I remember for sure. This isn't really what I mean. I am familiar with HA and that makes sense, cool. Here's what my dilemma is: say cluster 1 houses VMs that are...
  8. E

    [SOLVED] Questions about ZFS and Migration/Backups

    @narrateourale No worries, thanks for following back up nonetheless. I agree with the hard drive being the issue, but it was worse then this - the entire cluster slowed to a crawl and affected managing VMs. Oddly the VMs that we had solely on SSD pools were running fine but we couldn't manage...
  9. E

    [SOLVED] Questions about ZFS and Migration/Backups

    Okay, wow, so I had this wrong. For some reason I assumed my cluster wide pool was limited in storage capacity but my blunder is that each node individually manages the pool, allowing cluster wide access, but each has the space avail. In short, for whatever reason I expected the pool to...
  10. E

    [SOLVED] Questions about ZFS and Migration/Backups

    Hey @narrateourale Thanks for your response. Allow me to respond to your comments : We implemented a 10Gb backbone for ceph specifically. Using 5400 RPM drives I'm not sure I follow this. Can you give me an example of what you mean? Migration worked fine for ceph (we've done extensive...
  11. E

    [SOLVED] Questions about ZFS and Migration/Backups

    Also, I tried ceph and loved its feature set but it wasn't working as expected when used for production VMs.
  12. E

    [SOLVED] Questions about ZFS and Migration/Backups

    Hey all, I am writing this thread in regards to help understanding ZFS and its capabilities. From my testing, I see that ZFS can allow live migrations if you create a ZFS pool on all servers with the same name. Only the firs server you check "Add Storage:" option to allow for the cluster wide...
  13. E

    Ceph OSD Issues

    H200, for Dell R610 As coincidental as this may sound, I don't suspect it to be the backplane. Upon moving the VMs back onto this cluster, issues started arising once HA was enabled on critical VMs. I am currently testing by removing those resources from HA to see if it helps. Another note of...
  14. E

    Odd Ceph Issues

    @sg90 I don't NEED the data, no, I just want to figure out a way to delete the PGs that are giving me issues without having to delete the pool itself. So you know if that's the only way to resolve this? My concern is that this issue may rise up again in my other cluster and I can't just keep...
  15. E

    Odd Ceph Issues

    I did quite a bit of research and now see things a bit more clearly, but still don't have this figured out. The PGs are incomplete due to not enough instances of this PG. I tried to search for how to delete the PGs but there was no real clear answer. This is more of a ceph issue I take, but at...
  16. E

    Odd Ceph Issues

    After working on the warnings from Datacenter > Summary > Ceph Health, I managed to get the errors resolved (including lowering the min_size parameter). The most recent logs I am seeing (the last ones, in fact) are: pg 20.0 is incomplete, acting [13,7] pg 20.3 is incomplete, acting [7,0] pg...
  17. E

    Odd Ceph Issues

    Hey @mgiammarco , 1) No, but why is it that the SSDs on that SAME controller are just fine? The issue is with HDDs and how whatever process is doing whatever it does to provide the ceph services it does, there's no doubt about that. This is further exemplified by moving the VMs back to the...
  18. E

    Ceph OSD Issues

    @Alwin, Thank you for the follow up. No, this post is regarding a second cluster I have. The link you provided is for the first cluster. What further sets this post apart from the other post is that these VMs were previously running fine on the cluster this post is regarding (let's call it...
  19. E

    Odd Ceph Issues

    @sg90 + @Alwin Sorry for the lapse here. I removed the OSDs, replaced disks, and tried to troubleshoot quite a bit, all to no avail. To recap the issues we had started when we moved our production VMs from our other 10x node cluster to the 5x node cluster for other reasons related to our colo...
  20. E

    Ceph OSD Issues

    Hey all, I am writing today about some issues I'm having with some OSDs. To give some background data: - we have 2x separate clusters - 1x cluster with 5x nodes - another 1x cluster with 10x nodes (currently configured with only 9x for the time being) - all 15x nodes have 1x SAS drive with...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!