Search results

  1. 3

    CT throwing error

    I was able to finally clone the ct and the clone is able to be backed up. I then deleted the problem ct. I guess I'm good until the next time. Thanks for your help.
  2. 3

    CT throwing error

    These logs are pretty big. in the ceph.log there are thousands of the active+clean, but I do not see anything labeled as an error. I started a backup, had it fail and then went through the log. The log its self is concerning that it just cycles over and over. What am I actually looking for...
  3. 3

    CT throwing error

    The backup does seam to drag this thing down to its knees now. Also manual creation of snapshot does work.
  4. 3

    CT throwing error

    storage is a ceph storage across 18 drives on 3 nodes, it is stating it is healthy. arch: amd64 cores: 4 features: nesting=1 hostname: node006 memory: 4096 net0...
  5. 3

    CT throwing error

    I am getting the following error on a single CT. Just switched to Backup Server. Everything else seams ok so far. INFO: Starting Backup of VM 706 (lxc) INFO: Backup started at 2023-05-29 00:16:11 INFO: status = running INFO: CT Name: node006 INFO: including mount point rootfs ('/') in backup...
  6. 3

    Stopping Clone Process Midway

    I had made a mistake with the creating a clone, so I hit the stop button. This ended up leaving two containers stuck with the lock symbol. To clear the locks, I edited the files under /etc/pve/lxc for the CT's. Deleteing the line "Lock: Drive" from the file. Don't leave a blank line. Once the...
  7. 3

    Server crash, all vm's shut off, ceph messages.

    for those wondering I was able to edit the config files and remove the locks and the deleted snapshot. I will have to figure out how to get rid of orphaned snapshots later. For now after rebooting all nodes, everything is back up.
  8. 3

    Server crash, all vm's shut off, ceph messages.

    also some ct's are displaying a lock. when I try to unlock them, it states that the ct doesn't exist.
  9. 3

    Server crash, all vm's shut off, ceph messages.

    ceph health detail never returns with any results. ceph simply does not respond.
  10. 3

    Server crash, all vm's shut off, ceph messages.

    Thought I had gotten this fixed, but now nothing shows in ceph. I can reboot vm's and CT's, but not all of them. Some help would be appreciated, thank you gurubert.
  11. 3

    Server crash, all vm's shut off, ceph messages.

    I restarted Node C and then the error stated Node B, so I restarted Node B. Now it states Node A, after restarting A it now states Node C. Originally there were no VM's or CT's on Node C but now I have them spread out across all the nodes. Just restarting the the manager under Ceph, Monitor did...
  12. 3

    Server crash, all vm's shut off, ceph messages.

    Only error I found is the scrubs running. It keeps repeating the same for one node. Thanks for any thoughts. Error: 2023-02-01T18:09:19.419960-0700 mgr.Node-C (mgr.34310) 1363602 : cluster [DBG] pgmap v1363335: 513 pgs: 513 active+clean; 2.6 TiB data, 5.1 TiB used, 8.4 TiB / 14 TiB avail; 17...
  13. 3

    Ceph problem after removing node.

    This last step appears to be a time issue. It is now rebalancing again. Note that in my case all of my vm's and ct's are off line, even if the GUI has them running.. I did have to remove the server from /etc/ceph/ceph.conf. (Make a backup first! cp /etc/ceph/ceph.conf /etc/ceph/ceph.conf.bak)...
  14. 3

    Ceph problem after removing node.

    To remove an OSD: Note there is a waiting time to resync if you have not already done it. Let it finish! (Steps 7 & 8) Mine was already done. First, make sure that you have at least one other OSD in the cluster, as well as a quorum of monitors. SSH into the Proxmox node where the OSD you want...
  15. 3

    Ceph problem after removing node.

    No I forgot about it before I ran the command. Thanks I will go through the steps. Now I am unable to destroy the OSD disks via the gui. This is what I did to get the monitor portion cleaned up. First, make sure that you have a quorum of monitors in your Ceph cluster. This means that you...
  16. 3

    Ceph problem after removing node.

    I could add the node back in after a reload, if I really had to. But the node was deleted out of the cluster.
  17. 3

    Ceph problem after removing node.

    I have a 4 node cluster and removed one today. Ceph took 4+ hours to sync up but now shows a health warning and still shows the node that was removed as a monitor. How do I get this cleaned up? Everything is running ok, so far.
  18. 3

    How do I edit /etc/pve/corosync.conf?

    Thank you, just what I was looking for.
  19. 3

    How do I edit /etc/pve/corosync.conf?

    Screwed up a cluster setup and want to change the ip used for the connection. Can I just edit this file to the correct ip? Not looking forward to reloading it. This is a new cluster so nothing lost yet.
  20. 3

    Tried to update server and broke it

    I am in the process of doing a fresh installation. I had see the link you posted before and it did not fix the issue.