I have a situation where a node failed (due to the boot drive failing) and then another node failed (due to RAM failure). There are 7 nodes in the cluster, so things kept running, but eventually there were many writes that could not be redundantly stored and the whole thing ground to a halt...
I have a failed boot drive in a 7 node proxmox cluster with ceph. If I replace the drive and do a fresh install, I would need to trash the OSD's attached to that node. If I could somehow recover the OSD's instead it would be great and probably save time too. Is that possible?
PBS as a VM is definitely a good idea. However, PBS generates a fingerprint that you need to save, otherwise another instance won't be able to read to backups.
You can attach storage to PBS in many ways.
On the underlying OS (Debian) you can mount the storage and simply link to it in the...
I have found that fast-diff is very useful, which requires exclusive-lock and object-map to be enabled as well.
While the selection of features at RBD image create time is nicely documented, how to modify an existing volume is not easy to find.
I wanted to enable fast-diff on images to I can...
It's been a while, but it seems that all is not well with newer versions of ceph mgr and this...
I get this:
FT1-NodeA:~# apt-get install ceph-mgr-dashboard
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
ceph-mgr-dashboard is already the newest...
You should really look into Proxmox Backup server. It will take the pain out of backups. I configured some older metal into a proxmox ceph cluster, run PBS in a VM and it's works really well.
If you want to just protect yourself against user errors or similar, use snapshots. They're much...
This morning a restart of a node that had not been restarted for quite some time caused the same symptoms as those reported here. It dawned on me that this might be due to a new running kernel that this swapping of ports occurs. On further investigation, here's what I found.
NodeA was running...
:redface: Of course, the command has to run on the node on which the container is running...!
~# pct fstrim 192
/var/lib/lxc/192/rootfs/: 88.9 GiB (95446147072 bytes) trimmed
/var/lib/lxc/192/rootfs/home/user-data/owncloud: 1.6 TiB (1795599138816 bytes) trimmed
However, when I ask rbd for the...
I don't think it's a good idea to run privileged containers for clients, not? If a UID matches one of the host's UIDs that has rights to locations a client should not have access to, it may create a big problem...
Does it mean that if you have a mountpoint (over and above the boot drive), thin-provisioning doesn't work?
~# cat /etc/pve/lxc/192.conf
Of course that gives the same result. For some reason the container believes that the storage doesn't support trimming, i.e. it's not thin provisioned. However, some other volumes on the same ceph storage pool are completely ok with trimming.
Could there be something that's set in the...
I have an LXC that is provisioned with a 100GB boot drive using ceph RBD storage. However, see the following:
~# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/rbd10 98G 8.8G 85G 10% /
This is in the running container.
Checking the disk usage in ceph however, claims...
I know the matter has been resolved, but just for reference, here's what I was referring to:
14:17:19 up 21 days, 19:44, 1 user, load average: 5.37, 5.36, 5.55
This refers to CPU's, not percentages.