Is a good practice to separate Ceph data traffic in another interface or almost another VLAN and use 10 Gb or more for this traffic, You should keep in mind that modern SSDs have a very high transfer rate and need a network to match this...
To know which is the cause of this problem:
Do you have a separate VLAN or network for Ceph?
Check the network interfaces for errors, and also the smart values of disks
Is not a critical error, blueStore has recovered the failure, and the cause can be a puntual problem with the network, or some hardware element (controller, disk) but it appears to be for a low response of these OSD.
To remove the warning:
ceph...
There is no reason to loose backups, because you only are moving the storage network, not the storage itself. I mean, for example if you put another PVE and add storage pointing for this datastore yo can see all the backups of the first PVE. The...
Sometimes RAM errors in fact can be due a power supply failure. RAM needs precise voltage to reliably store data; too low voltage causes data corruption and may be the cause of the problem also.
iperf show performance in a short time, but as i told, if the buffers of the network equipment collapse the speed will be reduced. Also will be a reduced time if the destination PBS cannot write so fast.
I experienced the buffer collapse with...
It can be a problem of network congestion (packet loss/retransmissions), high latency or a problem with the buffers of your network equipment. If these problems occurs, then the TCP window size will be reduced.
How are the MTU settings?
Try to do...
Hey @ksl28,
instead of:
simply use:
This probably returns you your expected remotes, including a list of the nodes themselves and their types (pre, pbs). The output looks like:
Hope it helps! Afterwards, you can use this data as a...
hi,
this is currently expected behavior, but you can simply remove the token from the pve ui.
it's on our to-do list to remove the token as an opt-in mechanism, since the token may be used by other applications/clients too
Yo can try to use a tool like Clonezilla or Linux dd, if the disk has errors you can use ddrescue
https://clonezilla.org/downloads.php
https://www.gnu.org/software/ddrescue/ddrescue_es.html
You need to add only a cluster, and then all the nodes of the cluster appear. If you add each node of the cluster to PDM you may have duplicated (or x 3or x5 depending on the number of nodes) VM in PDM
Thank you very much for your replies. I can confirm that setting X2APIC, and IOAT DMA to enabled, solved the issue, now kernel 6.17.2-2 boots normally. I do NOT have Secure Boot enabled and SR-IOV is enabled.
My Apologies. I've set this exact config and now works fine
https://forum.proxmox.com/threads/dell-poweredge-r630-reboots-immediately-on-kernel-6-17-2-2-pve-pwr2262-cpu0704-uefi0078.177201/post-821716
Which version of Proxmox do you have at the destination? It will be a problem with the versión of Proxmox at the destination cluster, because live migration is only supported on latest PVE 8 versions