From the analysis until now, the IO pressure seems to be a cosmetic issue or rather accounting issue in the kernel. QEMU switched to using io_uring for event loops with QEMU 10.2. The issue appears in combination with IO threads, where a blocking...
please don't double post,
https://forum.proxmox.com/threads/applying-pve-qemu-kvm-10-2-1-1-may-cause-extremely-high-%E2%80%9Ci-o-delay%E2%80%9D-and-extremely-high-%E2%80%9Ci-o-pressure-stalls%E2%80%9D-patches-in-the-test-repository.182186/page-2...
the thing is that the host itself don't mount the rbd. It's done directly by each qemu process. the librbd is inside qemu, and this is the qemu process which is doing to connections to monitor && osd.
(with kbrd it's different, as it's done by...
if you have a vyos router, the best way is to peer with evpn with other proxmox nodes. (1 peer by vrf. 1 proxmox each zone is a vrf).
if your coke tenant is a layer2 behind vyos, vyos should announce their ip/mac inside the evpn zone.
It's really done when the qemu process is starting when librbd is initialized. (like other tuning that you can do in ceph.conf). It could be possible to pass params to qemu commandline, but qemu rbd driver have limitations of parameters that you...
I 'm not sure that hook can work with live migration, as the arp entry need to be set after source vm stop (the arp entry need to be flushed from source horst) and before the resume of the target vm.
This is a thing that we should implemented...
I finally solved problem thank u for help
The root was my network devices configurated as stack and corosync was connected to different units, thats caused link flapping, when i connected to 1 unit problem was gone
Why? Transmission is encrypted, you don't need Wireguard.
I reframe that so it gets easier for a beginner to understand. I know this, because I only learnt that a few days ago :)
When the backup is running, and the VM wants to do a write, the...
try to enable fleecing option in backup scheduling advanced option.
when backup is running, the new writes are going first to the pbs storage and that can give you slowdown if the pbs is slow or have more lantency or have limited bandwidth.
The...
If can really confirm that memory allocator impact librbd latency. you can with a simple fio 4k randread or write.
my last qemu patch from 2023 :
https://lists.proxmox.com/pipermail/pve-devel/2023-May/056815.html
default malloc: 60k iops 4k...
for hardening, you can do something like
DisableForwarding yes
Match User root Address <<10.0.10.0/24,10.0.20.0/24>>
DisableForwarding no
AllowTcpForwarding yes
X11Forwarding no
AllowAgentForwarding no
PermitTunnel no
(replace...
corosync mtu is always lower than real mtu.
in my production, I have also 1397 pmtud with 1500 mtu on nic
Feb 19 11:05:20 corosync[23618]: [KNET ] pmtud: PMTUD link change for host: 4 link: 0 from 469 to 1397
the current code is
my $secure_delete_cmd = sub {
my ($lvmpath) = @_;
my $stepsize = $scfg->{'saferemove-stepsize'} // 32;
$stepsize = $stepsize * 1024 * 1024;
my $bdev = abs_path($lvmpath);
my...
Hi Tchaikov, I think that the user setup is hyperconverged with proxmox/ceph on the 3 nodes. So, I think that the rbd client is able to handle it without hook . could you confirm this ? (I have seen other proxmox users doing it in hyperconverged...
it's look like to nic of the 3th is going down/up or flapping.
nic driver bug ? maybe bad cable ?
do you have any kernel log on the 3th node ? #dmesg ?
maybe also try without bonding/lacp with 2 corosync links