Hello,
I'm seeing 0.033x per second input discards on all my nodes in the cluster with a random jump to 0.1 per second. Its constant over the last 400 days of monitoring in check_mk and I see the same thing on the hypervisors.
I've upp'd RX and TX to 4096 on all node.
Unifi AGG switches...
Last night I moved all my data off the spinning disks, blew away the spinning disk pools, spent several hours troubleshooting random weird problems.
After finally getting everything going zapping and wiping the spinning discs and re-adding them I'm having a tremendous performance improvement...
I switched to the latest virtio drivers and have been fine since.
also turns out a lot of the "drops" were actually IO related issues with the spinning disks. The underlaying storage would "hang" which would cause SMB services to drop. Look at the logs in Windows to validate this. Since moved...
Hello again...
decided to BLOW AWAY all spinning disks from my cluster -> seemed to be working okay except for the summary page... its completely blank.
it shows 0 OSDs
it shows blank pgs
it shows no data about cluster activity...
Reboot the ENTIRE cluster (full shutdown) and its still...
Hello There,
I've had a cluster running since 2016 and been upgrading storage. The SSD pool has been excellent since day one but my spinning disks are causing me some serious headaches.
My largest drives are two 14TB WDC WD140EDFZ. These are almost ALWAYS at 250MS. They are on the same HBA as...
Here is what I setup (removed others). Keep in mind the spice section is minimal as I've been troubleshooting this:
##### /etc/nginx/nginx.conf #####
user www-data;
worker_processes auto;
pid /run/nginx.pid;
include /etc/nginx/modules-enabled/*.conf;
events { worker_connections 1024; }
http {...
Hello,
I have been fiddeling with a reverse proxy setup for my cluster. While I can 100% get this functional, spice/VNC are off the table.
Port 3128 has been added to nginx and shows access in the access log.
[04/Nov/2021:02:51:17 +0000] "CONNECT...
Found it.
For reference, my cluster only has two nodes with OSDs and three with monitors. The health pool was the problem. It was set to 3/2 instead of 2/2
Hello,
I followed the upgrade guide as I want to move my cluster to PVE7. Post upgrade I'm stuck with 15 objects on a single PG that REFUSE to repair. It's been in a "stuck" state now for hours. There are no slow IO issues. Ceph osd logs look clean. I followed the posted proxmox upgrade guide...
Shockingly the VMWare emulated NICs seem rock solid. My camera server hasn't needed a reboot in weeks. Monitoring server is happier too. I also switched backups to a NAS via CIFS share on the SAN subnet.
Passed through an inteli210 nic on my primary smb server to see if it works better - it does...
So virtio NIC is massively dropping packets while under load.
E1000 works perfectly under load BUT drops post backup over my 10Gbit nics. I assume its caused by the backup mode "Snapshot". In the GUI selecting disconnect for the NIC and reconnecting always instantly fixes it.
I'm trying the...
So I can reproduce this CONSTANTLY...
I'm running Blue Iris on a Windows Server 2016 VM. Once half or more of my cameras connect, recent VirtIO drivers (post .141) start producing some CRAZY latency. The e1000 works perfectly but will randomly drop after backups run FROM A COMPLETELY DIFFERENT...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.