I have 3 nodes and all CTs are running of NFS server and backing up (using STOP mode) to another NFS server all over 10Gb link.
Everything is running without problem but every lets say 14 days one node during backup of CT freezes, and until I manually stop it it shows just:
INFO: starting new...
What HW nics are you using?
Right now I have uptime 10 days from last proxmox update and restart and pfsense has transfered ~435 GB without errors on any of the interfaces so I can say that pfsense can be stable on proxmox. I also have three vpn tunnels which I would know immediately if some of...
I have:
Hardware Checksum Offloading: unchecked (enabled)
Hardware TCP Segmentation Offloading: checked (disabled)
Hardware Large Receive Offloading: checked (disabled)
I thing that you have it backwards. When the option is ticked/checked the feature is disabled. (or I just misunderstand you...
I have machine which was upgraded from old proxmox versions. Everything is vorking mostly fine now. But today I noticed that there are two kernel packages ready to be updated:
Shouldn't there be just 4.15.xx in case of proxmox 5.2? I did upgrade system previously according wiki so all upgrade...
I did upgrade to 0.7.6 on Friday and after 2 days of testing I didn't have any slowdown/freezing problems as with previous versions. I was able to do all backups including largest VM and everything run smoothly, no system load spikes. I will keep testing through week and report back if problem...
Thank you. I'm trying to solve this directly with ZFS on Linux team as there seems to be more people with this problem even though it has been marked as fixed in 0.7.6 (and now moved to 0.7.7) (curiously lot of people reporting this are running Proxmox - that probably just because Proxmox is one...
Next observation:
When IO gets high (I simulated high load with rsync) I took a look at dirty pages (after I stopped rsync because I noticed increasing system load again):
cat /proc/vmstat | egrep "dirty|writeback"
nr_dirty 59
nr_writeback 32
nr_writeback_temp 0
nr_dirty_threshold 904847...
Yesterday I tried to do backups again after few weeks, I managed to backup 7 out of 8 VMs without problem. Last one (around 200 GB in size) start slowing down server as usual but then suddenly I lost connection to server entirely, I left it over night but even after 10 hours it wasn't responding...
Ok server just froze again so I quickly grabbed some stats:
zpool iostat 2
capacity operations bandwidth
pool alloc free read write read write
---------- ----- ----- ----- ----- ----- -----
DP1 436G 452G 47 53 1.53M 1.40M
rpool...
I think that post #20 shows what is exactly happening to disc (I will try testing it again tonight). What I don't understand is that even when I try to operate with data on one pair of DATA SSDs io will always hit system SSDs.
Right now I will also test to drastically lower ARC so far it was...
Somehow it's actually worst now. Today I noticed that server is getting stuck every 4 hours. Load will shoot up to like 70 (max. should be 16) for like 10 minutes. And whole server will start slowing down so much that even VMs will get affected eventually and different services will stop...
I saw update to v0.7.4-1, is patch included in this version? If it is then I have some different problem because after update I'm still experiencing same problem - system hanging with any intense IO. I was almost sure that above mentioned problem was the cause because it described almost exactly...
Ok thank you very much. I'm rushing this because as I stated before I couldn't do any backups of VMs for now more than 3 weeks because (hopefully) this bug that totally kills any data transfers and will hang system so I'm jumping to every possible solution.
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.