Putting aside the silliness of the title, I do not really see why the LACP bonding in the scenario I'm about to describe does _not_ result in >1gig combined throughput.
In this scenario my Proxmox VE cluster is 2x physical nodes. Each node having 2x1gig LACP bond (layer2+3) and the ports on the switch are LACP Active (not static).
I have a test/dev TrueNAS system I built for $reasons and it's expected to be temporary, but plenty fast. It's a Dell R720 with 64GB RAM and has 2x1gig LACP in the same way, to the same switch the other PVE nodes are on. Rocking the latest TrueNAS SCALE (Linux not FreeBSD), and is using a SAS2 HBA to a SAS2 disk shelf populated with 12x600GB 10KRPM SAS2 Enterprise HDDs (I got them for this many dollaridoos $0). The zpool is striped-mirror, so all vdevs are mirrors, I believe sync is still on default, and I don't remember if atime is on/off, but generally this is default because this is just a temporary space for me to work on some dev kubernetes stuff. The TrueNAS OS is installed on a $shitCanWorthy USB stick because for this case I don't need better (so far?) as this is temporary. There are no SSD devices at-play here for this TrueNAS system.
The TrueNAS is serving an NFS export of a single dataset. Proxmox VE I've told the cluster to mount the NFS export with v4.2 because bigger numbers are obviously better (plus who knows if I get pNFS magic because $reasons). I do not recall using any special-magic mount flags but gladly will share if asked.
This TrueNAS only has the single dataset, nothing else going on. I don't recall even setting up snapshot tasks (don't need them for this current use-case)
So... the disks are in very good health, everything is great here from a hardware operational status (so far as I can tell), so I see no indications of component/device failure/errors causing problems.
I have 3x VMs that I run each as kubernetes (k8s) nodes, as in, each VM is a node in the k8s cluster.
2x of those VMs are on ONE of my Proxmox VE nodes.
1x of the other VM is on the other Proxmox VE node.
These 3x VMs are the ONLY VMdisks that are backed by this test/dev TrueNAS system, all other VMdisks and content are on another NAS (echo $longStory).
So in this test scenario, all 3x VMs are off. I turn them on all at the same time. They spin up going brrrttt and all that. However, when I monitor the metrics on the TrueNAS system, the peak tx (transmit) does NOT exceed 1gig, in fact peaks about 800Mb/s.
This performance result is the same whether the ARC on the system is fully hot (as in I've done multiple off/ons of the VMs) or if the ARC is fully cold.
So.... considering the TrueNAS is serving requests from two different Proxmox VE nodes... WHY AM I NOT ABLE TO EXCEED 1gig even in spikes??? (
Everything I read and understand is that the two PVE nodes _SHOULD_ represent multiple TCP/sessions instead of just one, which SHOULD make exceeding 1gig trivial. ESPECIALLY when the ARC is hot. And to be clear, the 64GB of RAM in the TrueNAS system is larger than the total capacity of the 3x VM disks in entirety (even though they're mostly white space) BEFORE ARC compression.
So.... what am I doing wrong here? ((
In this scenario my Proxmox VE cluster is 2x physical nodes. Each node having 2x1gig LACP bond (layer2+3) and the ports on the switch are LACP Active (not static).
I have a test/dev TrueNAS system I built for $reasons and it's expected to be temporary, but plenty fast. It's a Dell R720 with 64GB RAM and has 2x1gig LACP in the same way, to the same switch the other PVE nodes are on. Rocking the latest TrueNAS SCALE (Linux not FreeBSD), and is using a SAS2 HBA to a SAS2 disk shelf populated with 12x600GB 10KRPM SAS2 Enterprise HDDs (I got them for this many dollaridoos $0). The zpool is striped-mirror, so all vdevs are mirrors, I believe sync is still on default, and I don't remember if atime is on/off, but generally this is default because this is just a temporary space for me to work on some dev kubernetes stuff. The TrueNAS OS is installed on a $shitCanWorthy USB stick because for this case I don't need better (so far?) as this is temporary. There are no SSD devices at-play here for this TrueNAS system.
The TrueNAS is serving an NFS export of a single dataset. Proxmox VE I've told the cluster to mount the NFS export with v4.2 because bigger numbers are obviously better (plus who knows if I get pNFS magic because $reasons). I do not recall using any special-magic mount flags but gladly will share if asked.
This TrueNAS only has the single dataset, nothing else going on. I don't recall even setting up snapshot tasks (don't need them for this current use-case)
So... the disks are in very good health, everything is great here from a hardware operational status (so far as I can tell), so I see no indications of component/device failure/errors causing problems.
I have 3x VMs that I run each as kubernetes (k8s) nodes, as in, each VM is a node in the k8s cluster.
2x of those VMs are on ONE of my Proxmox VE nodes.
1x of the other VM is on the other Proxmox VE node.
These 3x VMs are the ONLY VMdisks that are backed by this test/dev TrueNAS system, all other VMdisks and content are on another NAS (echo $longStory).
So in this test scenario, all 3x VMs are off. I turn them on all at the same time. They spin up going brrrttt and all that. However, when I monitor the metrics on the TrueNAS system, the peak tx (transmit) does NOT exceed 1gig, in fact peaks about 800Mb/s.
This performance result is the same whether the ARC on the system is fully hot (as in I've done multiple off/ons of the VMs) or if the ARC is fully cold.
So.... considering the TrueNAS is serving requests from two different Proxmox VE nodes... WHY AM I NOT ABLE TO EXCEED 1gig even in spikes??? (
Everything I read and understand is that the two PVE nodes _SHOULD_ represent multiple TCP/sessions instead of just one, which SHOULD make exceeding 1gig trivial. ESPECIALLY when the ARC is hot. And to be clear, the 64GB of RAM in the TrueNAS system is larger than the total capacity of the 3x VM disks in entirety (even though they're mostly white space) BEFORE ARC compression.
So.... what am I doing wrong here? ((