Hi all!
"A long time ago" we try to migrate our small office PBX (Asterisk 11 + CentOS6) to Proxmox (LXC container) and discovered problem with voice that called standstill (one way audio).
We have up to 7 concurent SIP calls (maximum 1 Mbps of voice traffic).
Only one container running on node, LA near 1.
We try to create test cluster for smooth migration of container and quick change hardware (3 different servers, from "High": 2xX5650,128GB RAM to "Low": 2xX5250, 16GB RAM), only one container running on the node at the same time and problem persist anywhere.
We try to capture traffic in 4 places and simultaneously ping each place with 100ms interval: host, container, switch (SPAN), ISP and find that container freeze for 1-1,5 seconds.
Then problem occur:
tcpdump on host showed icmp ping replies from container(!) to PC and one way RTP traffic (from ISP to container, but not on reverse direction).
tcpdump on container showing nothing (but ping replies exist on bridge|wire!), 0(zero) packets was captured in 1-1,5 sec interval - as if the container was freeze(hung), but packets existed in capture after and before this "one-way voice standstill".
In general, it looks as if the container is frozen for 1 - 1,5 sec.
At first we thought it was a problem with network bridge or veeth, but container replies to ping and not showing this on tcpdump (nothing showed then problem exist).
In dmesg, syslog, etc - no suspicious messages (such as hung, timeout, etc) - all fine everywhere.
For example from "Low" server:
Kernel, lxc, etc updates, tuning, reading and hardware changes during year doesn`t solve the problem and we ask community to help us, please - how to find the root cause the problem (veeth, lxc, zfs....)?
"A long time ago" we try to migrate our small office PBX (Asterisk 11 + CentOS6) to Proxmox (LXC container) and discovered problem with voice that called standstill (one way audio).
We have up to 7 concurent SIP calls (maximum 1 Mbps of voice traffic).
Only one container running on node, LA near 1.
We try to create test cluster for smooth migration of container and quick change hardware (3 different servers, from "High": 2xX5650,128GB RAM to "Low": 2xX5250, 16GB RAM), only one container running on the node at the same time and problem persist anywhere.
We try to capture traffic in 4 places and simultaneously ping each place with 100ms interval: host, container, switch (SPAN), ISP and find that container freeze for 1-1,5 seconds.
Then problem occur:
tcpdump on host showed icmp ping replies from container(!) to PC and one way RTP traffic (from ISP to container, but not on reverse direction).
tcpdump on container showing nothing (but ping replies exist on bridge|wire!), 0(zero) packets was captured in 1-1,5 sec interval - as if the container was freeze(hung), but packets existed in capture after and before this "one-way voice standstill".
In general, it looks as if the container is frozen for 1 - 1,5 sec.
At first we thought it was a problem with network bridge or veeth, but container replies to ping and not showing this on tcpdump (nothing showed then problem exist).
In dmesg, syslog, etc - no suspicious messages (such as hung, timeout, etc) - all fine everywhere.
For example from "Low" server:
Code:
top - 11:41:09 up 15 days, 3:12, 2 users, load average: 0,61, 0,63, 0,52
Tasks: 327 total, 1 running, 266 sleeping, 0 stopped, 0 zombie
%Cpu(s): 0,9 us, 1,8 sy, 0,0 ni, 96,1 id, 1,2 wa, 0,0 hi, 0,0 si, 0,0 st
KiB Mem : 12296532 total, 4123032 free, 5691732 used, 2481768 buff/cache
KiB Swap: 8388604 total, 8369916 free, 18688 used. 6074492 avail Mem
Code:
pveversion -v
proxmox-ve: 5.2-2 (running kernel: 4.15.18-7-pve)
pve-manager: 5.2-10 (running version: 5.2-10/6f892b40)
pve-kernel-4.15: 5.2-10
pve-kernel-4.15.18-7-pve: 4.15.18-27
pve-kernel-4.15.17-1-pve: 4.15.17-9
corosync: 2.4.2-pve5
criu: 2.11.1-1~bpo90
glusterfs-client: 3.8.8-1
ksm-control-daemon: 1.2-2
libjs-extjs: 6.0.1-2
libpve-access-control: 5.0-8
libpve-apiclient-perl: 2.0-5
libpve-common-perl: 5.0-40
libpve-guest-common-perl: 2.0-18
libpve-http-server-perl: 2.0-11
libpve-storage-perl: 5.0-30
libqb0: 1.0.1-1
lvm2: 2.02.168-pve6
lxc-pve: 3.0.2+pve1-3
lxcfs: 3.0.2-2
novnc-pve: 1.0.0-2
proxmox-widget-toolkit: 1.0-20
pve-cluster: 5.0-30
pve-container: 2.0-29
pve-docs: 5.2-8
pve-firewall: 3.0-14
pve-firmware: 2.0-5
pve-ha-manager: 2.0-5
pve-i18n: 1.0-6
pve-libspice-server1: 0.14.1-1
pve-qemu-kvm: 2.12.1-1
pve-xtermjs: 1.0-5
pve-zsync: 1.7-1
qemu-server: 5.0-38
smartmontools: 6.5+svn4324-1
spiceterm: 3.0-5
vncterm: 1.5-3
zfsutils-linux: 0.7.11-pve1~bpo1
Kernel, lxc, etc updates, tuning, reading and hardware changes during year doesn`t solve the problem and we ask community to help us, please - how to find the root cause the problem (veeth, lxc, zfs....)?