KVM VM loses network access until it is stopped and restarted

jcpham

New Member
Apr 12, 2009
19
0
1
VM 104 is the problem child on this host, but I have this same similar problem on multiple hosts at multiple sites. The problem seems to stem from the line in syslog that reads like this:kernel: vmbr1: port 2(tap104i1) entering disabled stateI've searched the forums and this seems to be a common problem for other people as well that I have yet to find a resolution for, I may not be searching well enough.The VM goes offline for the network and any client using it in general. I can connect in to the VM and access it through console, however network connectivity is severely limited. No RDP, No telnet, No http(s).Stopping and starting the VM fixes the problem until it happens again, which could be a week or a month; or even the next day.Curiously, the VM can still ping the host when this problem manifests. IIRC pings to the WAN also respond. Seems to be a bridging issue.This VM is Windows Server 2008. I have other VM's that are different version of Windows including 2012.It doesn't seem to matter which NIC I choose: virtio, rtl8139, or e1000; the issue persists. Currently I'm using the latest virtio Windows drivers.The most recent occurrence was yesterday afternoon.

Code:
/etc/pve/qemu-server# cat 104.conf#192.168.1.98
boot: dcn
bootdisk: ide0
cores: 8
cpu: host
ide0: local:104/vm-104-disk-2.qcow2,format=qcow2,cache=directsync,size=250G
ide1: local:iso/virtio-win.iso,media=cdrom,size=156894K
memory: 32768
name: TERMSERV
net1: virtio=42:7E:55:E5:87:49,bridge=vmbr1
onboot: 1
ostype: win7
sockets: 1




tail -100 /var/log/syslog
...
Jul 30 17:39:02 stfh02 pvedaemon[388875]: WARNING: ipcc_send_rec failed: Transport endpoint is not connected
Jul 30 17:39:05 stfh02 pvedaemon[388875]: authentication failure; rhost=192.168.1.10 user=root@pam msg=Authentication failure
Jul 30 17:39:15 stfh02 pvedaemon[389090]: <root@pam> successful auth for user 'root@pam'
Jul 30 17:39:22 stfh02 pvedaemon[783394]: starting vnc proxy UPID:stfh02:000BF422:1E465838:55BA998A:vncproxy:104:root@pam:
Jul 30 17:39:22 stfh02 pvedaemon[388875]: <root@pam> starting task UPID:stfh02:000BF422:1E465838:55BA998A:vncproxy:104:root@pam:
Jul 30 17:40:29 stfh02 pvedaemon[783468]: stop VM 104: UPID:stfh02:000BF46C:1E46721B:55BA99CD:qmstop:104:root@pam:
Jul 30 17:40:29 stfh02 pvedaemon[388875]: <root@pam> starting task UPID:stfh02:000BF46C:1E46721B:55BA99CD:qmstop:104:root@pam:
Jul 30 17:40:29 stfh02 kernel: vmbr1: port 2(tap104i1) entering disabled state
Jul 30 17:40:30 stfh02 pvedaemon[388875]: <root@pam> end task UPID:stfh02:000BF46C:1E46721B:55BA99CD:qmstop:104:root@pam: OK
Jul 30 17:40:30 stfh02 ntpd[47748]: Deleting interface #36 tap104i1, fe80::5421:efff:fe22:be86#123, interface stats: received=0, sent=0, dropped=0, active_time=532562 secs
Jul 30 17:40:30 stfh02 ntpd[47748]: peers refreshed
Jul 30 17:40:36 stfh02 pvedaemon[388875]: <root@pam> end task UPID:stfh02:000BF422:1E465838:55BA998A:vncproxy:104:root@pam: OK
Jul 30 17:40:36 stfh02 pvedaemon[783478]: starting vnc proxy UPID:stfh02:000BF476:1E4674E4:55BA99D4:vncproxy:104:root@pam:
Jul 30 17:40:36 stfh02 pvedaemon[389090]: <root@pam> starting task UPID:stfh02:000BF476:1E4674E4:55BA99D4:vncproxy:104:root@pam:
Jul 30 17:40:37 stfh02 qm[783480]: VM 104 qmp command failed - VM 104 not running
Jul 30 17:40:37 stfh02 pvedaemon[783478]: command '/bin/nc -l -p 5900 -w 10 -c '/usr/sbin/qm vncproxy 104 2>/dev/null'' failed: exit code 255
Jul 30 17:40:37 stfh02 pvedaemon[389090]: <root@pam> end task UPID:stfh02:000BF476:1E4674E4:55BA99D4:vncproxy:104:root@pam: command '/bin/nc -l -p 5900 -w 10 -c '/usr/sbin/qm vncproxy 104 2>/dev/null'' failed: exit code 255
Jul 30 17:40:54 stfh02 pvedaemon[783498]: start VM 104: UPID:stfh02:000BF48A:1E467BF0:55BA99E6:qmstart:104:root@pam:
Jul 30 17:40:54 stfh02 pvedaemon[389090]: <root@pam> starting task UPID:stfh02:000BF48A:1E467BF0:55BA99E6:qmstart:104:root@pam:
Jul 30 17:40:55 stfh02 kernel: device tap104i1 entered promiscuous mode
Jul 30 17:40:55 stfh02 kernel: vmbr1: port 2(tap104i1) entering forwarding state
Jul 30 17:40:56 stfh02 pvedaemon[389090]: <root@pam> end task UPID:stfh02:000BF48A:1E467BF0:55BA99E6:qmstart:104:root@pam: OK
Jul 30 17:40:57 stfh02 ntpd[47748]: Listen normally on 37 tap104i1 fe80::a04e:73ff:fed6:2554 UDP 123
Jul 30 17:40:57 stfh02 ntpd[47748]: peers refreshed
Jul 30 17:41:05 stfh02 kernel: tap104i1: no IPv6 routers present
Jul 30 17:41:54 stfh02 pvedaemon[389090]: <root@pam> starting task UPID:stfh02:000BF4E4:1E46936B:55BA9A22:vncproxy:104:root@pam:
Jul 30 17:41:54 stfh02 pvedaemon[783588]: starting vnc proxy UPID:stfh02:000BF4E4:1E46936B:55BA9A22:vncproxy:104:root@pam:
Jul 30 17:44:00 stfh02 pvedaemon[389090]: <root@pam> successful auth for user 'root@pam'
Jul 30 17:44:12 stfh02 pvedaemon[783724]: starting vnc proxy UPID:stfh02:000BF56C:1E46C92E:55BA9AAC:vncproxy:104:root@pam:
Jul 30 17:44:12 stfh02 pvedaemon[389090]: <root@pam> starting task UPID:stfh02:000BF56C:1E46C92E:55BA9AAC:vncproxy:104:root@pam:
Jul 30 17:46:15 stfh02 pvedaemon[389090]: <root@pam> end task UPID:stfh02:000BF56C:1E46C92E:55BA9AAC:vncproxy:104:root@pam: OK
Jul 30 17:47:49 stfh02 pvedaemon[48060]: worker 386717 finished
Jul 30 17:47:49 stfh02 pvedaemon[48060]: starting 1 worker(s)
Jul 30 17:47:49 stfh02 pvedaemon[48060]: worker 783985 started
Jul 30 17:48:43 stfh02 pvedaemon[388875]: <root@pam> successful auth for user 'root@pam'
Jul 30 17:49:36 stfh02 pveproxy[48089]: worker 411443 finished
Jul 30 17:49:36 stfh02 pveproxy[48089]: starting 1 worker(s)
Jul 30 17:49:36 stfh02 pveproxy[48089]: worker 784088 started
Jul 30 17:50:18 stfh02 pveproxy[48089]: worker 411450 finished
Jul 30 17:50:18 stfh02 pveproxy[48089]: starting 1 worker(s)
Jul 30 17:50:18 stfh02 pveproxy[48089]: worker 784130 started




brctl show
bridge name	bridge id		STP enabled	interfaces
vmbr0		8000.d4ae52e83df4	no		eth0
							tap102i1
							tap103i0
							tap103i1
							tap201i1
vmbr1		8000.d4ae52e83df6	no		eth1
							tap104i1
vmbr2		8000.d4ae52e83df8	no		eth2
							tap100i0
vmbr3		8000.001018fc2bbc	no		eth4
vmbr4		8000.001018fc2bbe	no		eth5




kvm --version
QEMU emulator version 2.1.2, Copyright (c) 2003-2008 Fabrice Bellard




pveversion -v
proxmox-ve-2.6.32: 3.3-139 (running kernel: 2.6.32-34-pve)
pve-manager: 3.3-5 (running version: 3.3-5/bfebec03)
pve-kernel-2.6.32-19-pve: 2.6.32-96
pve-kernel-2.6.32-16-pve: 2.6.32-82
pve-kernel-2.6.32-34-pve: 2.6.32-140
pve-kernel-2.6.32-26-pve: 2.6.32-114
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.7-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.10-1
pve-cluster: 3.0-15
qemu-server: 3.3-3
pve-firmware: 1.1-3
libpve-common-perl: 3.0-19
libpve-access-control: 3.0-15
libpve-storage-perl: 3.0-25
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-8
vzctl: 4.0-1pve6
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 2.1-10
ksm-control-daemon: 1.1-1
glusterfs-client: 3.5.2-1
 
Last edited:
Hi,
the kernel you are using is no more actual and in the newer kernel there are many network fixes.
so I would try to update to an actual kernel.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!