Nothing special in INIT log but I made more test. See below.
I have 2 proxmox VE servers : one master, one slave. The VE with NFS is on the slave server.
The Proxmox servers are the last 1.6.
pve-manager: 1.6-5 (pve-manager/1.6/5261)
running kernel: 2.6.32-4-pve
proxmox-ve-2.6.32: 1.6-25
pve-kernel-2.6.32-3-pve: 2.6.32-14
pve-kernel-2.6.32-4-pve: 2.6.32-25
qemu-server: 1.1-22
pve-firmware: 1.0-9
libpve-storage-perl: 1.0-14
vncterm: 0.9-2
vzctl: 3.0.24-1pve4
vzdump: 1.2-8
vzprocps: 2.0.11-1dso2
vzquota: 3.0.11-1
pve-qemu-kvm: 0.12.5-2
ksm-control-daemon: 1.0-4
The VE is a Debian Lenny 32bits with only Apache and Darwing Streaming Server. The VE have the FEATURES="nfs
n ". I mount one remote volume using NFS.
Seems that I have 2 differents problems:
- if on the master with the GUI, I stop the VE, the master lost connection. I found "proxwww[13960]: 500 read timeout" in syslog on the master. I have this problem with and without NFS activated. I recreated the cluster clearing all configs, destroyed the VE and mounted a copy with vzrestore but I still have the problem. The vue is not refreshed in the GUI and got Unable to load cluster table.
- on the slave in shell, I can stop/start the VE if not NFS.
- If I activate NFS on VE with nfs
n, I have 2 cases:
-- NFS remote volume not mounted: I can stop/start. VE seems to work fine.
-- NFS remote volume mounted: I can't stop. I got timeout to the vzctl stop command and found this in syslog on the slave
Oct 31 10:16:58 cerimes-vs002 kernel: "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
Oct 31 10:16:58 cerimes-vs002 kernel: vzctl D ffff8802abc1a000 0 6302 1 0x00000000
Oct 31 10:16:58 cerimes-vs002 kernel: ffff8802ad8d8000 0000000000000082 0000000000000000 0000000000012ca0
Oct 31 10:16:58 cerimes-vs002 kernel: 0000000000000030 0000000000000002 000000000000fa40 ffff88029fa83fd8
Oct 31 10:16:58 cerimes-vs002 kernel: 0000000000016940 0000000000016940 ffff8802abc1a000 ffff8802abc1a2f8
Oct 31 10:16:58 cerimes-vs002 kernel: Call Trace:
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff81314587>] ? rwsem_down_failed_common+0x8c/0xa8
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff813145ea>] ? rwsem_down_read_failed+0x22/0x2b
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff81182244>] ? call_rwsem_down_read_failed+0x14/0x30
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff81313f9d>] ? down_read+0x17/0x19
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffffa0313f01>] ? do_env_enter+0x2d/0x157 [vzmon]
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffffa031520d>] ? real_env_create+0xda5/0xdea [vzmon]
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff810cdc6f>] ? __do_fault+0x425/0x455
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff810f3d82>] ? chrdev_open+0x0/0x13e
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff810efa44>] ? __dentry_open+0x1aa/0x2a5
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffffa0315485>] ? vzcalls_ioctl+0x233/0x501 [vzmon]
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffffa030d1cd>] ? vzctl_ioctl+0x39/0x54 [vzdev]
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff810fcf06>] ? vfs_ioctl+0x21/0x6c
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff810fd454>] ? do_vfs_ioctl+0x48d/0x4cb
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff810fd4cf>] ? sys_ioctl+0x3d/0x5c
Oct 31 10:16:58 cerimes-vs002 kernel: [<ffffffff81010c12>] ? system_call_fastpath+0x16/0x1b
Another problem: The first time I try to mount the NFS volume on the VE, the mount command respond "mount.nfs: No such device". I need to type "modprobe nfs" on the proxmox slave server. After, I can mount the NFS remote volume on the VE.
And the last: If the NFS is mounted on the VE, I should type 2 times reboot command on the slave to reboot it. The first reboot do nothing.
No fine. ;-)