Hi!
I am using Proxmox 4.4 on a IBM X3550 M2 with 8 cores, 128GB ram, and a pair of 15K rpm 146GB drives (raid0).
This server is used to run as many WinXP vms as possible to simulate a high number of clients running our software and swarming the server side of the soft.
The server is part of a cluster, but all vms run from the local lvm storage and are linked clones to a local template.
I have a few issues so far and couldn't find my answers with research on this forum and google.
1. During the cloning process, we use a script that calls pvesh to clone, when the host is fairly busy, we frequently get "500 internal server error", the script launches a clone every 30sec, it doesn't do it in one shot. I have seen others mention this issue but no solution.
2. When the Proxmox host runs 120 vms, the host will be consuming around 40% of its total cpu and around 60GB of ram, plenty of resources left, but for some reason, some of the last vms to boot will have no network connectivity. (using a virtio nic, with configured vlan). Doing a release/renew or disable/enable doesn't help, rebooting the vm will most of the time fix it. We are trying to see if this is a behavior of the virtio drivers or not. Currently attempting to clone 120vms with vmxnet3 instead to verify.
3. When using spice, we couldn't connect to the console of vms beyond the 100th because of "unable to find free port 61000-61099" ? (we only had 1 spice session open)
4. When hosting 120vms (still with plenty of resources left), for example: hotplugging a nic in a vm will usually give error "got timeout", but if you connect to the vm's console, it is very responsive...
5. Any suggestions of tuning I could apply that could help (increase some memory/buffers, change scheduler) ?
Any help is very appreciated,
thank you,
Phil,
I am using Proxmox 4.4 on a IBM X3550 M2 with 8 cores, 128GB ram, and a pair of 15K rpm 146GB drives (raid0).
This server is used to run as many WinXP vms as possible to simulate a high number of clients running our software and swarming the server side of the soft.
The server is part of a cluster, but all vms run from the local lvm storage and are linked clones to a local template.
I have a few issues so far and couldn't find my answers with research on this forum and google.
1. During the cloning process, we use a script that calls pvesh to clone, when the host is fairly busy, we frequently get "500 internal server error", the script launches a clone every 30sec, it doesn't do it in one shot. I have seen others mention this issue but no solution.
2. When the Proxmox host runs 120 vms, the host will be consuming around 40% of its total cpu and around 60GB of ram, plenty of resources left, but for some reason, some of the last vms to boot will have no network connectivity. (using a virtio nic, with configured vlan). Doing a release/renew or disable/enable doesn't help, rebooting the vm will most of the time fix it. We are trying to see if this is a behavior of the virtio drivers or not. Currently attempting to clone 120vms with vmxnet3 instead to verify.
3. When using spice, we couldn't connect to the console of vms beyond the 100th because of "unable to find free port 61000-61099" ? (we only had 1 spice session open)
4. When hosting 120vms (still with plenty of resources left), for example: hotplugging a nic in a vm will usually give error "got timeout", but if you connect to the vm's console, it is very responsive...
5. Any suggestions of tuning I could apply that could help (increase some memory/buffers, change scheduler) ?
Any help is very appreciated,
thank you,
Phil,