What is the best processor configuration (VM performance) for these specifications? I'm using "host", I intend to create a replication with the same specifications in the future.
Should I change the processor of each VM?
Hello there, I'm currently working with resolving some performance issues on a Proxmox installation.
There is a ZFS array on the network which is hosting the VMs, and the behavior is strange.
Currently, the server is hosting 3 Windows 10 VMs with the configuration shown in the attached image...
I installed proxmox (using ext4 filesystem) on a 5212MB SATA SSD (used consumer-grade Sandisk SD8SB8U512G1001) to evaluate it. I run some benchmarks using fio, to get an idea of how fast it is. Before installing proxmox, I benchmarked the disk with PassMark's Performance Test on Win1o...
I'm having really poor performance when copying data from one physical USB-C disk to another physical USB-A disk plugged into the same Proxmox VE-system.
The rsync below is only at 0.85kB/s. Filesystems are both ext4 and the disks are created as VM-disks: directories and raw.
In the VM doing...
I want to use the best performance configuration. I need help to find out which VM configuration is the best for me.
The best performance means the drivers use low CPU time.
I will use local storage with a separate disk. Which file system is best for sata qcow2 virtual machines?
I'd like to setup proxmox 7 to be as fast as it can possibly be with the hardware I have and that I am considering to get;
EDIT: (this will be for non critical, non server related workloads)
I would like to have a dedicated VM to pass GPU's to so that I can donate to the folding at home...
Hi, I've noticed that the CPU performance of the VMs is well below what it should be.
I have a newer one that has a CPU with less performance than this newer one, and the VM on geekbench delivers what it should deliver. A VPS on the newer one has less performance than the previous generation...
Since updating to Proxmox version 7, we have not been able to backup a single VM (Production affected)
The issue is that VZDUMP slows down and VM's on the Proxmox nodes become unresponsive.
I was able to get 450MB transfer using rsync to the VM locally to test access to the vm drives which go...
ich bin komplett neu in die Proxmox Materie eingetaucht.
Bisher hatte ich einen Home Server mit Hyper V am laufen und bin nun auf Proxmox umgestiegen.
Da ich nur einen kleinen 2HE Mini Server habe, habe ich 4 Nvmes verwendet (WD Black SN750 mit 500 GB PCIe 3.0x4) und auf diesen...
I'm having severe slowness on a VM with Server 2016. I'm not familiar with Proxmox so any advice would help. The VM is running SQL & RemoteApp. There are about 30 users RDPing into this server and I've thrown all the resources I can at it. The server and the programs on the server just crawl...
We have a large volume that we need to backup which contains 100.000.000 files, with a ∆ / day of about 50.000 files (400GB).
For the time being this file system is mounted directly in PBS using fuse kernel driver with mount -t ceph ip.srv.1,ip.srv.2,ip.srv.3,ipsrv.4:/ /mnt/mycephfs -o...
I'm fairly new to Proxmox and Linux, so please excuse my noobiness.
Trying to move away from a MacMini hosting SMB shares (the crooked Apple way), TimeMachines and running some Debian/Windows VMs via Virtualbox for Homelab stuff. Moving towards a "real" (home) server with Debian...
I have just set up an external CEPH cluster together with an external specialist.
This is configured as follows:
7.1TiB free storage
on 4 nodes with each:
12 core processors
Only NVMe SSDs & normal SSDs
Connected in Cluster Net with 10G
im trying to find out why zfs is pretty slow when it comes to read performance,
i have been testing with different systems, disks and seetings
testing directly on the disk im able to achieve some reasonable numbers not far away from specsheet => 400-650k IOPS (p4510 and some samsung based HPE)...
I setup a Proxmox cluster with 3 servers (Intel Xeon E5-2673 and 192 GB RAM each).
There are 2 Ceph Pools configured on them and separated into a NVMe- and a SSD-Pool through crush rules.
The public_network is using a dedicated 10 GBit network while the cluster_network is using a dedicated 40...
Thanks for the PVE 6.4 release! The Live-Restore feature is especially interesting to me, because I've always looked for ways to make the restore faster in order to keep disaster recovery times a minimum.
Main Node has 16 cores / 32 threads
VM 101 has 32 vCPUs, because the database...
I'm running a Server with 2 x 8 TB HDD and 1 x 240GB SSD Drive with the following config.
# zpool status
scan: scrub repaired 0B in 0 days 22:10:56 with 0 errors on Sun Apr 11 22:34:58 2021
Ich habe hier noch ein altes Board mit einem Intel J4005 rumliegen und habe als Testsystem Proxmox draufgehauen.
Proxmox selbst läuft auf einer NVMe SSD, zudem habe ich eine weitere SATA SSD und eine HDD als ZFS Single Disk hinzugefügt.
Bei einem Test via Samba ist mir aufgefallen, dass...
Hi, I plan to build my first ceph cluster and have some newbie questions. In the beginning I will start with 5 nodes, and plan to reach 50 nodes.
Those nodes quite old (CPU E3,16GB RAM, 2x1Gbps network), so I think to gain the performance in adding more nodes but not upgrading RAM or CPU.
I just started testing the pbs backup client for some advanced backup scenarios. One question of course is how to get the maximum performance out of the server that creates backups.
In multiple larger infrastructures there are so called 'backupworkers' (VMs) who have plenty of CPU and RAM as...