Recent content by gk_emmo

  1. G

    New 3 node cluster, unable to install Ceph

    I kinda sorted the problem out partly by myself. The reason why i used bonding, that i was lacking an extra 10Gbit interface for VM-NET, and for file services, it would be useful for client to move data if needed. I thought, that if i can ping with large MTU's, and even telnet the monitor on...
  2. G

    New 3 node cluster, unable to install Ceph

    I am trying to build a new cluster. This will be my 3rd one in the last 3 years. I'd say im not a complete beginner. There are 3 nodes as a start. All nodes connected via switches with VLANs and bonding for Ceph network (2x10G). Corosync and management are on separate physical interfaces. Via...
  3. G

    Windows Server 2019 - vioscsi Warning (129) - locking up data drive

    Wow, so CEPH can't be, or most likely not the issue. That is a surprise for me. Then it can only be virtio / QEMU? Awesome.
  4. G

    Windows Server 2019 - vioscsi Warning (129) - locking up data drive

    I just read through myself on the Github link before i posted here :) I am a 100% sure that this error is occuring at others. We started to think, that maybe it is caused by the lack of bandwidth on the backend. We are using 2x10Gbit with 9000 MTU for public and cluster networks. Switches are...
  5. G

    Windows Server 2019 - vioscsi Warning (129) - locking up data drive

    Did anybody figure this out? We started to see these warnings. and problems arose on VMs with any sensitive service like Exchange, or standard SQL. We are using PVE 8.03 with Ceph, in a 7-node cluster. I've tried all the suggestions in this and other threads, without any success. It is...
  6. G

    Long heartbeat ping times on back interface seen

    Weird thing is, that the issue gone away and sorted itself. There were no change in config or kernel.
  7. G

    Long heartbeat ping times on back interface seen

    We started to see the same with a 7 node cluster after upgrading to the newest version. Kernel 6.2.16-4-pve, Ceph 17.2.6 . We never saw this before, no other changes in cluster. 60 SSD's via 2x10G fiber to switches. [WRN] OSD_SLOW_PING_TIME_BACK: Slow OSD heartbeats on back (longest 4854.443ms)...
  8. G

    Proxmox host no boot after reboot

    I tried to rule this out, DELL's built in tests were ran fine. Idrac reports no issues with the boot SSD's which are in a ZFS RAID1. Now i try to open the system discs via grub rescue, but ls gives me an incorrect dnode type error. Thanks for the tip, i will try memtest to be sure.
  9. G

    Proxmox host no boot after reboot

    Hi! I have a 5 node Ceph cluster. One of the nodes dead today, and won't boot since. It is a Dell R740, with 10 SSD's, and 2 Nvme for journal and DB. It worked for weeks. I updated packages, and did a reboot. I have 2 kernels in Grub, 5.15.85-1 and 5.15.30-2 . Neither of these (even recovery)...
  10. G

    Windows is still setting up the class configuration for this device. (Code 56)

    Thx m8 After 4 hours of fighting we finally found this thread :D
  11. G

    Migration is a nightmare

    Sorry Sir, maybe i didn't explained myself correctly. When i wrote slow, i meant it for the whole cluster. The deletion proccess from an imported disk of 500GB is still ongoing since saturday, and it is at 14%. That's what makes me mad, because all the OSD's are up, there is no network usage...
  12. G

    Migration is a nightmare

    We've tried many different configs. At the point where i exported the config there were no network added, to close out it, cuz we thought that it can cause trouble. IDE is because in the old HyperV environment it was used with an IDE controller. CPU was also switched many times. We needed...
  13. G

    Migration is a nightmare

    As an addition, i was curious, and created a new pool, with less PG. It does the same. I can't see the bottleneck, and i think we suspend the deployment, because i don't know what else i can do with it now.
  14. G

    Migration is a nightmare

    We tried with many options. It is an older VM without UEFI, with 4 cpu cores, 1 sockets, 1 virtual disk, 1 network card and 10gigs of RAM. The config now looks like this (we tried to change the CPU, and the network already, the disk is IDE from the start.) bootdisk: ide0 cores: 4 ide0...
  15. G

    Migration is a nightmare

    Dear Members! We are quite new to Proxmox, and we are in the proccess to deploy our first small cluster with 5 nodes: 5x Dell R730xd with Dual Xeons 192GB RAM each 8xSamsung PM SSD 1TB for OSD's in each node 2xSamsung NVME for WAL and DB in each node We are trying to migrate from an old VMware...