Search results

  1. A

    HP DL 380 Gen 9 issues on 5.4.73 & 5.4.78 kernel

    Yep I have tried that as well. No change. The only thing that works so far is going back to 5.4.65.
  2. A

    HP DL 380 Gen 9 issues on 5.4.73 & 5.4.78 kernel

    Here it is. Keep in mind the same thing is happening for my bnx2x adapters. Networking is 100% once I get on it and run "systemctl restart networking.service"
  3. A

    HP DL 380 Gen 9 issues on 5.4.73 & 5.4.78 kernel

    I attached some screen shots for ethtool and the service status outputs. Also attached dmesg and journalctl outputs. Keep in mind, networking starts fine once I run systemctl restart networking.service. I am starting to wonder if its related to the SCSI HP MSA Central Storage. The below...
  4. A

    HP DL 380 Gen 9 issues on 5.4.73 & 5.4.78 kernel

    Service that are failing to start on these latest kernel's. root@supprox1:~# systemctl --failed UNIT LOAD ACTIVE SUB DESCRIPTION ● pvesr.service loaded failed failed Proxmox VE replication runner ● systemd-udev-settle.service loaded failed failed udev...
  5. A

    HP DL 380 Gen 9 issues on 5.4.73 & 5.4.78 kernel

    See lots of these entries as well. [Tue Dec 1 07:03:03 2020] scsi host10: BC_298 : MBX Cmd Completion timed out [Tue Dec 1 07:03:03 2020] scsi host10: BG_1108 : MBX CMD get_boot_target Failed [Tue Dec 1 07:03:31 2020] INFO: task systemd-udevd:504 blocked for more than 120 seconds. [Tue Dec...
  6. A

    HP DL 380 Gen 9 issues on 5.4.73 & 5.4.78 kernel

    Same issue on 5.4.78.1 from the testing repo as well.
  7. A

    HP DL 380 Gen 9 issues on 5.4.73 & 5.4.78 kernel

    Attached a screen shot of the kernel message I am seeing in the logs. But that is about it.
  8. A

    HP DL 380 Gen 9 issues on 5.4.73 & 5.4.78 kernel

    Updated some of my test nodes this morning. On one of them, which is a HP DL 380 Gen9, it boots up without any networking. This is specific to 5.4.73-1, if I go back to 5.4.65-1 all is well. I don't think its a driver issue as the host is using ixgbe and bnx2x. Ethtool just reports the...
  9. A

    External Ceph Monitor

    That is pretty much along the lines I was thinking but I wanted to confirm. Appreciate the input.
  10. A

    External Ceph Monitor

    Hey guys we are getting ready to roll out a new production ceph cluster using proxmox. I wanted to get some input on monitors. Does everyone always use external monitors, or do most use their OSD nodes? I want to have 3 monitors, but was thinking of maybe 1-2 external monitors and 1 monitor...
  11. A

    Odd network issue after upgrade to Proxmox6.2

    Finally pinned this issue down. At some point I installed the numad package when the cluster was on proxmox5. Once moved to proxmox6 the the numad package was causing the VM to hang because it was moving it around to different numa nodes. Basically causing the VM to hang for periods of time.
  12. A

    Ceph Octopus

    Anyone have any input on Octopus testing? Going pretty well? We are very interested in the new replication features!
  13. A

    Ceph Issues

    I managed to totally mess up my ceph cluster. Is there any possible way to start ceph from scratch but re-use the OSD's and preserve the data?
  14. A

    Odd network issue after upgrade to Proxmox6.2

    The only other issue I can seem to find is the following. 2020-07-30 05:14:32 starting migration of VM 100 to node 'fpracprox1' (10.211.45.1) can't deactivate LV '/dev/Data/vm-100-disk-1': Logical volume Data/vm-100-disk-1 is used by another device. 2020-07-30 05:14:33 ERROR: volume...
  15. A

    Odd network issue after upgrade to Proxmox6.2

    Yep storage network is solid. Its a private dedicated 10G network.
  16. A

    Odd network issue after upgrade to Proxmox6.2

    4 node cluster of 2x HP DL 560's, 1x HP DL 380, 1x Supermicro and Nimble storage. This cluster was on 5.x for a couple years now with no issues. The issue only exists on one of the four nodes. Along with that the issue seems to only exist for the VM. At random times the VM looses network...
  17. A

    IGD Passthrough setup broke with Proxmox 6.2

    I want to confirm that these steps have resolved my issues, but I still have to use the args option in the VM config to actually get a display on a physical monitor. 1) Moved /etc/modprobe.d/vfio.conf to /etc/modprobe.d/vfio.conf.old 2) Edited /etc/default/grub by adding...
  18. A

    IGD Passthrough setup broke with Proxmox 6.2

    Yikes that doesn't sound very promising. What are the chances of getting hostpci to work properly? I would love to just add a hostpci line but the display never works.
  19. A

    IGD Passthrough setup broke with Proxmox 6.2

    Interesting I had 7 machines I was able to correct with the 5.3.18-3-pve kernel. With 5.4.34-1-pve if I add the video adapter as a PCI device in the GUI and start the VM one time, /dev/vfio/1 gets created and then I can boot the VM with the original args line. Hoping the dev's can figure...
  20. A

    IGD Passthrough setup broke with Proxmox 6.2

    Ok, so when I went back a kernel, I went back to the latest Proxmox5 kernel (4.15.18-28-pve). However, when I go back to 5.3.18-3-pve the issue is resolved and the vm starts and displays as expected. So this does look like some type of kernel issue with 5.4.34-1-pve