Search results

  1. D

    LXC container reboot fails - LXC becomes unusable

    As far as I know, this issue is only resolved by 4.18+. You may be able to use a kernel from Ubuntu or Debian Backports, but I didn't have any luck due to missing ZFS support and/or hardware modules in those kernels. I'm currently building my own kernels to track 4.19 + ZFS + hardware I need...
  2. D

    Backing up a proxmox server

    You would have had to manually select it (it's not the default).
  3. D

    Backing up a proxmox server

    Proxmox is (Debian) Linux so you may want to Google "Linux backup software". If you used ZFS for your install, you can take a snapshot and send it to a file. The issue with just copying files is that the system won't be in a crash-consistent state and you won't be getting a copy of the Master...
  4. D

    [SOLVED] ZFS Raid 10 with 4 SSD and cache...SLOW.

    You may want to check out: https://www.phoronix.com/scan.php?page=article&item=freebsd-12-zfs&num=1 Every filesystem has a use case where it shines. If you're looking for raw sequential throughput, no CoW filesystem is going to compete with ext4 in RAID0. You can try these safe tuning options...
  5. D

    [SOLVED] Proxmox 5.1.46 LXC cluster error Job for pve-container@101.service failed

    The issue is still present but less frequently encountered in the 4.15.x line. See: https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1779678 I saw it as recently as 4.15.18-8-pve and moved to custom 4.18 and 4.19 kernels afterwards. As in the bug report, I haven't seen the issue on these...
  6. D

    ZFS resilvering gone bad

    If the new drive is at least as big as the one you're replacing, you can add it without any partitioning; ZFS will take care of that for you during the replace operation. You can use either format to reference the drive but /dev/disk/by-id is the recommended approach as it won't vary if you...
  7. D

    ZFS resilvering gone bad

    Looking at the output of your zpool status, I see the "old" (original?) drive and another drive that was presumably the first replacement that failed. I would be inclined to leave those for now, add the new drive and do: zpool replace POOLNAME ORIGINAL_DRIVE SECOND_NEW_DRIVE If that does not...
  8. D

    ZFS data loss

    A little trick for next time: zfs mount -O -a That tells ZFS that it's ok to "O"verlay existing directories which will allow the mounts to succeed.
  9. D

    LXC container reboot fails - LXC becomes unusable

    Yes, a reboot will clear it up -- I'm not aware of any way to recover a system in this state without a reboot. My experience has been the same as in that Ubuntu kernel bug report; it's an infrequent condition that presents like a deadlock. We typically go months between incidents on 4.15 kernels...
  10. D

    LXC container reboot fails - LXC becomes unusable

    When the startup is hung, do grep copy_net_ns /proc/*/stack If that returns anything, you're having this issue. I can confirm that the issue is still present (but less frequent) on recent 4.15.x Proxmox PVE kernels. The Ubuntu kernel team acknowledged this bug...
  11. D

    LXC container reboot fails - LXC becomes unusable

    As far as I know, it's not related to ZFS. All of the patches are in the mainline kernel (which doesn't contain any ZFS code) related to network namespacing, NFS and cgroups. I suspect it's related to specific workloads within the containers as we only ever saw it on a hypervisors in one...
  12. D

    Need Help and/or suggestions on troubleshooting hard drive issue.

    You may have a cable or SATA port issue. check dmesg and /var/log/syslog for ATA errors. Then reconcile against: https://lime-technology.com/wiki/index.php/The_Analysis_of_Drive_Issues#Drive_Interface_Issues
  13. D

    [SOLVED] No space left on device

    What filesystem type do you have on /dev/md1? Is it BTRFS by chance? If it is and you're using snapshots, they will continue to reference any files you delete from the live filesystem. So to clear space you would need to remove files from the live filesystem AND delete any snapshots that were...
  14. D

    LXC container reboot fails - LXC becomes unusable

    NOTE: This is just a caution for the Proxmox kernel team and anyone that might be building their own kernels. As far as I know, the problematic change is not present in the official Proxmox 4.15 kernel. An issue with identical symptoms has emerged in 4.16 and is patched in 4.17. See...
  15. D

    Web GUI died

    Some things to check / try: * Use arp-scan to make sure you don't have any duplicated MAC addresses on your network * Install / use a different NIC * Confirm there is no firewall running on your workstation or anywhere between you and the Proxmox server * If you're using an ethernet cable on...
  16. D

    [SOLVED] Migrating Proxmox LXC containers with low downtime

    Also, if you're running zfs auto snapshot on the receiving end, you need disable snapshots for the receiving filesystem on HOST2 while the first send is in progress: zfs set com.sun:auto-snapshot=false zfs-vps1/subvol-105-disk-1 After step V, re-enable snapshots on the receiving end: zfs inherit...
  17. D

    [SOLVED] Migrating Proxmox LXC containers with low downtime

    "zfs receive -F" on step V should take care of your issue.
  18. D

    Random reboot - HW Failure or something else

    That does sounds like your bases are covered on the power side unless the failover / detection circuit is faulty. You can test this by removing and reinserting the power supplies one at a time to confirm the power transfers properly.
  19. D

    Random reboot - HW Failure or something else

    If you haven't explicitly set up the kernel to reboot on panic, then it would normally hang rather than reboot for most hardware faults (with a trace on console). If that server has a management device, you can check the log entries (SEL) using ipmitool. If not, reboot and enter the BIOS to see...
  20. D

    unregistered_netdevice: waiting for lo to become free

    I was seeing those errors intermittently under the 4.13 kernel line. I haven't seen any under the new testing kernel line: pve-kernel-4.15.

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!