Search results

  1. is it possible to auto trim for lxc disks?

    I have a cluster that has relatively heavy IO and consequently free space on the ceph storage is constantly constrained. I'm finding myself performing fstrim on a more and more frequent interval. Is there a way to auto trim a disk for an lxc container?
  2. preconfigured firewall rules and overrun conntrack table

    I am encountering a problem on busy servers were the nodes "inexplicably" lost connectivity with cluster partners and fence themselves off. Some investigation shows that when this happens, pve-firewall is enabled and conntrack table is full. a quick look at "virgin" iptables rules has entries...
  3. Problems resizing rbd

    Just noticed this on one of my clusters; disk resize is failing with the following error message: Resizing image: 100% complete...done. mount.nfs: Failed to resolve server rbd: Name or service not known Failed to update the container's filesystem: command 'unshare -m -- sh -c 'mount...
  4. container features deployment limited to root

    I have recently began deployment of nested containers following an orderly upgrade to 5.3, when I noticed that only a root user may actually flag the features. What is the rationale for this limitation? What are the implications of setting these flags I'm not considering?
  5. get nodes/$node/storage showing 0 byte for ceph pool

    I have this intermittent problem with storage returning 0 values for a specific rbd pool. Its only happening on one cluster, and there doesnt seem to be a corrolation to which node context is being called...
  6. lxc containers and nfs mounts

    I have a need to have a specific container access NFS mounts present on the hypervisor (they're attached via IB which is not bridgeable to the container.) I can mount them as individual bind mounts and that works, but as I need to attach approx 20 of them, it will exceed the 10 mount limit...
  7. all but one nodes showing grey ?, cluster apparently healthy

    I have an odd problem with one of my clusters. all the sudden all but one of the nodes turned grey in the GUI: Whats odd is that pvecm status shows normal; pvestatd and pveproxy are functioning normally on all nodes, there are no hung mounts, pct/qm commands process without interruption, and...
  8. [SOLVED] pvesh output changes

    Within the last month or so pvesh has changed from outputting json to "pretty" output by default. I'm trying to obtain json output again but cant figure out how; pvesh refers to format options: # pvesh USAGE: pvesh set <api_path> [OPTIONS] [FORMAT_OPTIONS] Call API PUT on <api_path>...
  9. Long delay from command issue to command execution

    I have a (what appears to be) intermittent problem with container shutdowns taking a LONG time. For example: As you can see, there is a NEARLY 7 MINUTE delay from the stop request end time to the shutdown command. What is the cause of this delay and how can it be mitigated?
  10. adventures with snapshots

    I have a new problem (well, it could be old and I just noticed it.) I have a number of containers that show any number of snapshots but when I look at the disk those snapshots dont exist. Example: pvesh get /nodes/sky12/lxc/16980/snapshot 200 OK [ { "description" : "Automatic snapshot...
  11. Question: How to externally detect a hung API on a proxmox node

    Every once in a while I have a node that no longer responds to api calls; its usually because a container is not responding to pvestatd and has to be killed in order to release pveproxy. How can I have my monitoring identify this condition? since both pveproxy and pvestatd are technically...
  12. cluster node out and cannot be restarted. how to take posession of its assets

    I have a node that was shut down properly but will no longer boot. It will be a while until I can get to it to fix but in the meantime, how do I take over its assets (VMs and CTs) that are now showing as ? status in the defunct node?
  13. ceph performance is really poor

    I have a cluster of 6 nodes, each containing 8x Intel SSDSC2BB016T7R for a total of 48 OSDs. each node has 384GB ram and 40 logical cpus. For some reason, this cluster performance is really low in comparison to other deployments. deploying the gitlab template took well over 5 minutes...
  14. Problem with ha-manager

    Today I had an interesting failure. I have a node that was misbehaving, and corosync was not able to synchronize (was getting [TOTEM ] Received message has invalid digest... ignoring.) I proceeded to move containers off of it but part way through the process the node crashed. Here is the first...
  15. snapshot reversion fails via vzdump; succeeds using rbd snap

    I have a situation where a snapshot reversion for a container (rbd backed) is failing with the error "unable to read tail (got 0 byte)" in the tasklog. doing a manual reversion using rbd snap works fine. proxmox-ve: 5.1-42 (running kernel: 4.15.3-1-pve) pve-manager: 5.1-46 (running version...
  16. Cannot add node(s) to cluster

    I am having a hard time adding SOME nodes to an existing cluster. all nodes are freshly updated and running same version of proxmox-ve, as follows: # pveversion -v proxmox-ve: 5.1-42 (running kernel: 4.15.15-1-pve) pve-manager: 5.1-51 (running version: 5.1-51/96be5354) pve-kernel-4.13: 5.1-44...
  17. rbd snapshot "write 1000 at 0 result -30" error

    While running backups, vzdump got stuck on a specific container; there is no outward indication of fault but the task isnt moving, and the syslog is getting spammed with rbd: rbd54: write 1000 at 0 result -30 the vzdump processes are not in a D state so all appears normal, but the filesize of...
  18. designate node for non compute purposes

    Is there a way to designate a cluster node as a non compute node which will disallow it to accept QM/CT? I know this can be sort-of accomplished by using HA groups but that requires tight adherence to ha-group management; I'm looking to make the behavior default on ceph OSD nodes, etc.
  19. LXC Container stuck on startup, hangs pveproxy

    I have a container that has failed to start, and hung pveproxy denying any new activity to the node, which is showing with a grey question mark in the gui. Pveproxy itself appears to be running: # service pveproxy status ● pveproxy.service - PVE API Proxy Server Loaded: loaded...
  20. snapshots through api fail. manual snapshots working

    I have a weird problem with specific containers failing to snapshot; the end result is that scheduled vzdump tasks effectively fail and disallow any further vzdump jobs until manually killed. The thing is, if I MANUALLY take a snapshot using rbd snap create it works fine- its only via api that...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE and Proxmox Mail Gateway. We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get your own in 60 seconds.

Buy now!