Recent content by jaxx

  1. J

    [Enhancement Request] Multi-Pool based backup jobs

    Hi, Unless I've totally missed it, or there's a trick to do such a thing, I'd have a little wish for a little enhancement on pool based backup: by allowing multiple pools being selected in a job (or excluded for an inverted effect, as you can do with the "include/exclude VM") Today, when we...
  2. J

    "complex" VMs get timeout on start

    Hi ! I'm having a bit of a pickle on starting a fairly horrible VM (because it's Windows). It has 3 disks and a GPU passthrough. I have a local NVMe Drive and a fairly good CEPH pool on this little cluster (18 spinners with NVMe WAL/journal and 2x10Gb loop) It manages to start (sometimes) if...
  3. J

    BUG: kernel NULL pointer dereference, address: 0000000000000402

    Hi all ! I'm hitting a kernel bug, possibly linked to ceph(/cephfs) that doesn't exactly look like other threads I've come across and would like to get some help :) Setup is: Three servers, two have OSDs for ceph (I know, we'll get another one one day, third (more dev and local nvme for...
  4. J

    [SOLVED] Stray Mons stuck in cephfs mountpoint

    Well... I have a few containers with sub mounted MPs to the cephfs inside ... cut the little traffic there was to be safe and umount worked a charm even without stopping the containers (but yeah, I'll plan a reboot anyways)
  5. J

    [SOLVED] Stray Mons stuck in cephfs mountpoint

    #facepalm I would have thought disabling and re-enabling the share would have naturally taken care of the mount. # mount | grep cephfs 10.137.99.1,10.137.99.10,10.137.99.2:/ on /mnt/pve/cephfs type ceph (rw,relatime,name=admin,secret=<hidden>,acl) # umount /mnt/pve/cephfs # mount | grep cephfs...
  6. J

    [SOLVED] Stray Mons stuck in cephfs mountpoint

    Hi all ! I've been moving servers around (adding a new and removing one that was 'too far' network wise, waaaay better btw for those who know :) ) I've removed a node that had OSDs, the whole bang, and accounted as a mon and before that added a server with the same setup, acting as a mon as...
  7. J

    Ceph latency remediation?

    Well, the initial plan was one node in GRA RBX and SBG, we were hitting the hurdle with SBG being too far (11ms, OVH's smokeping ( http://sbg1-sbg.smokeping.ovh.net/smokeping?target=OVH.DCs.RBX ) is on par with the latencies we see on the vRack network), so we'll be regrouping to GRA+RBX, one...
  8. J

    Ceph latency remediation?

    I actually have that page open since a while :) We're gonna order closer nodes... two same DC ones (<0.2ms), third and worst one would be at ~2ms away... though that one might end up being a mirror of some sort if it remains suboptimal, it's just not what we were hoping for (cephfs mountpoints...
  9. J

    Ceph latency remediation?

    I, basically, know, I just never had to worry about that until recently... I was merely asking if there where any crush rules or setting that would allow to have it ack writes with once it arrives to the primary OSD (letting it replicate in background to the remaining OSDs) and if it were...
  10. J

    Ceph latency remediation?

    Hi all, I'm having what seems to be a network nottleneck. Context is: one of my clients wants to revamp it's infrastructure and was happy already with PVE servers despite having only local zfs backed images, missing out on the broad possibilities offered by Ceph... I wanted to push him to go...
  11. J

    [SOLVED] CT/VM (+br/vlan) issues post 6 to 7 upgrade

    Well... explicit bridge-vids on vmbr410 did the trick :facepalm: Thanks ! I guess holidays weren't long enough JB./.
  12. J

    [SOLVED] CT/VM (+br/vlan) issues post 6 to 7 upgrade

    Yeah, It was a perfectly working solution with ifupdown1… Already beer-time here, but I’ll scroll my terminals tomorrow, bridge vlan dev bridge10 showed the three vlans as expected if I remember well, i don’t remember for vmbr410 which was holding the tagged CTs/VMs, I’ll give it a shot… I...
  13. J

    [SOLVED] CT/VM (+br/vlan) issues post 6 to 7 upgrade

    Side note: It isn't PVE7 generating the issue (nor a too recent kernel) but ifupdown2 that isn't creating bridges exactly the same way, it wasn't installed (or needed) before ... in the upgrade path, I include now ifupdown2 to be installed, and the issue arises even on a PVE6 with ifupdown2. I...
  14. J

    [SOLVED] CT/VM (+br/vlan) issues post 6 to 7 upgrade

    Oh, cr*p ... I have >60 VM/CTs on the cluster (and a couple of /23 and a handful of /24s) for half a dozen clients, that's why my numbering scheme is vast ... it hopefully hinders only hotplugging, because I don't feel like renumbering everything :) Errrm... bridge10.410 is created fine, so...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!