Search results

  1. P

    Windows Guest - Slow disk performance in RBD Pool

    I recently built a dev cluster to test ceph performance, using a Windows Server 2019 guest with CrystalDiskMark I am getting very very slow speeds in read and write testing. Reads:140MB/s vs 4000MB/s testing on a disk attached to NFS storage. Writes: 90MB/s vs 1643MB/s ceph.conf [global]...
  2. P

    Removed ceph, restarted node and all nodes went down. Why?!

    I think what may have happened is after removing the 2 "failed" nodes they were still HA targets. Right now ceph in the cluster is totally broken and I've disabled it everywhere that I can. I won't be able to test again until the weekend in case it takes everything down. Here is Pve -v...
  3. P

    Removed ceph, restarted node and all nodes went down. Why?!

    We had a node failure that took down the ceph manager service, i know there should have been more than one running but ceph -s said their were 2 on standby that never took over. Ceph was completely pooched and we had to do restorations from backup and luckily managed to recover some stuff from...
  4. P

    Unable to create Ceph monitor - No Active IP for public network

    I have a 12 node cluster, 6 at each of two location. Location one nodes use .2.0/24, the other .39.0/24 Nodes can all ping one another but when trying to create a ceph monitor on any node at the second location (.39) the error states: Multiple Ceph public networks detected on putsproxp07...
  5. P

    Prevent SystemD from renaming after upgrade.

    When I upgraded my test cluster from 6.x to 7.x there were no issues. Today when upgrading one of my production nodes it appears that systemd used a new naming structure and all my interfaces changes as follows: ens3f0 - enp175s0f0 ens3f1 - enp175s0f1 ens6f0 - enp24s0f0 ens6f1 - enp24s0f1...
  6. P

    Filter or modify displayed syslog?

    My syslog on all nodes is basically page after page of: Dec 12 13:03:27 putsproxp10 corosync[3147]: [KNET ] pmtud: Starting PMTUD for host: 7 link: 0 Dec 12 13:03:27 putsproxp10 corosync[3147]: [KNET ] udp: detected kernel MTU: 1500 Dec 12 13:03:27 putsproxp10 corosync[3147]: [KNET ]...
  7. P

    Rebooting VHD host for updates.

    Is there a best practice for restarting the host of the virtual disks? The boot drives are all held in a local volume that is replicated to all the nodes but the data/storage/database disks are housed on network attached storage. I'd like to avoid manually shutting down 100+ VMs running...
  8. P

    [SOLVED] vzdump fails - sysfs write failed

    Thank you for the explanation. You can close this post.
  9. P

    [SOLVED] vzdump fails - sysfs write failed

    Unchecked the KRBD flag in the RBD config - seems to have fixed the sysfs write failed issue. If someone can explain what the krbd flag does in the RBD config that would be great Thank you
  10. P

    [SOLVED] vzdump fails - sysfs write failed

    I have searched all the logs and I do not see an indicator of what could be at fault here. I just updated the ceph to 14.2.10. same error messges 2020-08-05 07:44:40 starting migration of VM 107 to node 'putsproxp01' (192.168.2.95) 2020-08-05 07:44:41 starting VM 107 on remote node 'putsproxp01'...
  11. P

    [SOLVED] vzdump fails - sysfs write failed

    At this point I have this issue with RBD - can't map rbd volume for a disk image when trying these functions: Migrate a VM from any node (1-6) to any node (1-6) back up any VM on any node (1-6) start a new VM on any node (1-6)
  12. P

    [SOLVED] vzdump fails - sysfs write failed

    begin migration from node 2 to node 1 attached node 2 syslog output and node 1 syslog output
  13. P

    [SOLVED] vzdump fails - sysfs write failed

    pvesm status Name Type Status Total Used Available % isoRepo nfs active 2108488704 336919552 1664457728 15.98% local dir active 25413876 14911972 9187908 58.68% local-lvm...
  14. P

    [SOLVED] vzdump fails - sysfs write failed

    same on all nodes pveversion -v root@putsproxp06:~# pveversion -v proxmox-ve: 6.2-1 (running kernel: 5.4.44-2-pve) pve-manager: 6.2-10 (running version: 6.2-10/a20769ed) pve-kernel-5.4: 6.2-4 pve-kernel-helper: 6.2-4 pve-kernel-5.3: 6.1-6 pve-kernel-5.0: 6.0-11 pve-kernel-5.4.44-2-pve: 5.4.44-2...
  15. P

    [SOLVED] vzdump fails - sysfs write failed

    migration fails from any to any, not just putsproxp04
  16. P

    [SOLVED] vzdump fails - sysfs write failed

    All the same root@putsproxp06:~# cat /etc/vzdump.conf # vzdump default settings #tmpdir: DIR #dumpdir: DIR #storage: STORAGE_ID #mode: snapshot|suspend|stop #bwlimit: KBPS #ionice: PRI #lockwait: MINUTES #stopwait: MINUTES #size: MB #stdexcludes: BOOLEAN #mailto: ADDRESSLIST #maxfiles: N...
  17. P

    [SOLVED] vzdump fails - sysfs write failed

    root@putsproxp01:~# ceph osd pool ls detail pool 1 'cephfs01_data' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 autoscale_mode warn last_change 8816 flags hashpspool stripe_width 0 application cephfs pool 2 'cephfs01_metadata' replicated size 3 min_size 2...
  18. P

    [SOLVED] vzdump fails - sysfs write failed

    root@putsproxp01:~# ceph -s cluster: id: fbfde291-2831-4005-802f-5a01e95c9615 health: HEALTH_OK services: mon: 6 daemons, quorum putsproxp06,putsproxp05,putsproxp04,proxp03,proxp01,proxp02 (age 24m) mgr: putsproxp03(active, since 3w), standbys: putsproxp02, putsproxp01...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!