Search results

  1. P

    Windows Guest - Slow disk performance in RBD Pool

    I recently built a dev cluster to test ceph performance, using a Windows Server 2019 guest with CrystalDiskMark I am getting very very slow speeds in read and write testing. Reads:140MB/s vs 4000MB/s testing on a disk attached to NFS storage. Writes: 90MB/s vs 1643MB/s ceph.conf [global]...
  2. P

    Removed ceph, restarted node and all nodes went down. Why?!

    I think what may have happened is after removing the 2 "failed" nodes they were still HA targets. Right now ceph in the cluster is totally broken and I've disabled it everywhere that I can. I won't be able to test again until the weekend in case it takes everything down. Here is Pve -v...
  3. P

    Removed ceph, restarted node and all nodes went down. Why?!

    We had a node failure that took down the ceph manager service, i know there should have been more than one running but ceph -s said their were 2 on standby that never took over. Ceph was completely pooched and we had to do restorations from backup and luckily managed to recover some stuff from...
  4. P

    Unable to create Ceph monitor - No Active IP for public network

    I have a 12 node cluster, 6 at each of two location. Location one nodes use .2.0/24, the other .39.0/24 Nodes can all ping one another but when trying to create a ceph monitor on any node at the second location (.39) the error states: Multiple Ceph public networks detected on putsproxp07...
  5. P

    Prevent SystemD from renaming after upgrade.

    When I upgraded my test cluster from 6.x to 7.x there were no issues. Today when upgrading one of my production nodes it appears that systemd used a new naming structure and all my interfaces changes as follows: ens3f0 - enp175s0f0 ens3f1 - enp175s0f1 ens6f0 - enp24s0f0 ens6f1 - enp24s0f1...
  6. P

    Filter or modify displayed syslog?

    My syslog on all nodes is basically page after page of: Dec 12 13:03:27 putsproxp10 corosync[3147]: [KNET ] pmtud: Starting PMTUD for host: 7 link: 0 Dec 12 13:03:27 putsproxp10 corosync[3147]: [KNET ] udp: detected kernel MTU: 1500 Dec 12 13:03:27 putsproxp10 corosync[3147]: [KNET ]...
  7. P

    Rebooting VHD host for updates.

    Is there a best practice for restarting the host of the virtual disks? The boot drives are all held in a local volume that is replicated to all the nodes but the data/storage/database disks are housed on network attached storage. I'd like to avoid manually shutting down 100+ VMs running...
  8. P

    [SOLVED] vzdump fails - sysfs write failed

    Thank you for the explanation. You can close this post.
  9. P

    [SOLVED] vzdump fails - sysfs write failed

    Unchecked the KRBD flag in the RBD config - seems to have fixed the sysfs write failed issue. If someone can explain what the krbd flag does in the RBD config that would be great Thank you
  10. P

    [SOLVED] vzdump fails - sysfs write failed

    I have searched all the logs and I do not see an indicator of what could be at fault here. I just updated the ceph to 14.2.10. same error messges 2020-08-05 07:44:40 starting migration of VM 107 to node 'putsproxp01' (192.168.2.95) 2020-08-05 07:44:41 starting VM 107 on remote node 'putsproxp01'...
  11. P

    [SOLVED] vzdump fails - sysfs write failed

    At this point I have this issue with RBD - can't map rbd volume for a disk image when trying these functions: Migrate a VM from any node (1-6) to any node (1-6) back up any VM on any node (1-6) start a new VM on any node (1-6)
  12. P

    [SOLVED] vzdump fails - sysfs write failed

    begin migration from node 2 to node 1 attached node 2 syslog output and node 1 syslog output
  13. P

    [SOLVED] vzdump fails - sysfs write failed

    pvesm status Name Type Status Total Used Available % isoRepo nfs active 2108488704 336919552 1664457728 15.98% local dir active 25413876 14911972 9187908 58.68% local-lvm...
  14. P

    [SOLVED] vzdump fails - sysfs write failed

    same on all nodes pveversion -v root@putsproxp06:~# pveversion -v proxmox-ve: 6.2-1 (running kernel: 5.4.44-2-pve) pve-manager: 6.2-10 (running version: 6.2-10/a20769ed) pve-kernel-5.4: 6.2-4 pve-kernel-helper: 6.2-4 pve-kernel-5.3: 6.1-6 pve-kernel-5.0: 6.0-11 pve-kernel-5.4.44-2-pve: 5.4.44-2...
  15. P

    [SOLVED] vzdump fails - sysfs write failed

    migration fails from any to any, not just putsproxp04
  16. P

    [SOLVED] vzdump fails - sysfs write failed

    All the same root@putsproxp06:~# cat /etc/vzdump.conf # vzdump default settings #tmpdir: DIR #dumpdir: DIR #storage: STORAGE_ID #mode: snapshot|suspend|stop #bwlimit: KBPS #ionice: PRI #lockwait: MINUTES #stopwait: MINUTES #size: MB #stdexcludes: BOOLEAN #mailto: ADDRESSLIST #maxfiles: N...
  17. P

    [SOLVED] vzdump fails - sysfs write failed

    root@putsproxp01:~# ceph osd pool ls detail pool 1 'cephfs01_data' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 autoscale_mode warn last_change 8816 flags hashpspool stripe_width 0 application cephfs pool 2 'cephfs01_metadata' replicated size 3 min_size 2...
  18. P

    [SOLVED] vzdump fails - sysfs write failed

    root@putsproxp01:~# ceph -s cluster: id: fbfde291-2831-4005-802f-5a01e95c9615 health: HEALTH_OK services: mon: 6 daemons, quorum putsproxp06,putsproxp05,putsproxp04,proxp03,proxp01,proxp02 (age 24m) mgr: putsproxp03(active, since 3w), standbys: putsproxp02, putsproxp01...