Search results

  1. T

    cluster21 kernel: nfs: server ... not responding, timed out

    Ok, nachdem wir auf ALLEN Nodes den mount gelöscht haben per SSH umount -f und nach ca. 1h, kommen nun keine mehr rein. Ist also gelöst.
  2. T

    cluster21 kernel: nfs: server ... not responding, timed out

    Wir haben einen externen NFS gelöscht. Im Proxmox unter Rechenzentrum (cluster) -> Storage ist der auch nicht mehr eingetragen. Dennoch haben wie 4 / 7 Nodes die im syslog May 28 09:45:55 cluster21 kernel: nfs: server .... not responding, timed out werfen. Jede Minute. Au einen dieser nodes...
  3. T

    Ceph add monitor any valid prefix is expected rather than "192.168.0.0/24, 192.168.1.0/24".

    thanks very much. ok, it's from ceph monitor: Apr 14 13:19:13 storage4 ceph-mon[168545]: 2020-04-14 13:19:13.760 7f4f6446e700 -1 mon.storage4@4(electing) e9 failed to get devid for : fallback method has serial ''but no model Apr 14 13:19:13 storage4 ceph-mon[168545]: 2020-04-14 13:19:13.760...
  4. T

    Ceph add monitor any valid prefix is expected rather than "192.168.0.0/24, 192.168.1.0/24".

    Thanks, but with pveceph mon create we get: mon.storage4@-1(probing) e0 handle_auth_bad_method hmm, they didn't like 2 result (13) Permission denied we edit the /var/lib/ceph/mon/ceph-storage4/keyring to the key all other mon nodes uses, now all is fine hope this is ok. sometimes we get...
  5. T

    Ceph add monitor any valid prefix is expected rather than "192.168.0.0/24, 192.168.1.0/24".

    Thanks for answer. What do you mean with: "Both networks need to be routed between each other." ? ok now we have 4 mon nodes with all the same /var/lib/ceph/mon/ceph-..../keyring file. hope this is ok. storage1-3 hase the same file (key). add new monitor on storage4...
  6. T

    Ceph add monitor any valid prefix is expected rather than "192.168.0.0/24, 192.168.1.0/24".

    ok we edit manually the ceph config. add monitor and add the other network again to ceph.conf. But with monitor we get: Apr 14 12:43:37 storage4 ceph-mon[125140]: 2020-04-14 12:43:37.908 7f68e236e700 -1 mon.storage4@-1(probing) e0 handle_auth_bad_method hmm, they didn't like 2 result (13)...
  7. T

    Ceph add monitor any valid prefix is expected rather than "192.168.0.0/24, 192.168.1.0/24".

    thanks. Ok, is there an other workaround? f.e. create mon with ceph not with proxmox? or how are the steps? should i edit the ceph.conf ? we need the two networks. Thanks so much. system is productive, so our ceph cluster have to stay online.
  8. T

    Ceph add monitor any valid prefix is expected rather than "192.168.0.0/24, 192.168.1.0/24".

    pveceph mon create Error: any valid prefix is expected rather than "192.168.0.0/24, 192.168.1.0/24". command '/sbin/ip address show to '192.168.0.0/24, 192.168.1.0/24' up' failed: exit code 1 what is the reason? proxmox 6.1-8 ceph 14.2.8 ceph.conf: [global] <------> auth_client_required =...
  9. T

    8 Node Cluster / Host key verification failed.

    ok, we restart the node. now it works as espected. is there a way without reboot the hole node and withput impact on runnung VMs ?
  10. T

    8 Node Cluster / Host key verification failed.

    solved too, was wrong /etc/hosts 2nd line. local name cluster23 was mapped to public ip , but how can i upgrade the wrong entry in /etc/pve/.members "cluster22": { "id": 7, "online": 1, "ip": "192.168.0.22"}, "cluster23": { "id": 8, "online": 1, "ip": "51.N.N.N"}, Thanks.
  11. T

    8 Node Cluster / Host key verification failed.

    ok, after 2 hows, we got it: edit /root/.ssh/known_hosts seemse there are problems with different host keys from files: debug3: hostkeys_foreach: reading file "/root/.ssh/known_hosts" debug3: hostkeys_foreach: reading file "/etc/ssh/ssh_known_hosts" debug3: record_hostkey: found key type RSA in...
  12. T

    8 Node Cluster / Host key verification failed.

    We can migrate from all of our nodes. but from cluster22 to cluster23 it is not working due Host key verification failed. Migrate from cluster22 to cluster21 works well. migrate from cluster21 to cluster23 works well too. check: /usr/bin/ssh -v -e none -o 'BatchMode=yes' -o...
  13. T

    Snapshot hangs if qemu-guest-agent is running / Cloudlinux

    Thanks, but there is no load on storage. iotop
  14. T

    Snapshot hangs if qemu-guest-agent is running / Cloudlinux

    "If you stop the guest-agent inside the VM - you also need to disable it in the VM's options! " Thanks :) The Server is productive, so we can not test so much :/ But with running guest-agent snapshot+backup hung due freeze. It is unsafe to run without guest-agent?
  15. T

    Snapshot hangs if qemu-guest-agent is running / Cloudlinux

    ok, but if this happens this is not possible. server hungs, we can not do anythink on the server. with qemu-agent disabled we try to create backup: INFO: issuing guest-agent 'fs-freeze' command and server hungs.
  16. T

    Snapshot hangs if qemu-guest-agent is running / Cloudlinux

    Thanks. Our workarround is disable qemu agent. Can you explain how exactly to journal?
  17. T

    Snapshot hangs if qemu-guest-agent is running / Cloudlinux

    We use KVM. We use Centos7 with and Cpanel (think it does not matters) and installed qemu-guest-agent. Snapshots working fine. we install cloudlinux yesterday on this machine (based on centos7). Today we like create a new snapshot. it takes over 1h and we abborded it. some log of an other try...
  18. T

    A lot of cluster fails after upgrade 5.4 to 6.0.4

    Thanks. Ok we do corosync restart. "you really need to upgrade all nodes " ok we do start of next week. we do apt install libknet1 and then restart corosync. We have it on 2 of our 6 nodes yet. One node is on 1.11 We do have vRACK with 3Gbit/s quaranteed. We configured a tinc private Network...
  19. T

    A lot of cluster fails after upgrade 5.4 to 6.0.4

    Ok we will wait. Thanks. What are "best practice" config in "slow" latency networks? We use OVH vRACK as Corosync ring. What suggestions do you have? PS: We have one node with libknet1/stable,now 1.11-pve1 amd64 [Installiert,aktualisierbar auf: 1.13-pve1] No link errors. But all other nodes...
  20. T

    A lot of cluster fails after upgrade 5.4 to 6.0.4

    we are on ovh too. do you have a solution? we get lots of : Oct 17 18:47:48 storage2 corosync[3488]: [KNET ] link: host: 5 link: 1 is down Oct 17 18:47:48 storage2 corosync[3488]: [KNET ] host: host: 6 (passive) best link: 0 (pri: 1) Oct 17 18:47:48 storage2 corosync[3488]: [KNET ]...

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!