Recent content by dkorzhevin

  1. CVE-2021-29657

    Hello, Please check:
  2. CEPH replication to different hosts instead of OSDs (PVE 6.2)

    Crush rules was created with commands: ceph osd crush rule create-replicated ssd-pool-rule default osd ssd ceph osd crush rule create-replicated hdd-pool-rule default osd hdd
  3. CEPH replication to different hosts instead of OSDs (PVE 6.2)

    Here is current tree from crushmap: # ceph osd crush tree --show-shadow ID CLASS WEIGHT TYPE NAME -12 ssd 10.47583 root default~ssd -9 ssd 2.61896 host cloud1~ssd 3 ssd 0.87299 osd.3 4 ssd 0.87299 osd.4 5 ssd 0.87299...
  4. CEPH replication to different hosts instead of OSDs (PVE 6.2)

    Guys could you please briefly look if steps I've mentioned are correct?
  5. CEPH replication to different hosts instead of OSDs (PVE 6.2)

    As I correctly understood, next steps should be taken: 1. ceph osd getcrushmap -o /tmp/mycrushmap 2. crushtool -d /tmp/mycrushmap > /tmp/mycrushmap.txt 3. Change ssd-pool-rule and hdd-pool-rule with (step chooseleaf firstn 0 type host): rule ssd-pool-rule { id 1 type replicated...
  6. CEPH replication to different hosts instead of OSDs (PVE 6.2)

    Hello, Could you please advice on how to safely change the replica's to be on different hosts, instead of OSDs for next crush map (PVE 6.2): # begin crush map tunable choose_local_tries 0 tunable choose_local_fallback_tries 0 tunable choose_total_tries 50 tunable chooseleaf_descend_once 1...
  7. [SOLVED] Shutting down any node makes VMs unavailable

    How I can check that distribution goes to different nodes, not to OSDs?
  8. cephfs mount error: exit code 16 (500)

    Hello, On some servers in cloud I see this error, while trying to check cephfs - content: mount error: exit code 16 (500) I have next package versions:
  9. [SOLVED] Closing xterm.js results in closing active command ?

    Try to run command inside terminal multiplexer such as tmux.
  10. Kubernates

    Hello, I'd wish to understand if anyone here running production-grade Kubernates on proxmox (either in KVM or in LXC). I found many articles on web, but most of them are about test or small k8s small or test clusters. As far as I understood (please correct me if I wrong - Kubernetes needs some...
  11. Ceph - Monitor clock skew

    No-matter what, but something should be enable by default, to sync time :)
  12. Ceph - Monitor clock skew

    Meet time sync problem with PVE 5.4 strange, that ntp time sync not enabled by default.
  13. Server notes - Permission denied (500)

    proxmox-ve: 5.4-1 (running kernel: 4.15.18-12-pve) pve-manager: 5.4-3 (running version: 5.4-3/0a6eaa62) pve-kernel-4.15: 5.3-3 pve-kernel-4.15.18-12-pve: 4.15.18-35 ceph: 12.2.11-pve1 corosync: 2.4.4-pve1 criu: 2.11.1-1~bpo90 glusterfs-client: 3.8.8-1 ksm-control-daemon: 1.2-2 libjs-extjs...


The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get your own in 60 seconds.

Buy now!