My setup:
3 nodes with proxmox 4.4-92 and Ceph Jewel, it was setup like in the video described.
So I have 2 pools: ceph-vm and a ceph-lxc.
Cluster is fine, ceph is running over sperate 10Gb network.
new node is a fresh installed and patched 4.4-92 proxmox, in the same cluster and storage network like the others.
I just executed pvecm add 192.168.3.21 on the new node.
Afterward it was part of the cluster:
Code:
Quorum information
------------------
Date: Wed Aug 9 14:19:56 2017
Quorum provider: corosync_votequorum
Nodes: 4
Node ID: 0x00000004
Ring ID: 4/244
Quorate: Yes
Votequorum information
----------------------
Expected votes: 4
Highest expected: 4
Total votes: 4
Quorum: 3
Flags: Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000004 1 192.168.3.11 (local)
0x00000001 1 192.168.3.21
0x00000002 1 192.168.3.22
0x00000003 1 192.168.3.23
If I open the ceph-lxc pool on new 4th node, I can see the lxc disk images. They are all listed.
But if I click on ceph-vm pool nothing is listed, no vm disk images.
Correct me if I am wrong: but nodes access the ceph-lxc pool with krbd drivers which are loaded.
So If the node want to access the ceph-vm node, the proxmox rbd drivers has to be installed?
This I would do with a "pveceph install --version jewel"?
Can you approve my suggestion?
I just don't want to crash my ceph cluster ;-)
Ceph nodes:
Code:
proxmox-ve: 4.4-92 (running kernel: 4.4.67-1-pve)
pve-manager: 4.4-15 (running version: 4.4-15/7599e35a)
pve-kernel-4.4.35-1-pve: 4.4.35-77
pve-kernel-4.4.67-1-pve: 4.4.67-92
lvm2: 2.02.116-pve3
corosync-pve: 2.4.2-2~pve4+1
libqb0: 1.0.1-1
pve-cluster: 4.0-52
qemu-server: 4.0-110
pve-firmware: 1.1-11
libpve-common-perl: 4.0-95
libpve-access-control: 4.0-23
libpve-storage-perl: 4.0-76
pve-libspice-server1: 0.12.8-2
vncterm: 1.3-2
pve-docs: 4.4-4
pve-qemu-kvm: 2.7.1-4
pve-container: 1.0-101
pve-firewall: 2.0-33
pve-ha-manager: 1.0-41
ksm-control-daemon: 1.2-1
glusterfs-client: 3.5.2-2+deb8u3
lxc-pve: 2.0.7-4
lxcfs: 2.0.6-pve1
criu: 1.6.0-1
novnc-pve: 0.5-9
smartmontools: 6.5+svn4324-1~pve80
zfsutils: 0.6.5.9-pve15~bpo80
ceph: 10.2.9-1~bpo80+1
Storage.cfg
Code:
dir: local
path /var/lib/vz
content vztmpl,backup,iso
zfspool: local-zfs
pool rpool/data
content images,rootdir
sparse 1
rbd: ceph-vm
monhost 172.30.3.21 172.30.3.22 172.30.3.23
content images
krbd 0
pool ceph-vm
username admin
rbd: ceph-lxc
monhost 172.30.3.21 172.30.3.22 172.30.3.23
content rootdir
krbd 1
pool ceph-lxc
username admin
nfs: Backup
export /mnt/Backup
path /mnt/pve/Backup
server 172.30.3.50
content rootdir,images,iso,vztmpl,backup
maxfiles 10
options vers=3,async