I'm trying to test out a Proxmox/Ceph cluster and the gui/storage seems to stop working for all storage related tasks/info when I setup Ceph.
I setup a nested proxmox cluster (wiki/Nested_Virtualization) and everything seems to work with that. Hardware server is running pve-manager: 3.4-9 with kernel: 3.10.0-11-pve. I have three VMs setup:
OS: PVE 3.4-9, Kernel 2.6.32-40-pv,
NICs: VIRTIO0 - Bridge to Internet connected NIC
VIRTIO1 - Bridge to be used for Proxmox/VMs
VIRTIO2 - Bridge to be used for Ceph
Hard Drives: VIRTIO0 - For Proxmox
VIRTIO1 - For Ceph Journal
VIRTIO2 - For Ceph disk #1
VIRTIO3 - For Ceph disk #2
Everything seems to run fine with the nested proxmox cluster until I setup ceph (wiki - Ceph_Server). I have ceph installed, monitors setup, osd disks setup and a pool created. Now if I try to check content through the gui I get communication failure. If I try and create a VM, the Hard Disk -> Storage box is grey/unavailable. Once I get this communication failure, everything related to storage gets a communication failure as well. Anything that makes a call to /api2/json/nodes/pc1/storage generates a error. Once the storage timeouts start, I may get timeouts in other parts of the gui and the graphs do not show any data. Making a request through the gui for storage related information seems to start this. The same thing happens on all three nested proxmox nodes. I've started over from bare metal install on this configuration several times, double-checking every step along the way.
I've tried creating a VM from the command line with local storage and that works, but the storage is not created.
Any ideas on what is going on would be appreciated.
From access.log
From syslog
Top level version information
Nested version information
Virtualization working on VM. CPUINFO
VM Config file
VM network (have tried with configuring eth1/eth2 instead of vmbr1/vmbr2)
Storage.cfg
Ceph Status
I setup a nested proxmox cluster (wiki/Nested_Virtualization) and everything seems to work with that. Hardware server is running pve-manager: 3.4-9 with kernel: 3.10.0-11-pve. I have three VMs setup:
OS: PVE 3.4-9, Kernel 2.6.32-40-pv,
NICs: VIRTIO0 - Bridge to Internet connected NIC
VIRTIO1 - Bridge to be used for Proxmox/VMs
VIRTIO2 - Bridge to be used for Ceph
Hard Drives: VIRTIO0 - For Proxmox
VIRTIO1 - For Ceph Journal
VIRTIO2 - For Ceph disk #1
VIRTIO3 - For Ceph disk #2
Everything seems to run fine with the nested proxmox cluster until I setup ceph (wiki - Ceph_Server). I have ceph installed, monitors setup, osd disks setup and a pool created. Now if I try to check content through the gui I get communication failure. If I try and create a VM, the Hard Disk -> Storage box is grey/unavailable. Once I get this communication failure, everything related to storage gets a communication failure as well. Anything that makes a call to /api2/json/nodes/pc1/storage generates a error. Once the storage timeouts start, I may get timeouts in other parts of the gui and the graphs do not show any data. Making a request through the gui for storage related information seems to start this. The same thing happens on all three nested proxmox nodes. I've started over from bare metal install on this configuration several times, double-checking every step along the way.
I've tried creating a VM from the command line with local storage and that works, but the storage is not created.
Any ideas on what is going on would be appreciated.
From access.log
xxx - root-at-pam [20/Aug/2015:10:29:52 -0700] "GET /api2/json/nodes/pc1/storage?content=images HTTP/1.1" 596 -
xxx - root-at-pam [20/Aug/2015:10:29:52 -0700] "GET /api2/json/nodes/pc1/storage?content=iso HTTP/1.1" 596 -
xxx - root-at-pam [20/Aug/2015:10:30:19 -0700] "GET /api2/json/nodes/pc1/storage/local/status HTTP/1.1" 596 -
xxx - root-at-pam [20/Aug/2015:10:30:49 -0700] "GET /api2/json/nodes/pc1/storage/RBD_Pool1/status HTTP/1.1" 596 -
xxx - root-at-pam [20/Aug/2015:10:32:20 -0700] "GET /api2/json/nodes/pc1/storage/local/status HTTP/1.1" 596 -
xxx - root-at-pam [20/Aug/2015:10:32:50 -0700] "GET /api2/json/nodes/pc1/storage/RBD_Pool1/status HTTP/1.1" 596 -
From syslog
Aug 20 10:29:36 pc1 pveproxy[87782]: proxy detected vanished client connection
Aug 20 10:33:06 pc1 pvestatd[3807]: status update time (300.102 seconds)
Aug 20 10:35:22 pc1 pveproxy[87780]: proxy detected vanished client connection
Top level version information
proxmox-ve-2.6.32: 3.4-160 (running kernel: 3.10.0-11-pve)
pve-manager: 3.4-9 (running version: 3.4-9/4b51d87a)
pve-kernel-2.6.32-40-pve: 2.6.32-160
pve-kernel-3.10.0-11-pve: 3.10.0-36
pve-kernel-2.6.32-29-pve: 2.6.32-126
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.7-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.10-3
pve-cluster: 3.0-18
qemu-server: 3.4-6
pve-firmware: 1.1-4
libpve-common-perl: 3.0-24
libpve-access-control: 3.0-16
libpve-storage-perl: 3.0-33
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-8
vzctl: 4.0-1pve6
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 2.2-11
ksm-control-daemon: 1.1-1
glusterfs-client: 3.5.2-1
Nested version information
proxmox-ve-2.6.32: 3.4-160 (running kernel: 2.6.32-40-pve)
pve-manager: 3.4-9 (running version: 3.4-9/4b51d87a)
pve-kernel-2.6.32-40-pve: 2.6.32-160
pve-kernel-2.6.32-39-pve: 2.6.32-157
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.7-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.10-3
pve-cluster: 3.0-18
qemu-server: 3.4-6
pve-firmware: 1.1-4
libpve-common-perl: 3.0-24
libpve-access-control: 3.0-16
libpve-storage-perl: 3.0-33
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-8
vzctl: 4.0-1pve6
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 2.2-11
ksm-control-daemon: 1.1-1
glusterfs-client: 3.5.2-1
Virtualization working on VM. CPUINFO
flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx lm
constant_tsc arch_perfmon rep_good unfair_spinlock pni vmx ssse3 cx16 sse4_1 x2apic hypervisor lahf_lm vnmi
VM Config file
args: -enable-kvm
bootdisk: virtio0
cores: 2
cpu: host
ide2: local:iso/proxmox-ve_3.4-102d4547-6.iso,media=cdrom
memory: 12288
name: PC1
net0: virtio=92:E7:5E:18:CC:E7,bridge=vmbr0
net1: virtio=56:07:4F:EE:E6:5F,bridge=vmbr1
net2: virtio=96:781:F5:11:70,bridge=vmbr2
numa: 0
onboot: 1
ostype: l26
smbios1: uuid=e73284d5-4878-4904-beec-3b3100829b4f
sockets: 2
virtio0: local:101/vm-101-disk-1.qcow2,format=qcow2,size=15G
virtio1: local:101/vm-101-disk-2.qcow2,format=qcow2,size=15G
virtio2: local:101/vm-101-disk-3.qcow2,format=qcow2,size=25G
virtio3: local:101/vm-101-disk-4.qcow2,format=qcow2,size=25G
VM network (have tried with configuring eth1/eth2 instead of vmbr1/vmbr2)
auto lo
iface lo inet loopback
auto vmbr1
iface vmbr1 inet static
address 10.10.10.101
netmask 255.255.255.0
bridge_ports eth1
bridge_stp off
bridge_fd 0
auto vmbr2
iface vmbr2 inet static
address 10.10.11.101
netmask 255.255.255.0
bridge_ports eth2
bridge_stp off
bridge_fd 0
auto vmbr0
iface vmbr0 inet static
address xx.xx.xx.xx
netmask xx.xx.xx.xx
gateway xx.xx.xx.xx
bridge_ports eth0
bridge_stp off
bridge_fd 0
Storage.cfg
rbd: RBD_Pool1
monhost pc1
pool pool1
content images
username admin
dir: local
path /var/lib/vz
content images,iso,vztmpl,rootdir
maxfiles 0
Ceph Status
cluster fe377151-e3ac-498f-8fac-daaf98defe56
health HEALTH_OK
monmap e3: 3 mons at {0=10.10.11.101:6789/0,1=10.10.11.102:6789/0,2=10.10.11.103:6789/0}
election epoch 40, quorum 0,1,2 0,1,2
osdmap e185: 6 osds: 5 up, 5 in
pgmap v550: 320 pgs, 2 pools, 0 bytes data, 0 objects
188 MB used, 124 GB / 124 GB avail
320 active+clean