[SOLVED] Cluster Problem after Ceph Jewel to Luminous update

alexx

Active Member
Apr 28, 2017
7
1
43
Hi,

we updated our Ceph nodes from Jewel to Luminous. After the update the cluster seems to have a problem (see screenshot), though on console it seems fine:

Code:
# pvecm status
Quorum information
------------------
Date:             Wed Dec 27 10:32:56 2017
Quorum provider:  corosync_votequorum
Nodes:            8
Node ID:          0x00000003
Ring ID:          1/24060
Quorate:          Yes

Votequorum information
----------------------
Expected votes:   8
Highest expected: 8
Total votes:      8
Quorum:           5 
Flags:            Quorate

Membership information
----------------------
    Nodeid      Votes Name
0x00000001          1 10.1.4.1
0x00000002          1 10.1.4.2
0x00000003          1 10.1.4.3 (local)
0x00000008          1 10.1.4.101
0x00000006          1 10.1.4.102
0x00000007          1 10.1.4.103
0x00000004          1 10.1.4.106
0x00000005          1 10.1.4.107

Code:
# ceph -s
  cluster:
    id:     835fa7e9-f9ef-459e-a949-b6cfe76b9dd8
    health: HEALTH_WARN
            too many PGs per OSD (256 > max 200)
 
  services:
    mon: 3 daemons, quorum 0,1,2
    mgr: ceph01(active), standbys: ceph03, ceph02
    osd: 12 osds: 12 up, 12 in
 
  data:
    pools:   1 pools, 1024 pgs
    objects: 927k objects, 3677 GB
    usage:   10922 GB used, 33768 GB / 44690 GB avail
    pgs:     1022 active+clean
             2    active+clean+scrubbing+deep
 
  io:
    client:   70813 B/s rd, 3332 kB/s wr, 4 op/s rd, 296 op/s wr

Code:
root@ceph02:~# ceph osd versions
{
    "ceph version 12.2.2 (cf0baeeeeba3b47f9427c6c97e2144b094b7e5ba) luminous (stable)": 12
}
root@ceph02:~# ceph mon versions
{
    "ceph version 12.2.2 (cf0baeeeeba3b47f9427c6c97e2144b094b7e5ba) luminous (stable)": 3
}

Alle nodes are running the latest proxmox 4.4, the 3 ceph nodes are updated to ceph Luminous (https://pve.proxmox.com/wiki/Ceph_Jewel_to_Luminous)
 

Attachments

  • image.png
    image.png
    4.5 KB · Views: 11
I had to use the ceph Luminous repository on the none ceph nodes, too.
dist-upgrade and restarting the pvestatd fixed the problem.
 
PVE4.x comes with Ceph Hammer packages, a compatibility between one version is usually given, but Luminous is already two versions apart, that's why you needed to upgrade the packages or do a PVE 5.1 upgrade completely.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!