Hi,
I was attempting to rebalance my 3 node cluster by balancing the amount of space between my nodes. I had 14 600GB drives on 1 host and 14 146GB drives on another, with 6 600GB drives on the 3rd host.
I pulled 7 drives from host 1 and 7 drives from host 2, down out and destroyed. When I put them into their new hosts I am having an issue getting them added back in.
In this instance I am trying to add osd 7 after having just added osd 6 to this node.
I have removed the auth and the osd's using:
ceph auth del osd.7
ceph osd rm ods.7
Now when I try to add a new osd I receive an error. The process I follow is:
ceph-disk zap /dev/sdX
ceph-disk prepare --bluestore /dev/sdX --osd-id 7 --osd-uuid XXXXxxxxXXX
ceph-disk activate /dev/sdX1
The error I get is:
command_with_stdin: Error EEXIST: entity osd.6 exists but key does not match
mount_activate: Failed to activate
'['ceph', '--cluster', 'ceph', '--name', 'client.bootstrap-osd', '--keyring', '/var/lib/ceph/bootstrap-osd/ceph.keyring', '-i', '-', 'osd', 'new', u'XXXXxxxxXXX']' failed with status code 17
osd.6 does exist, because I created it a few minutes prior to this, so I don't know why ceph-disk is trying to make this new disk use osd.6, especially when I have assigned it the osd-id of 7.
I am only able to add 1 osd back to this node, any others that I try just fail. I would like to learn from this and not just revert to reinstalling ceph altogether. This should be the same process that is followed to replace a failed osd
Is this not the correct way to add new osd's?
I have googled extensively, but if anyone has any insight or has had this issue before I would really appreciate the help.
I'm using Proxmox 5.1 and ceph 12.2.2
Thanks
I was attempting to rebalance my 3 node cluster by balancing the amount of space between my nodes. I had 14 600GB drives on 1 host and 14 146GB drives on another, with 6 600GB drives on the 3rd host.
I pulled 7 drives from host 1 and 7 drives from host 2, down out and destroyed. When I put them into their new hosts I am having an issue getting them added back in.
In this instance I am trying to add osd 7 after having just added osd 6 to this node.
I have removed the auth and the osd's using:
ceph auth del osd.7
ceph osd rm ods.7
Now when I try to add a new osd I receive an error. The process I follow is:
ceph-disk zap /dev/sdX
ceph-disk prepare --bluestore /dev/sdX --osd-id 7 --osd-uuid XXXXxxxxXXX
ceph-disk activate /dev/sdX1
The error I get is:
command_with_stdin: Error EEXIST: entity osd.6 exists but key does not match
mount_activate: Failed to activate
'['ceph', '--cluster', 'ceph', '--name', 'client.bootstrap-osd', '--keyring', '/var/lib/ceph/bootstrap-osd/ceph.keyring', '-i', '-', 'osd', 'new', u'XXXXxxxxXXX']' failed with status code 17
osd.6 does exist, because I created it a few minutes prior to this, so I don't know why ceph-disk is trying to make this new disk use osd.6, especially when I have assigned it the osd-id of 7.
I am only able to add 1 osd back to this node, any others that I try just fail. I would like to learn from this and not just revert to reinstalling ceph altogether. This should be the same process that is followed to replace a failed osd
Is this not the correct way to add new osd's?
I have googled extensively, but if anyone has any insight or has had this issue before I would really appreciate the help.
I'm using Proxmox 5.1 and ceph 12.2.2
Thanks