Sage Weil
2014-09-24 13:03:50 UTC
I run ceph on Red Hat Enterprise Linux Server 6.4 Santiago, and when I
# service ceph start
ERROR:ceph-disk:Failed to activate
/var/lib/ceph/tmp/mnt.I71N5T
mount: /dev/hioa1 already mounted or /var/lib/ceph/tmp/mnt.02sVHj busy
ceph-disk: Mounting filesystem failed: Command '['/bin/mount', '-t',
'xfs', '-o', 'noatime', '--',
'/dev/disk/by-parttypeuuid/4fbd7e29-9d25-41b8-afd0-062c0ceff05d.6d726c93-41f9-453d-858e-ab4132b5c8fd',# service ceph start
ERROR:ceph-disk:Failed to activate
/var/lib/ceph/tmp/mnt.I71N5T
mount: /dev/hioa1 already mounted or /var/lib/ceph/tmp/mnt.02sVHj busy
ceph-disk: Mounting filesystem failed: Command '['/bin/mount', '-t',
'xfs', '-o', 'noatime', '--',
'/var/lib/ceph/tmp/mnt.02sVHj']' returned non-zero exit status 32
ceph-disk: Error: One or more partitions failed to activate
Someone told me "service ceph start" still tries to call ceph-disk which
will create a filestore type OSD, and create a journal partition, is it
true?
ls -l /dev/disk/by-parttypeuuid/
lrwxrwxrwx. 1 root root 11 9? 23 16:56
45b0969e-9b03-4f30-b4c6-b4b80ceff106.00dbee5e-fb68-47c4-aa58-924c904c4383ceph-disk: Error: One or more partitions failed to activate
Someone told me "service ceph start" still tries to call ceph-disk which
will create a filestore type OSD, and create a journal partition, is it
true?
ls -l /dev/disk/by-parttypeuuid/
lrwxrwxrwx. 1 root root 11 9? 23 16:56
-> ../../hioa2
lrwxrwxrwx. 1 root root 10 9? 23 17:02
45b0969e-9b03-4f30-b4c6-b4b80ceff106.c30e5b97-b914-4eb8-8306-a9649e1c20balrwxrwxrwx. 1 root root 10 9? 23 17:02
-> ../../sdb2
lrwxrwxrwx. 1 root root 11 9? 23 16:56
4fbd7e29-9d25-41b8-afd0-062c0ceff05d.6d726c93-41f9-453d-858e-ab4132b5c8fdlrwxrwxrwx. 1 root root 11 9? 23 16:56
-> ../../hioa1
lrwxrwxrwx. 1 root root 10 9? 23 17:02
4fbd7e29-9d25-41b8-afd0-062c0ceff05d.b56ec699-e134-4b90-8f55-4952453e1b7elrwxrwxrwx. 1 root root 10 9? 23 17:02
-> ../../sdb1
lrwxrwxrwx. 1 root root 11 9? 23 16:52
89c57f98-2fe5-4dc0-89c1-f3ad0ceff2be.6d726c93-41f9-453d-858e-ab4132b5c8fdlrwxrwxrwx. 1 root root 11 9? 23 16:52
-> ../../hioa1
There seems to be two hioa1 partitions there, maybe remained from last
time I create the OSD using ceph-deploy osd prepare?
Crap - it is fighting you, yes - looks like the startup script has triedThere seems to be two hioa1 partitions there, maybe remained from last
time I create the OSD using ceph-deploy osd prepare?
to build an osd for you using ceph-disk (which will make two partitions
by default). So that's toasted the setup that your script did.
Growl - that's made it more complicated for sure.
will avoid creating a journal [partition], and we need to make sure that
the journal behavior is all conditional on the journal symlink being
present. Do you mind opening a bug for this? It could condition itself
off of the osd objectstore option (we'd need to teach ceph-diska bout the
varoius backends), or we could add a secondary option (awkware to
configure), or we could call into ceph-osd with something like 'ceph-osd
-i 0 --does-backend-need-journal' so that a call into the backend
code itself can control things. The latter is probably ideal.
Opened http://tracker.ceph.com/issues/9580 and copying ceph-devel
sage