Discussion:
the state of cephfs in giant
Sage Weil
2014-10-13 18:16:31 UTC
Permalink
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.

What we've working on:

* better mds/cephfs health reports to the monitor
* mds journal dump/repair tool
* many kernel and ceph-fuse/libcephfs client bug fixes
* file size recovery improvements
* client session management fixes (and tests)
* admin socket commands for diagnosis and admin intervention
* many bug fixes

We started using CephFS to back the teuthology (QA) infrastructure in the
lab about three months ago. We fixed a bunch of stuff over the first
month or two (several kernel bugs, a few MDS bugs). We've had no problems
for the last month or so. We're currently running 0.86 (giant release
candidate) with a single MDS and ~70 OSDs. Clients are running a 3.16
kernel plus several fixes that went into 3.17.


With Giant, we are at a point where we would ask that everyone try
things out for any non-production workloads. We are very interested in
feedback around stability, usability, feature gaps, and performance. We
recommend:

* Single active MDS. You can run any number of standby MDS's, but we are
not focusing on multi-mds bugs just yet (and our existing multimds test
suite is already hitting several).
* No snapshots. These are disabled by default and require a scary admin
command to enable them. Although these mostly work, there are
several known issues that we haven't addressed and they complicate
things immensely. Please avoid them for now.
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.

The key missing feature right now is fsck (both check and repair). This is
*the* development focus for Hammer.


Here's a more detailed rundown of the status of various features:

* multi-mds: implemented. limited test coverage. several known issues.
use only for non-production workloads and expect some stability
issues that could lead to data loss.

* snapshots: implemented. limited test coverage. several known issues.
use only for non-production workloads and expect some stability issues
that could lead to data loss.

* hard links: stable. no known issues, but there is somewhat limited
test coverage (we don't test creating huge link farms).

* direct io: implemented and tested for kernel client. no special
support for ceph-fuse (the kernel fuse driver handles this).

* xattrs: implemented, stable, tested. no known issues (for both kernel
and userspace clients).

* ACLs: implemented, tested for kernel client. not implemented for
ceph-fuse.

* file locking (fcntl, flock): supported and tested for kernel client.
limited test coverage. one known minor issue for kernel with fix
pending. implemention in progress for ceph-fuse/libcephfs.

* kernel fscache support: implmented. no test coverage. used in
production by adfin.

* hadoop bindings: implemented, limited test coverage. a few known
issues.

* samba VFS integration: implemented, limited test coverage.

* ganesha NFS integration: implemented, no test coverage.

* kernel NFS reexport: implemented. limited test coverage. no known
issues.


Anybody who has experienced bugs in the past should be excited by:

* new MDS admin socket commands to look at pending operations and client
session states. (Check them out with "ceph daemon mds.a help"!) These
will make diagnosing, debugging, and even fixing issues a lot simpler.

* the cephfs_journal_tool, which is capable of manipulating mds journal
state without doing difficult exports/imports and using hexedit.

Thanks!
sage
Wido den Hollander
2014-10-13 18:20:33 UTC
Permalink
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
* better mds/cephfs health reports to the monitor
* mds journal dump/repair tool
* many kernel and ceph-fuse/libcephfs client bug fixes
* file size recovery improvements
* client session management fixes (and tests)
* admin socket commands for diagnosis and admin intervention
* many bug fixes
We started using CephFS to back the teuthology (QA) infrastructure in the
lab about three months ago. We fixed a bunch of stuff over the first
month or two (several kernel bugs, a few MDS bugs). We've had no problems
for the last month or so. We're currently running 0.86 (giant release
candidate) with a single MDS and ~70 OSDs. Clients are running a 3.16
kernel plus several fixes that went into 3.17.
With Giant, we are at a point where we would ask that everyone try
things out for any non-production workloads. We are very interested in
feedback around stability, usability, feature gaps, and performance. We
A question to clarify this for anybody out there. Do you think it is
safe to run CephFS on a cluster which is doing production RBD/RGW I/O?

Will it be the MDS/CephFS part which breaks or are there potential issue
due to OSD classes which might cause OSDs to crash due to bugs in CephFS?

I know you can't fully rule it out, but it would be useful to have this
clarified.
Post by Sage Weil
* Single active MDS. You can run any number of standby MDS's, but we are
not focusing on multi-mds bugs just yet (and our existing multimds test
suite is already hitting several).
* No snapshots. These are disabled by default and require a scary admin
command to enable them. Although these mostly work, there are
several known issues that we haven't addressed and they complicate
things immensely. Please avoid them for now.
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
The key missing feature right now is fsck (both check and repair). This is
*the* development focus for Hammer.
* multi-mds: implemented. limited test coverage. several known issues.
use only for non-production workloads and expect some stability
issues that could lead to data loss.
* snapshots: implemented. limited test coverage. several known issues.
use only for non-production workloads and expect some stability issues
that could lead to data loss.
* hard links: stable. no known issues, but there is somewhat limited
test coverage (we don't test creating huge link farms).
* direct io: implemented and tested for kernel client. no special
support for ceph-fuse (the kernel fuse driver handles this).
* xattrs: implemented, stable, tested. no known issues (for both kernel
and userspace clients).
* ACLs: implemented, tested for kernel client. not implemented for
ceph-fuse.
* file locking (fcntl, flock): supported and tested for kernel client.
limited test coverage. one known minor issue for kernel with fix
pending. implemention in progress for ceph-fuse/libcephfs.
* kernel fscache support: implmented. no test coverage. used in
production by adfin.
* hadoop bindings: implemented, limited test coverage. a few known
issues.
* samba VFS integration: implemented, limited test coverage.
* ganesha NFS integration: implemented, no test coverage.
* kernel NFS reexport: implemented. limited test coverage. no known
issues.
* new MDS admin socket commands to look at pending operations and client
session states. (Check them out with "ceph daemon mds.a help"!) These
will make diagnosing, debugging, and even fixing issues a lot simpler.
* the cephfs_journal_tool, which is capable of manipulating mds journal
state without doing difficult exports/imports and using hexedit.
Thanks!
sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
More majordomo info at http://vger.kernel.org/majordomo-info.html
--
Wido den Hollander
42on B.V.

Phone: +31 (0)20 700 9902
Skype: contact42on
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Sage Weil
2014-10-13 18:26:50 UTC
Permalink
Post by Wido den Hollander
Post by Sage Weil
With Giant, we are at a point where we would ask that everyone try
things out for any non-production workloads. We are very interested in
feedback around stability, usability, feature gaps, and performance. We
A question to clarify this for anybody out there. Do you think it is
safe to run CephFS on a cluster which is doing production RBD/RGW I/O?
Will it be the MDS/CephFS part which breaks or are there potential issue
due to OSD classes which might cause OSDs to crash due to bugs in CephFS?
I know you can't fully rule it out, but it would be useful to have this
clarified.
I can't think of any issues that this would cause with the OSDs. CephFS
isn't using any rados classes; just core rados functionality that RGW also
uses.

On the monitor side, there is a reasonably probability of triggering a
CephFS related health warning. There is also the potential for code in
the MDSMonitor.cc code to crash the mon, but I don't think we've seen any
problems there any time recently.

So, probably safe.

sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Eric Eastman
2014-10-13 19:03:56 UTC
Permalink
I would be interested in testing the Samba VFS and Ganesha NFS

integration with CephFS. Are there any notes on how to configure these

two interfaces with CephFS?



Eric
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months.
This
Post by Sage Weil
is an update on the current state of things as of Giant.
...
* samba VFS integration: implemented, limited test coverage.
* ganesha NFS integration: implemented, no test coverage.
...
Thanks!
sage
_______________________________________________


--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Sage Weil
2014-10-13 20:56:57 UTC
Permalink
I would be interested in testing the Samba VFS and Ganesha NFS integration
with CephFS. Are there any notes on how to configure these two interfaces
with CephFS?
For samba, based on
https://github.com/ceph/ceph-qa-suite/blob/master/tasks/samba.py#L106
I think you need something like

[myshare]
path = /
writeable = yes
vfs objects = ceph
ceph:config_file = /etc/ceph/ceph.conf

Not sure what the ganesha config looks like. Matt and the other folks at
cohortfs would know more.

sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Amon Ott
2014-10-14 07:31:56 UTC
Permalink
We've been doing a lot of work on CephFS over the past few months. Th=
is
is an update on the current state of things as of Giant.
=2E..
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-=
fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.

=46or kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?

Thanks again,

Amon Ott
--=20
Dr. Amon Ott
m-privacy GmbH Tel: +49 30 24342334
Werner-Vo=DF-Damm 62 Fax: +49 30 99296856
12101 Berlin http://www.m-privacy.de

Amtsgericht Charlottenburg, HRB 84946

Gesch=E4ftsf=FChrer:
Dipl.-Kfm. Holger Maczkowsky,
Roman Maczkowsky

GnuPG-Key-ID: 0x2DD3A649

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" i=
n
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Sage Weil
2014-10-14 13:09:10 UTC
Permalink
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
We haven't been backporting CephFS bug fixes to the stable kernels the
same way we've been doing RBD bugs; it's a bit of a chore. This can be
done retroactively but no promises. Probably 3.14 makes the most sense.
The RHEL7/CentOS7 kernel is also a likely target.

sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Sage Weil
2014-10-14 14:23:40 UTC
Permalink
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
There are important bug fixes missing from 3.14. IIRC, the EC, cache
tiering, and firefly CRUSH changes aren't there yet either (they landed in
3.15), and that is not appropriate for a stable series.

They can be backported, but no commitment yet on that :)

sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Alphe Salas
2014-10-15 00:16:41 UTC
Permalink
Hello sage, last time I used CephFS it had a strange behaviour when if
used in conjunction with a nfs reshare of the cephfs mount point, I
experienced a partial random disapearance of the tree folders.

According to people in the mailing list it was a kernel module bug (not
using ceph-fuse) do you know if any work has been done recently in that
topic?

best regards

Alphe Salas
I.T ingeneer
Post by Sage Weil
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
There are important bug fixes missing from 3.14. IIRC, the EC, cache
tiering, and firefly CRUSH changes aren't there yet either (they landed in
3.15), and that is not appropriate for a stable series.
They can be backported, but no commitment yet on that :)
sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
More majordomo info at http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Sage Weil
2014-10-15 02:06:41 UTC
Permalink
This sounds like any number of readdir bugs that Zheng has fixed over the
last 6 months.

sage
Hello sage, last time I used CephFS it had a strange behaviour when if used in
conjunction with a nfs reshare of the cephfs mount point, I experienced a
partial random disapearance of the tree folders.
According to people in the mailing list it was a kernel module bug (not using
ceph-fuse) do you know if any work has been done recently in that topic?
best regards
Alphe Salas
I.T ingeneer
Post by Sage Weil
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
There are important bug fixes missing from 3.14. IIRC, the EC, cache
tiering, and firefly CRUSH changes aren't there yet either (they landed in
3.15), and that is not appropriate for a stable series.
They can be backported, but no commitment yet on that :)
sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
More majordomo info at http://vger.kernel.org/majordomo-info.html
_______________________________________________
ceph-users mailing list
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Amon Ott
2014-10-15 06:43:42 UTC
Permalink
Post by Sage Weil
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
There are important bug fixes missing from 3.14. IIRC, the EC, cache
tiering, and firefly CRUSH changes aren't there yet either (they landed in
3.15), and that is not appropriate for a stable series.
They can be backported, but no commitment yet on that :)
If the bugfixes are easily identified in one of your Ceph git branches,
I would even try to backport them myself. Still, I would rather see
someone from the Ceph team with deeper knowledge of the code port them.

IMHO, it would be good for Ceph to have stable support in at least the
latest longterm kernel. No need for new features, but bugfixes should be
there.

Amon Ott
--
Dr. Amon Ott
m-privacy GmbH Tel: +49 30 24342334
Werner-Voß-Damm 62 Fax: +49 30 99296856
12101 Berlin http://www.m-privacy.de

Amtsgericht Charlottenburg, HRB 84946

Geschäftsführer:
Dipl.-Kfm. Holger Maczkowsky,
Roman Maczkowsky

GnuPG-Key-ID: 0x2DD3A649
Ric Wheeler
2014-10-15 12:11:10 UTC
Permalink
Post by Amon Ott
Post by Sage Weil
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
There are important bug fixes missing from 3.14. IIRC, the EC, cache
tiering, and firefly CRUSH changes aren't there yet either (they landed in
3.15), and that is not appropriate for a stable series.
They can be backported, but no commitment yet on that :)
If the bugfixes are easily identified in one of your Ceph git branches,
I would even try to backport them myself. Still, I would rather see
someone from the Ceph team with deeper knowledge of the code port them.
IMHO, it would be good for Ceph to have stable support in at least the
latest longterm kernel. No need for new features, but bugfixes should be
there.
Amon Ott
Long term support and aggressive, tedious backports are what you go to distro
vendors for normally - I don't think that it is generally a good practice to
continually backport anything to stable series kernels that is not a
bugfix/security issue (or else, the stable branches rapidly just a stale version
of the upstream tip :)).

Not meant as a commercial for RH, other vendors also do this kind of thing of
course...

Regards,

Ric

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Amon Ott
2014-10-15 13:13:57 UTC
Permalink
Post by Ric Wheeler
Post by Amon Ott
Post by Sage Weil
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
There are important bug fixes missing from 3.14. IIRC, the EC, cache
tiering, and firefly CRUSH changes aren't there yet either (they landed in
3.15), and that is not appropriate for a stable series.
They can be backported, but no commitment yet on that :)
If the bugfixes are easily identified in one of your Ceph git branches,
I would even try to backport them myself. Still, I would rather see
someone from the Ceph team with deeper knowledge of the code port them.
IMHO, it would be good for Ceph to have stable support in at least the
latest longterm kernel. No need for new features, but bugfixes should be
there.
Amon Ott
Long term support and aggressive, tedious backports are what you go to
distro vendors for normally - I don't think that it is generally a good
practice to continually backport anything to stable series kernels that
is not a bugfix/security issue (or else, the stable branches rapidly
just a stale version of the upstream tip :)).
bugfix/security is exactly what I am looking for.

Amon Ott
--
Dr. Amon Ott
m-privacy GmbH Tel: +49 30 24342334
Werner-Voß-Damm 62 Fax: +49 30 99296856
12101 Berlin http://www.m-privacy.de

Amtsgericht Charlottenburg, HRB 84946

Geschäftsführer:
Dipl.-Kfm. Holger Maczkowsky,
Roman Maczkowsky

GnuPG-Key-ID: 0x2DD3A649
Sage Weil
2014-10-15 14:58:19 UTC
Permalink
Post by Amon Ott
Post by Ric Wheeler
Post by Amon Ott
Post by Sage Weil
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
There are important bug fixes missing from 3.14. IIRC, the EC, cache
tiering, and firefly CRUSH changes aren't there yet either (they landed in
3.15), and that is not appropriate for a stable series.
They can be backported, but no commitment yet on that :)
If the bugfixes are easily identified in one of your Ceph git branches,
I would even try to backport them myself. Still, I would rather see
someone from the Ceph team with deeper knowledge of the code port them.
IMHO, it would be good for Ceph to have stable support in at least the
latest longterm kernel. No need for new features, but bugfixes should be
there.
Amon Ott
Long term support and aggressive, tedious backports are what you go to
distro vendors for normally - I don't think that it is generally a good
practice to continually backport anything to stable series kernels that
is not a bugfix/security issue (or else, the stable branches rapidly
just a stale version of the upstream tip :)).
bugfix/security is exactly what I am looking for.
Right; sorry if I was unclear. We make a point of sending bug fixes to
***@vger.kernel.org but haven't been aggressive with cephfs because
the code is less stable. There will be catch-up required to get 3.14 in
good working order.

Definitely hear you that this important, just can't promise when we'll
have the time to do it. There's probably a half day's effort to pick out
the right patches and make sure they build properly, and then some time to
feed it through the test suite.

sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Alphe Salas
2014-10-15 16:47:40 UTC
Permalink
For the humble ceph user I am it is really hard to follow what version
of what product will get the changes I requiere.

Let me explain myself. I use ceph in my company is specialised in disk
recovery, my company needs a flexible, easy to maintain, trustable way
to store the data from the disks of our clients.

We try the usual way jbod boxes connected to a single server with a SAS
raid card and ZFS mirror to handle replicas and merging disks into a big
disk. result is really slow. (used to use zfs and solaris 11 on x86
servers... with openZfs and ubuntu 14.04 the perf are way better but not
any were comparable with ceph (on a giga ethernet lan you can get data
transfer betwin client and ceph cluster around 80MB/s...while client to
openzfs/ubuntu is around 25MB/S)

Along my path with ceph I first used cephfs, worked fine! until I
noticed that part of the folder tree suddently randomly disapeared
forcing a constant periodical remount of the partitions.

Then I choose to forget about cephfs and use rbd images, worked fine!
Until I noticed that rbd replicas where never freed or overwriten and
that for a replicas set to 2 (data and 1 replica) and an image of 13 TB
after some time of write erase cycles on the same rbd image I get an
overall data use of 34 TB over the 36TB available on my cluster I
noticed that there was a real problem with "space management". The data
part of the rbd image was properly managed using overwrites on old
deleted data at OS level, so the only logical explaination of the
overall data use growth was that the replicas where never freed.

All along that time I was pending of the bugs/ features and advances of
ceph.
But those isues are not really ceph related they are kernel modules for
using "ceph clients" so the release of feature add and bug fix are in
part to be given in the ceph-common package (for the server related
machanics) and the other part is then to be provided at the kernel level.

For comodity I use Ubuntu which is not really top notch using the very
lastest brew of the kernel and all the bug fixed modules.

So when I see this great news about giant and the fact that alot of work
has been done in solving most of the problems we all faced with
ceph then I notice that it will be around a year or so for those fix to
be production available in ubuntu. There is some inertia there that
doesn t match with the pace of the work on ceph.

Then people can arg with me "why you use ubuntu?"
and the answers are simple I have a cluster of 10 machines and 1 proxy
if I need to compile from source lastest brew of ceph and lastest brew
of kernel then my maintainance time will be way bigger. And I am more
intended to get something that isn t properly done and have a machine
that doesn t reboot.
I know what I am talking about I used during several month ceph in
archlinux compiling kernel and ceph from source until the gcc installed
on my test server was too new and a compile option had been removed then
ceph wasn t compiling. That way to proceed was descarted because not
stable enough to bring production level quality.

So as far as I understand things I will have cephfs enhanced and rbd
discard ability available at same time using the couple ceph giant and
linux kernel 3.18 and up ?

regards and thank you again for your hardwork, I wish I could do more to
help.


---
Alphe Salas
I.T ingeneer
Post by Sage Weil
Post by Amon Ott
Post by Ric Wheeler
Post by Amon Ott
Post by Sage Weil
Post by Amon Ott
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
...
Post by Sage Weil
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
Thanks for all the work and specially for concentrating on CephFS! We
have been watching and testing for years by now and really hope to
change our Clusters to CephFS soon.
For kernel maintenance reasons, we only want to run longterm stable
kernels. And for performance reasons and because of severe known
problems we want to avoid Fuse. How good are our chances of a stable
system with the kernel client in the latest longterm kernel 3.14? Will
there be further bugfixes or feature backports?
There are important bug fixes missing from 3.14. IIRC, the EC, cache
tiering, and firefly CRUSH changes aren't there yet either (they landed in
3.15), and that is not appropriate for a stable series.
They can be backported, but no commitment yet on that :)
If the bugfixes are easily identified in one of your Ceph git branches,
I would even try to backport them myself. Still, I would rather see
someone from the Ceph team with deeper knowledge of the code port them.
IMHO, it would be good for Ceph to have stable support in at least the
latest longterm kernel. No need for new features, but bugfixes should be
there.
Amon Ott
Long term support and aggressive, tedious backports are what you go to
distro vendors for normally - I don't think that it is generally a good
practice to continually backport anything to stable series kernels that
is not a bugfix/security issue (or else, the stable branches rapidly
just a stale version of the upstream tip :)).
bugfix/security is exactly what I am looking for.
Right; sorry if I was unclear. We make a point of sending bug fixes to
the code is less stable. There will be catch-up required to get 3.14 in
good working order.
Definitely hear you that this important, just can't promise when we'll
have the time to do it. There's probably a half day's effort to pick out
the right patches and make sure they build properly, and then some time to
feed it through the test suite.
sage
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
More majordomo info at http://vger.kernel.org/majordomo-info.html
Thomas Lemarchand
2014-10-14 09:57:11 UTC
Permalink
Thanks for theses informations.

I plan to use CephFS on Giant, with production workload, knowing the
risks and having a hot backup near. I hope to be able to provide useful
feedback.

My cluster is made of 7 servers (3mon, 3osd (27 osd inside), 1mds). I
use ceph-fuse on clients.

You wrote about hardlinks, but what about symlinks ? I use some (on
cephFS firefly) without any problem for now.

Do you suggest something for backup of CephFS ? For now I use a simple
rsync, it works quite well.

Thanks !
--
Thomas Lemarchand
Cloud Solutions SAS - Responsable des systèmes d'information
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
* better mds/cephfs health reports to the monitor
* mds journal dump/repair tool
* many kernel and ceph-fuse/libcephfs client bug fixes
* file size recovery improvements
* client session management fixes (and tests)
* admin socket commands for diagnosis and admin intervention
* many bug fixes
We started using CephFS to back the teuthology (QA) infrastructure in the
lab about three months ago. We fixed a bunch of stuff over the first
month or two (several kernel bugs, a few MDS bugs). We've had no problems
for the last month or so. We're currently running 0.86 (giant release
candidate) with a single MDS and ~70 OSDs. Clients are running a 3.16
kernel plus several fixes that went into 3.17.
With Giant, we are at a point where we would ask that everyone try
things out for any non-production workloads. We are very interested in
feedback around stability, usability, feature gaps, and performance. We
* Single active MDS. You can run any number of standby MDS's, but we are
not focusing on multi-mds bugs just yet (and our existing multimds test
suite is already hitting several).
* No snapshots. These are disabled by default and require a scary admin
command to enable them. Although these mostly work, there are
several known issues that we haven't addressed and they complicate
things immensely. Please avoid them for now.
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
The key missing feature right now is fsck (both check and repair). This is
*the* development focus for Hammer.
* multi-mds: implemented. limited test coverage. several known issues.
use only for non-production workloads and expect some stability
issues that could lead to data loss.
* snapshots: implemented. limited test coverage. several known issues.
use only for non-production workloads and expect some stability issues
that could lead to data loss.
* hard links: stable. no known issues, but there is somewhat limited
test coverage (we don't test creating huge link farms).
* direct io: implemented and tested for kernel client. no special
support for ceph-fuse (the kernel fuse driver handles this).
* xattrs: implemented, stable, tested. no known issues (for both kernel
and userspace clients).
* ACLs: implemented, tested for kernel client. not implemented for
ceph-fuse.
* file locking (fcntl, flock): supported and tested for kernel client.
limited test coverage. one known minor issue for kernel with fix
pending. implemention in progress for ceph-fuse/libcephfs.
* kernel fscache support: implmented. no test coverage. used in
production by adfin.
* hadoop bindings: implemented, limited test coverage. a few known
issues.
* samba VFS integration: implemented, limited test coverage.
* ganesha NFS integration: implemented, no test coverage.
* kernel NFS reexport: implemented. limited test coverage. no known
issues.
* new MDS admin socket commands to look at pending operations and client
session states. (Check them out with "ceph daemon mds.a help"!) These
will make diagnosing, debugging, and even fixing issues a lot simpler.
* the cephfs_journal_tool, which is capable of manipulating mds journal
state without doing difficult exports/imports and using hexedit.
Thanks!
sage
_______________________________________________
ceph-users mailing list
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
--
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.
Sage Weil
2014-10-14 13:11:42 UTC
Permalink
Post by Thomas Lemarchand
Thanks for theses informations.
I plan to use CephFS on Giant, with production workload, knowing the
risks and having a hot backup near. I hope to be able to provide useful
feedback.
My cluster is made of 7 servers (3mon, 3osd (27 osd inside), 1mds). I
use ceph-fuse on clients.
Cool! Please be careful, and have a plan B. :)
Post by Thomas Lemarchand
You wrote about hardlinks, but what about symlinks ? I use some (on
cephFS firefly) without any problem for now.
Symlinks are simple and cheap; no issues there.
Post by Thomas Lemarchand
Do you suggest something for backup of CephFS ? For now I use a simple
rsync, it works quite well.
rsync is fine. There is some opportunity to do clever things with the
recursive ctime metadata, but nobody has wired it up to any tools yet.

sage
Post by Thomas Lemarchand
Thanks !
--
Thomas Lemarchand
Cloud Solutions SAS - Responsable des syst?mes d'information
Post by Sage Weil
We've been doing a lot of work on CephFS over the past few months. This
is an update on the current state of things as of Giant.
* better mds/cephfs health reports to the monitor
* mds journal dump/repair tool
* many kernel and ceph-fuse/libcephfs client bug fixes
* file size recovery improvements
* client session management fixes (and tests)
* admin socket commands for diagnosis and admin intervention
* many bug fixes
We started using CephFS to back the teuthology (QA) infrastructure in the
lab about three months ago. We fixed a bunch of stuff over the first
month or two (several kernel bugs, a few MDS bugs). We've had no problems
for the last month or so. We're currently running 0.86 (giant release
candidate) with a single MDS and ~70 OSDs. Clients are running a 3.16
kernel plus several fixes that went into 3.17.
With Giant, we are at a point where we would ask that everyone try
things out for any non-production workloads. We are very interested in
feedback around stability, usability, feature gaps, and performance. We
* Single active MDS. You can run any number of standby MDS's, but we are
not focusing on multi-mds bugs just yet (and our existing multimds test
suite is already hitting several).
* No snapshots. These are disabled by default and require a scary admin
command to enable them. Although these mostly work, there are
several known issues that we haven't addressed and they complicate
things immensely. Please avoid them for now.
* Either the kernel client (kernel 3.17 or later) or userspace (ceph-fuse
or libcephfs) clients are in good working order.
The key missing feature right now is fsck (both check and repair). This is
*the* development focus for Hammer.
* multi-mds: implemented. limited test coverage. several known issues.
use only for non-production workloads and expect some stability
issues that could lead to data loss.
* snapshots: implemented. limited test coverage. several known issues.
use only for non-production workloads and expect some stability issues
that could lead to data loss.
* hard links: stable. no known issues, but there is somewhat limited
test coverage (we don't test creating huge link farms).
* direct io: implemented and tested for kernel client. no special
support for ceph-fuse (the kernel fuse driver handles this).
* xattrs: implemented, stable, tested. no known issues (for both kernel
and userspace clients).
* ACLs: implemented, tested for kernel client. not implemented for
ceph-fuse.
* file locking (fcntl, flock): supported and tested for kernel client.
limited test coverage. one known minor issue for kernel with fix
pending. implemention in progress for ceph-fuse/libcephfs.
* kernel fscache support: implmented. no test coverage. used in
production by adfin.
* hadoop bindings: implemented, limited test coverage. a few known
issues.
* samba VFS integration: implemented, limited test coverage.
* ganesha NFS integration: implemented, no test coverage.
* kernel NFS reexport: implemented. limited test coverage. no known
issues.
* new MDS admin socket commands to look at pending operations and client
session states. (Check them out with "ceph daemon mds.a help"!) These
will make diagnosing, debugging, and even fixing issues a lot simpler.
* the cephfs_journal_tool, which is capable of manipulating mds journal
state without doing difficult exports/imports and using hexedit.
Thanks!
sage
_______________________________________________
ceph-users mailing list
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
--
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to ***@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Loading...