From: "Jim Schutt" <jaschut@sandia.gov>
To: "ceph-devel@vger.kernel.org" <ceph-devel@vger.kernel.org>
Subject: scaling issues
Date: Thu, 8 Mar 2012 16:31:23 -0700 [thread overview]
Message-ID: <4F59414B.3000403@sandia.gov> (raw)
Hi,
I've been trying to scale up a Ceph filesystem to as big
as I have hardware for - up to 288 OSDs right now.
(I'm using commit ed0f605365e - tip of master branch from
a few days ago.)
My problem is that I cannot get a 288 OSD filesystem to go active
(that's with 1 mon and 1 MDS). Pretty quickly I start seeing
"mds e4 e4: 1/1/1 up {0=cs33=up:creating(laggy or crashed)}".
Note that as this is happening all the OSDs and the MDS are
essentially idle; only the mon is busy.
While tailing the mon log I noticed there was a periodic pause;
after adding a little more debug printing, I learned that the
pause was due to encoding pg_stat_t before writing the pg_map to disk.
Here's the result of a scaling study I did on startup time for
a freshly created filesystem. I normally run 24 OSDs/server on
these machines with no trouble, for small numbers of OSDs.
seconds from seconds from seconds to
OSD PG store() mount store() mount encode
to to all PGs pg_stat_t Notes
up:active active+clean*
48 9504 58 63 0.30
72 14256 70 89 0.65
96 19008 93 117 1.1
120 23760 132 138 1.7
144 28512 92 165 2.3
168 33264 215 218 3.2 periods of "up:creating(laggy or crashed)"
192 38016 392 344 4.0 periods of "up:creating(laggy or crashed)"
240 47520 1189 644 6.3 periods of "up:creating(laggy or crashed)"
288 57024 >14400 >14400 9.0 never went active; >200 OSDs out, reporting "wrongly marked me down"
* active+clean includes active+clean+scrubbing, i.e., no peering or creating
** all runs up to 288 used mon osd down out interval = 30; 288 used that for
first hour, then switched to 300
It might be that the filesystem never went to active at 288 OSDs due
to some lurking bugs, but even so, the results for time to encode
pg_stat_t is worrisome; gnuplot fit it for me to
2.18341 * exp(OSDs/171.373) - 2.67065
----
After 79 iterations the fit converged.
final sum of squares of residuals : 0.0363573
rel. change during last iteration : -4.77639e-06
degrees of freedom (FIT_NDF) : 6
rms of residuals (FIT_STDFIT) = sqrt(WSSR/ndf) : 0.0778431
variance of residuals (reduced chisquare) = WSSR/ndf : 0.00605955
Final set of parameters Asymptotic Standard Error
======================= ==========================
a = 2.18341 +/- 0.2276 (10.42%)
b = 171.373 +/- 8.344 (4.869%)
c = -2.67065 +/- 0.3049 (11.42%)
----
I haven't dug deeply into what all goes into a pg_stat_t; how is that
expected to scale? I tried to fit it to some other functions, but
they didn't look as good to me (not very scientific).
If that fit is correct, and I had the hardware to double my cluster
size to 576 OSDs, the time to encode pg_stat_t for such a cluster
would be ~60 seconds. That seems unlikely to work well, and what
I'd really like to get to is thousands of OSDs.
Let me know if there is anything I can do to help with this. I've still
got the mon logs for the above runs, with debug ms = 1 and debug mon = 10;
-- Jim
P.S. Here's how I instrumented to get above results:
diff --git a/src/mon/PGMap.cc b/src/mon/PGMap.cc
index d961ac1..58198d7 100644
--- a/src/mon/PGMap.cc
+++ b/src/mon/PGMap.cc
@@ -5,6 +5,7 @@
#define DOUT_SUBSYS mon
#include "common/debug.h"
+#include "common/Clock.h"
#include "common/Formatter.h"
@@ -311,8 +312,17 @@ void PGMap::encode(bufferlist &bl) const
__u8 v = 3;
::encode(v, bl);
::encode(version, bl);
+
+ utime_t start = ceph_clock_now(g_ceph_context);
::encode(pg_stat, bl);
+ utime_t end = ceph_clock_now(g_ceph_context);
+ dout(10) << "PGMap::encode pg_stat took " << end - start << dendl;
+
+ start = end;
::encode(osd_stat, bl);
+ end = ceph_clock_now(g_ceph_context);
+ dout(10) << "PGMap::encode osd_stat took " << end - start << dendl;
+
::encode(last_osdmap_epoch, bl);
::encode(last_pg_scan, bl);
::encode(full_ratio, bl);
--
1.7.8.2
next reply other threads:[~2012-03-08 23:31 UTC|newest]
Thread overview: 8+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-03-08 23:31 Jim Schutt [this message]
2012-03-09 0:26 ` scaling issues Sage Weil
2012-03-09 19:39 ` Jim Schutt
2012-03-09 23:21 ` Jim Schutt
2012-04-10 16:22 ` Jim Schutt
2012-04-10 16:39 ` Sage Weil
2012-04-10 19:01 ` [EXTERNAL] " Jim Schutt
2012-04-10 22:38 ` Sage Weil
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4F59414B.3000403@sandia.gov \
--to=jaschut@sandia.gov \
--cc=ceph-devel@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.