All of lore.kernel.org
 help / color / mirror / Atom feed
* still recovery issues with cuttlefish
@ 2013-08-01  8:22 Stefan Priebe - Profihost AG
  2013-08-01 14:50 ` Andrey Korolyov
  2013-08-01 18:34 ` Samuel Just
  0 siblings, 2 replies; 40+ messages in thread
From: Stefan Priebe - Profihost AG @ 2013-08-01  8:22 UTC (permalink / raw)
  To: ceph-devel

Hi,

i still have recovery issues with cuttlefish. After the OSD comes back
it seem to hang for around 2-4 minutes and then recovery seems to start
(pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
get a lot of slow request messages an hanging VMs.

What i noticed today is that if i leave the OSD off as long as ceph
starts to backfill - the recovery and "re" backfilling wents absolutely
smooth without any issues and no slow request messages at all.

Does anybody have an idea why?

Greets,
Stefan

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01  8:22 still recovery issues with cuttlefish Stefan Priebe - Profihost AG
@ 2013-08-01 14:50 ` Andrey Korolyov
  2013-08-01 18:38   ` Samuel Just
  2013-08-01 18:34 ` Samuel Just
  1 sibling, 1 reply; 40+ messages in thread
From: Andrey Korolyov @ 2013-08-01 14:50 UTC (permalink / raw)
  To: Stefan Priebe - Profihost AG; +Cc: ceph-devel

Second this. Also for long-lasting snapshot problem and related
performance issues I may say that cuttlefish improved things greatly,
but creation/deletion of large snapshot (hundreds of gigabytes of
commited data) still can bring down cluster for a minutes, despite
usage of every possible optimization.

On Thu, Aug 1, 2013 at 12:22 PM, Stefan Priebe - Profihost AG
<s.priebe@profihost.ag> wrote:
> Hi,
>
> i still have recovery issues with cuttlefish. After the OSD comes back
> it seem to hang for around 2-4 minutes and then recovery seems to start
> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
> get a lot of slow request messages an hanging VMs.
>
> What i noticed today is that if i leave the OSD off as long as ceph
> starts to backfill - the recovery and "re" backfilling wents absolutely
> smooth without any issues and no slow request messages at all.
>
> Does anybody have an idea why?
>
> Greets,
> Stefan
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01  8:22 still recovery issues with cuttlefish Stefan Priebe - Profihost AG
  2013-08-01 14:50 ` Andrey Korolyov
@ 2013-08-01 18:34 ` Samuel Just
  2013-08-01 18:34   ` Stefan Priebe
  2013-08-01 18:54   ` Mike Dawson
  1 sibling, 2 replies; 40+ messages in thread
From: Samuel Just @ 2013-08-01 18:34 UTC (permalink / raw)
  To: Stefan Priebe - Profihost AG; +Cc: ceph-devel

Can you reproduce and attach the ceph.log from before you stop the osd
until after you have started the osd and it has recovered?
-Sam

On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
<s.priebe@profihost.ag> wrote:
> Hi,
>
> i still have recovery issues with cuttlefish. After the OSD comes back
> it seem to hang for around 2-4 minutes and then recovery seems to start
> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
> get a lot of slow request messages an hanging VMs.
>
> What i noticed today is that if i leave the OSD off as long as ceph
> starts to backfill - the recovery and "re" backfilling wents absolutely
> smooth without any issues and no slow request messages at all.
>
> Does anybody have an idea why?
>
> Greets,
> Stefan
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 18:34 ` Samuel Just
@ 2013-08-01 18:34   ` Stefan Priebe
  2013-08-01 18:36     ` Samuel Just
  2013-08-01 18:54   ` Mike Dawson
  1 sibling, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-01 18:34 UTC (permalink / raw)
  To: Samuel Just; +Cc: ceph-devel

m 01.08.2013 20:34, schrieb Samuel Just:
> Can you reproduce and attach the ceph.log from before you stop the osd
> until after you have started the osd and it has recovered?
> -Sam

Sure which log levels?

> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
> <s.priebe@profihost.ag> wrote:
>> Hi,
>>
>> i still have recovery issues with cuttlefish. After the OSD comes back
>> it seem to hang for around 2-4 minutes and then recovery seems to start
>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>> get a lot of slow request messages an hanging VMs.
>>
>> What i noticed today is that if i leave the OSD off as long as ceph
>> starts to backfill - the recovery and "re" backfilling wents absolutely
>> smooth without any issues and no slow request messages at all.
>>
>> Does anybody have an idea why?
>>
>> Greets,
>> Stefan
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 18:34   ` Stefan Priebe
@ 2013-08-01 18:36     ` Samuel Just
  2013-08-01 18:36       ` Samuel Just
  2013-08-01 18:46       ` Stefan Priebe
  0 siblings, 2 replies; 40+ messages in thread
From: Samuel Just @ 2013-08-01 18:36 UTC (permalink / raw)
  To: Stefan Priebe; +Cc: ceph-devel

For now, just the main ceph.log.
-Sam

On Thu, Aug 1, 2013 at 11:34 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> m 01.08.2013 20:34, schrieb Samuel Just:
>
>> Can you reproduce and attach the ceph.log from before you stop the osd
>> until after you have started the osd and it has recovered?
>> -Sam
>
>
> Sure which log levels?
>
>
>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>> <s.priebe@profihost.ag> wrote:
>>>
>>> Hi,
>>>
>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>> it seem to hang for around 2-4 minutes and then recovery seems to start
>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>> get a lot of slow request messages an hanging VMs.
>>>
>>> What i noticed today is that if i leave the OSD off as long as ceph
>>> starts to backfill - the recovery and "re" backfilling wents absolutely
>>> smooth without any issues and no slow request messages at all.
>>>
>>> Does anybody have an idea why?
>>>
>>> Greets,
>>> Stefan
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 18:36     ` Samuel Just
@ 2013-08-01 18:36       ` Samuel Just
  2013-08-01 18:46       ` Stefan Priebe
  1 sibling, 0 replies; 40+ messages in thread
From: Samuel Just @ 2013-08-01 18:36 UTC (permalink / raw)
  To: Stefan Priebe; +Cc: ceph-devel

It doesn't have log levels, should be in /var/log/ceph/ceph.log.
-Sam

On Thu, Aug 1, 2013 at 11:36 AM, Samuel Just <sam.just@inktank.com> wrote:
> For now, just the main ceph.log.
> -Sam
>
> On Thu, Aug 1, 2013 at 11:34 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> m 01.08.2013 20:34, schrieb Samuel Just:
>>
>>> Can you reproduce and attach the ceph.log from before you stop the osd
>>> until after you have started the osd and it has recovered?
>>> -Sam
>>
>>
>> Sure which log levels?
>>
>>
>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>> <s.priebe@profihost.ag> wrote:
>>>>
>>>> Hi,
>>>>
>>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>>> it seem to hang for around 2-4 minutes and then recovery seems to start
>>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>>> get a lot of slow request messages an hanging VMs.
>>>>
>>>> What i noticed today is that if i leave the OSD off as long as ceph
>>>> starts to backfill - the recovery and "re" backfilling wents absolutely
>>>> smooth without any issues and no slow request messages at all.
>>>>
>>>> Does anybody have an idea why?
>>>>
>>>> Greets,
>>>> Stefan
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 14:50 ` Andrey Korolyov
@ 2013-08-01 18:38   ` Samuel Just
  2013-08-02 17:56     ` Andrey Korolyov
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-01 18:38 UTC (permalink / raw)
  To: Andrey Korolyov; +Cc: Stefan Priebe - Profihost AG, ceph-devel

Is there a bug open for this?  I suspect we don't sufficiently
throttle the snapshot removal work.
-Sam

On Thu, Aug 1, 2013 at 7:50 AM, Andrey Korolyov <andrey@xdel.ru> wrote:
> Second this. Also for long-lasting snapshot problem and related
> performance issues I may say that cuttlefish improved things greatly,
> but creation/deletion of large snapshot (hundreds of gigabytes of
> commited data) still can bring down cluster for a minutes, despite
> usage of every possible optimization.
>
> On Thu, Aug 1, 2013 at 12:22 PM, Stefan Priebe - Profihost AG
> <s.priebe@profihost.ag> wrote:
>> Hi,
>>
>> i still have recovery issues with cuttlefish. After the OSD comes back
>> it seem to hang for around 2-4 minutes and then recovery seems to start
>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>> get a lot of slow request messages an hanging VMs.
>>
>> What i noticed today is that if i leave the OSD off as long as ceph
>> starts to backfill - the recovery and "re" backfilling wents absolutely
>> smooth without any issues and no slow request messages at all.
>>
>> Does anybody have an idea why?
>>
>> Greets,
>> Stefan
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 18:36     ` Samuel Just
  2013-08-01 18:36       ` Samuel Just
@ 2013-08-01 18:46       ` Stefan Priebe
  1 sibling, 0 replies; 40+ messages in thread
From: Stefan Priebe @ 2013-08-01 18:46 UTC (permalink / raw)
  To: Samuel Just; +Cc: ceph-devel

[-- Attachment #1: Type: text/plain, Size: 1543 bytes --]

here it is

Am 01.08.2013 20:36, schrieb Samuel Just:
> For now, just the main ceph.log.
> -Sam
>
> On Thu, Aug 1, 2013 at 11:34 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> m 01.08.2013 20:34, schrieb Samuel Just:
>>
>>> Can you reproduce and attach the ceph.log from before you stop the osd
>>> until after you have started the osd and it has recovered?
>>> -Sam
>>
>>
>> Sure which log levels?
>>
>>
>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>> <s.priebe@profihost.ag> wrote:
>>>>
>>>> Hi,
>>>>
>>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>>> it seem to hang for around 2-4 minutes and then recovery seems to start
>>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>>> get a lot of slow request messages an hanging VMs.
>>>>
>>>> What i noticed today is that if i leave the OSD off as long as ceph
>>>> starts to backfill - the recovery and "re" backfilling wents absolutely
>>>> smooth without any issues and no slow request messages at all.
>>>>
>>>> Does anybody have an idea why?
>>>>
>>>> Greets,
>>>> Stefan
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>

[-- Attachment #2: ceph.log.gz --]
[-- Type: application/gzip, Size: 40224 bytes --]

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 18:34 ` Samuel Just
  2013-08-01 18:34   ` Stefan Priebe
@ 2013-08-01 18:54   ` Mike Dawson
  2013-08-01 19:07     ` Stefan Priebe
  1 sibling, 1 reply; 40+ messages in thread
From: Mike Dawson @ 2013-08-01 18:54 UTC (permalink / raw)
  To: Samuel Just; +Cc: Stefan Priebe - Profihost AG, ceph-devel

I am also seeing recovery issues with 0.61.7. Here's the process:

- ceph osd set noout

- Reboot one of the nodes hosting OSDs
     - VMs mounted from RBD volumes work properly

- I see the OSD's boot messages as they re-join the cluster

- Start seeing active+recovery_wait, peering, and active+recovering
     - VMs mounted from RBD volumes become unresponsive.

- Recovery completes
     - VMs mounted from RBD volumes regain responsiveness

- ceph osd unset noout

Would joshd's async patch for qemu help here, or is there something else 
going on?

Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY

Thanks,

Mike Dawson
Co-Founder & Director of Cloud Architecture
Cloudapt LLC
6330 East 75th Street, Suite 170
Indianapolis, IN 46250

On 8/1/2013 2:34 PM, Samuel Just wrote:
> Can you reproduce and attach the ceph.log from before you stop the osd
> until after you have started the osd and it has recovered?
> -Sam
>
> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
> <s.priebe@profihost.ag> wrote:
>> Hi,
>>
>> i still have recovery issues with cuttlefish. After the OSD comes back
>> it seem to hang for around 2-4 minutes and then recovery seems to start
>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>> get a lot of slow request messages an hanging VMs.
>>
>> What i noticed today is that if i leave the OSD off as long as ceph
>> starts to backfill - the recovery and "re" backfilling wents absolutely
>> smooth without any issues and no slow request messages at all.
>>
>> Does anybody have an idea why?
>>
>> Greets,
>> Stefan
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 18:54   ` Mike Dawson
@ 2013-08-01 19:07     ` Stefan Priebe
  2013-08-01 21:23       ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-01 19:07 UTC (permalink / raw)
  To: Mike Dawson; +Cc: Samuel Just, ceph-devel

Mike we already have the async patch running. Yes it helps but only 
helps it does not solve. It just hides the issue ...
Am 01.08.2013 20:54, schrieb Mike Dawson:
> I am also seeing recovery issues with 0.61.7. Here's the process:
>
> - ceph osd set noout
>
> - Reboot one of the nodes hosting OSDs
>      - VMs mounted from RBD volumes work properly
>
> - I see the OSD's boot messages as they re-join the cluster
>
> - Start seeing active+recovery_wait, peering, and active+recovering
>      - VMs mounted from RBD volumes become unresponsive.
>
> - Recovery completes
>      - VMs mounted from RBD volumes regain responsiveness
>
> - ceph osd unset noout
>
> Would joshd's async patch for qemu help here, or is there something else
> going on?
>
> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>
> Thanks,
>
> Mike Dawson
> Co-Founder & Director of Cloud Architecture
> Cloudapt LLC
> 6330 East 75th Street, Suite 170
> Indianapolis, IN 46250
>
> On 8/1/2013 2:34 PM, Samuel Just wrote:
>> Can you reproduce and attach the ceph.log from before you stop the osd
>> until after you have started the osd and it has recovered?
>> -Sam
>>
>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>> <s.priebe@profihost.ag> wrote:
>>> Hi,
>>>
>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>> it seem to hang for around 2-4 minutes and then recovery seems to start
>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>> get a lot of slow request messages an hanging VMs.
>>>
>>> What i noticed today is that if i leave the OSD off as long as ceph
>>> starts to backfill - the recovery and "re" backfilling wents absolutely
>>> smooth without any issues and no slow request messages at all.
>>>
>>> Does anybody have an idea why?
>>>
>>> Greets,
>>> Stefan
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 19:07     ` Stefan Priebe
@ 2013-08-01 21:23       ` Samuel Just
  2013-08-02  7:44         ` Stefan Priebe
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-01 21:23 UTC (permalink / raw)
  To: Stefan Priebe; +Cc: Mike Dawson, ceph-devel

Can you dump your osd settings?
sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
-Sam

On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> Mike we already have the async patch running. Yes it helps but only helps it
> does not solve. It just hides the issue ...
> Am 01.08.2013 20:54, schrieb Mike Dawson:
>
>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>
>> - ceph osd set noout
>>
>> - Reboot one of the nodes hosting OSDs
>>      - VMs mounted from RBD volumes work properly
>>
>> - I see the OSD's boot messages as they re-join the cluster
>>
>> - Start seeing active+recovery_wait, peering, and active+recovering
>>      - VMs mounted from RBD volumes become unresponsive.
>>
>> - Recovery completes
>>      - VMs mounted from RBD volumes regain responsiveness
>>
>> - ceph osd unset noout
>>
>> Would joshd's async patch for qemu help here, or is there something else
>> going on?
>>
>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>
>> Thanks,
>>
>> Mike Dawson
>> Co-Founder & Director of Cloud Architecture
>> Cloudapt LLC
>> 6330 East 75th Street, Suite 170
>> Indianapolis, IN 46250
>>
>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>
>>> Can you reproduce and attach the ceph.log from before you stop the osd
>>> until after you have started the osd and it has recovered?
>>> -Sam
>>>
>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>> <s.priebe@profihost.ag> wrote:
>>>>
>>>> Hi,
>>>>
>>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>>> it seem to hang for around 2-4 minutes and then recovery seems to start
>>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>>> get a lot of slow request messages an hanging VMs.
>>>>
>>>> What i noticed today is that if i leave the OSD off as long as ceph
>>>> starts to backfill - the recovery and "re" backfilling wents absolutely
>>>> smooth without any issues and no slow request messages at all.
>>>>
>>>> Does anybody have an idea why?
>>>>
>>>> Greets,
>>>> Stefan
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 21:23       ` Samuel Just
@ 2013-08-02  7:44         ` Stefan Priebe
  2013-08-02 17:35           ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-02  7:44 UTC (permalink / raw)
  To: Samuel Just; +Cc: Mike Dawson, ceph-devel

Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd settings?
 > sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show

Sure.



{ "name": "osd.0",
   "cluster": "ceph",
   "none": "0\/5",
   "lockdep": "0\/0",
   "context": "0\/0",
   "crush": "0\/0",
   "mds": "0\/0",
   "mds_balancer": "0\/0",
   "mds_locker": "0\/0",
   "mds_log": "0\/0",
   "mds_log_expire": "0\/0",
   "mds_migrator": "0\/0",
   "buffer": "0\/0",
   "timer": "0\/0",
   "filer": "0\/0",
   "striper": "0\/1",
   "objecter": "0\/0",
   "rados": "0\/0",
   "rbd": "0\/0",
   "journaler": "0\/0",
   "objectcacher": "0\/0",
   "client": "0\/0",
   "osd": "0\/0",
   "optracker": "0\/0",
   "objclass": "0\/0",
   "filestore": "0\/0",
   "journal": "0\/0",
   "ms": "0\/0",
   "mon": "0\/0",
   "monc": "0\/0",
   "paxos": "0\/0",
   "tp": "0\/0",
   "auth": "0\/0",
   "crypto": "1\/5",
   "finisher": "0\/0",
   "heartbeatmap": "0\/0",
   "perfcounter": "0\/0",
   "rgw": "0\/0",
   "hadoop": "0\/0",
   "javaclient": "1\/5",
   "asok": "0\/0",
   "throttle": "0\/0",
   "host": "cloud1-1268",
   "fsid": "00000000-0000-0000-0000-000000000000",
   "public_addr": "10.255.0.90:0\/0",
   "cluster_addr": "10.255.0.90:0\/0",
   "public_network": "10.255.0.1\/24",
   "cluster_network": "10.255.0.1\/24",
   "num_client": "1",
   "monmap": "",
   "mon_host": "",
   "lockdep": "false",
   "run_dir": "\/var\/run\/ceph",
   "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
   "daemonize": "true",
   "pid_file": "\/var\/run\/ceph\/osd.0.pid",
   "chdir": "\/",
   "max_open_files": "0",
   "fatal_signal_handlers": "true",
   "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
   "log_max_new": "1000",
   "log_max_recent": "10000",
   "log_to_stderr": "false",
   "err_to_stderr": "true",
   "log_to_syslog": "false",
   "err_to_syslog": "false",
   "log_flush_on_exit": "true",
   "log_stop_at_utilization": "0.97",
   "clog_to_monitors": "true",
   "clog_to_syslog": "false",
   "clog_to_syslog_level": "info",
   "clog_to_syslog_facility": "daemon",
   "mon_cluster_log_to_syslog": "false",
   "mon_cluster_log_to_syslog_level": "info",
   "mon_cluster_log_to_syslog_facility": "daemon",
   "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
   "key": "",
   "keyfile": "",
   "keyring": "\/etc\/ceph\/osd.0.keyring",
   "heartbeat_interval": "5",
   "heartbeat_file": "",
   "heartbeat_inject_failure": "0",
   "perf": "true",
   "ms_tcp_nodelay": "true",
   "ms_tcp_rcvbuf": "0",
   "ms_initial_backoff": "0.2",
   "ms_max_backoff": "15",
   "ms_nocrc": "false",
   "ms_die_on_bad_msg": "false",
   "ms_die_on_unhandled_msg": "false",
   "ms_dispatch_throttle_bytes": "104857600",
   "ms_bind_ipv6": "false",
   "ms_bind_port_min": "6800",
   "ms_bind_port_max": "7100",
   "ms_rwthread_stack_bytes": "1048576",
   "ms_tcp_read_timeout": "900",
   "ms_pq_max_tokens_per_priority": "4194304",
   "ms_pq_min_cost": "65536",
   "ms_inject_socket_failures": "0",
   "ms_inject_delay_type": "",
   "ms_inject_delay_max": "1",
   "ms_inject_delay_probability": "0",
   "ms_inject_internal_delays": "0",
   "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
   "mon_initial_members": "",
   "mon_sync_fs_threshold": "5",
   "mon_compact_on_start": "false",
   "mon_compact_on_bootstrap": "false",
   "mon_compact_on_trim": "true",
   "mon_tick_interval": "5",
   "mon_subscribe_interval": "300",
   "mon_osd_laggy_halflife": "3600",
   "mon_osd_laggy_weight": "0.3",
   "mon_osd_adjust_heartbeat_grace": "true",
   "mon_osd_adjust_down_out_interval": "true",
   "mon_osd_auto_mark_in": "false",
   "mon_osd_auto_mark_auto_out_in": "true",
   "mon_osd_auto_mark_new_in": "true",
   "mon_osd_down_out_interval": "300",
   "mon_osd_down_out_subtree_limit": "rack",
   "mon_osd_min_up_ratio": "0.3",
   "mon_osd_min_in_ratio": "0.3",
   "mon_stat_smooth_intervals": "2",
   "mon_lease": "5",
   "mon_lease_renew_interval": "3",
   "mon_lease_ack_timeout": "10",
   "mon_clock_drift_allowed": "0.05",
   "mon_clock_drift_warn_backoff": "5",
   "mon_timecheck_interval": "300",
   "mon_accept_timeout": "10",
   "mon_pg_create_interval": "30",
   "mon_pg_stuck_threshold": "300",
   "mon_osd_full_ratio": "0.95",
   "mon_osd_nearfull_ratio": "0.85",
   "mon_globalid_prealloc": "100",
   "mon_osd_report_timeout": "900",
   "mon_force_standby_active": "true",
   "mon_min_osdmap_epochs": "500",
   "mon_max_pgmap_epochs": "500",
   "mon_max_log_epochs": "500",
   "mon_max_osd": "10000",
   "mon_probe_timeout": "2",
   "mon_slurp_timeout": "10",
   "mon_slurp_bytes": "262144",
   "mon_client_bytes": "104857600",
   "mon_daemon_bytes": "419430400",
   "mon_max_log_entries_per_event": "4096",
   "mon_health_data_update_interval": "60",
   "mon_data_avail_crit": "5",
   "mon_data_avail_warn": "30",
   "mon_config_key_max_entry_size": "4096",
   "mon_sync_trim_timeout": "30",
   "mon_sync_heartbeat_timeout": "30",
   "mon_sync_heartbeat_interval": "5",
   "mon_sync_backoff_timeout": "30",
   "mon_sync_timeout": "30",
   "mon_sync_max_retries": "5",
   "mon_sync_max_payload_size": "1048576",
   "mon_sync_debug": "false",
   "mon_sync_debug_leader": "-1",
   "mon_sync_debug_provider": "-1",
   "mon_sync_debug_provider_fallback": "-1",
   "mon_debug_dump_transactions": "false",
   "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
   "mon_sync_leader_kill_at": "0",
   "mon_sync_provider_kill_at": "0",
   "mon_sync_requester_kill_at": "0",
   "mon_leveldb_write_buffer_size": "33554432",
   "mon_leveldb_cache_size": "268435456",
   "mon_leveldb_block_size": "65536",
   "mon_leveldb_bloom_size": "0",
   "mon_leveldb_max_open_files": "0",
   "mon_leveldb_compression": "false",
   "mon_leveldb_paranoid": "false",
   "mon_leveldb_log": "",
   "paxos_stash_full_interval": "25",
   "paxos_max_join_drift": "100",
   "paxos_propose_interval": "1",
   "paxos_min_wait": "0.05",
   "paxos_min": "500",
   "paxos_trim_min": "500",
   "paxos_trim_max": "1000",
   "paxos_trim_disabled_max_versions": "108000",
   "paxos_service_trim_min": "500",
   "paxos_service_trim_max": "1000",
   "clock_offset": "0",
   "auth_cluster_required": "none",
   "auth_service_required": "none",
   "auth_client_required": "none",
   "auth_supported": "none",
   "cephx_require_signatures": "false",
   "cephx_cluster_require_signatures": "false",
   "cephx_service_require_signatures": "false",
   "cephx_sign_messages": "true",
   "auth_mon_ticket_ttl": "43200",
   "auth_service_ticket_ttl": "3600",
   "auth_debug": "false",
   "mon_client_hunt_interval": "3",
   "mon_client_ping_interval": "10",
   "mon_client_max_log_entries_per_message": "1000",
   "mon_max_pool_pg_num": "65536",
   "mon_pool_quota_warn_threshold": "0",
   "mon_pool_quota_crit_threshold": "0",
   "client_cache_size": "16384",
   "client_cache_mid": "0.75",
   "client_use_random_mds": "false",
   "client_mount_timeout": "300",
   "client_tick_interval": "1",
   "client_trace": "",
   "client_readahead_min": "131072",
   "client_readahead_max_bytes": "0",
   "client_readahead_max_periods": "4",
   "client_snapdir": ".snap",
   "client_mountpoint": "\/",
   "client_notify_timeout": "10",
   "client_caps_release_delay": "5",
   "client_oc": "true",
   "client_oc_size": "209715200",
   "client_oc_max_dirty": "104857600",
   "client_oc_target_dirty": "8388608",
   "client_oc_max_dirty_age": "5",
   "client_oc_max_objects": "1000",
   "client_debug_force_sync_read": "false",
   "client_debug_inject_tick_delay": "0",
   "fuse_use_invalidate_cb": "false",
   "fuse_allow_other": "true",
   "fuse_default_permissions": "true",
   "fuse_big_writes": "true",
   "fuse_atomic_o_trunc": "true",
   "fuse_debug": "false",
   "objecter_tick_interval": "5",
   "objecter_timeout": "10",
   "objecter_inflight_op_bytes": "104857600",
   "objecter_inflight_ops": "1024",
   "journaler_allow_split_entries": "true",
   "journaler_write_head_interval": "15",
   "journaler_prefetch_periods": "10",
   "journaler_prezero_periods": "5",
   "journaler_batch_interval": "0.001",
   "journaler_batch_max": "0",
   "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
   "mds_max_file_size": "1099511627776",
   "mds_cache_size": "100000",
   "mds_cache_mid": "0.7",
   "mds_mem_max": "1048576",
   "mds_dir_commit_ratio": "0.5",
   "mds_dir_max_commit_size": "90",
   "mds_decay_halflife": "5",
   "mds_beacon_interval": "4",
   "mds_beacon_grace": "15",
   "mds_enforce_unique_name": "true",
   "mds_blacklist_interval": "1440",
   "mds_session_timeout": "60",
   "mds_session_autoclose": "300",
   "mds_reconnect_timeout": "45",
   "mds_tick_interval": "5",
   "mds_dirstat_min_interval": "1",
   "mds_scatter_nudge_interval": "5",
   "mds_client_prealloc_inos": "1000",
   "mds_early_reply": "true",
   "mds_use_tmap": "true",
   "mds_default_dir_hash": "2",
   "mds_log": "true",
   "mds_log_skip_corrupt_events": "false",
   "mds_log_max_events": "-1",
   "mds_log_segment_size": "0",
   "mds_log_max_segments": "30",
   "mds_log_max_expiring": "20",
   "mds_bal_sample_interval": "3",
   "mds_bal_replicate_threshold": "8000",
   "mds_bal_unreplicate_threshold": "0",
   "mds_bal_frag": "false",
   "mds_bal_split_size": "10000",
   "mds_bal_split_rd": "25000",
   "mds_bal_split_wr": "10000",
   "mds_bal_split_bits": "3",
   "mds_bal_merge_size": "50",
   "mds_bal_merge_rd": "1000",
   "mds_bal_merge_wr": "1000",
   "mds_bal_interval": "10",
   "mds_bal_fragment_interval": "5",
   "mds_bal_idle_threshold": "0",
   "mds_bal_max": "-1",
   "mds_bal_max_until": "-1",
   "mds_bal_mode": "0",
   "mds_bal_min_rebalance": "0.1",
   "mds_bal_min_start": "0.2",
   "mds_bal_need_min": "0.8",
   "mds_bal_need_max": "1.2",
   "mds_bal_midchunk": "0.3",
   "mds_bal_minchunk": "0.001",
   "mds_bal_target_removal_min": "5",
   "mds_bal_target_removal_max": "10",
   "mds_replay_interval": "1",
   "mds_shutdown_check": "0",
   "mds_thrash_exports": "0",
   "mds_thrash_fragments": "0",
   "mds_dump_cache_on_map": "false",
   "mds_dump_cache_after_rejoin": "false",
   "mds_verify_scatter": "false",
   "mds_debug_scatterstat": "false",
   "mds_debug_frag": "false",
   "mds_debug_auth_pins": "false",
   "mds_debug_subtrees": "false",
   "mds_kill_mdstable_at": "0",
   "mds_kill_export_at": "0",
   "mds_kill_import_at": "0",
   "mds_kill_link_at": "0",
   "mds_kill_rename_at": "0",
   "mds_kill_openc_at": "0",
   "mds_kill_journal_at": "0",
   "mds_kill_journal_expire_at": "0",
   "mds_kill_journal_replay_at": "0",
   "mds_inject_traceless_reply_probability": "0",
   "mds_wipe_sessions": "false",
   "mds_wipe_ino_prealloc": "false",
   "mds_skip_ino": "0",
   "max_mds": "1",
   "mds_standby_for_name": "",
   "mds_standby_for_rank": "-1",
   "mds_standby_replay": "false",
   "osd_auto_upgrade_tmap": "true",
   "osd_tmapput_sets_uses_tmap": "false",
   "osd_max_backfills": "5",
   "osd_backfill_full_ratio": "0.85",
   "osd_backfill_retry_interval": "10",
   "osd_uuid": "00000000-0000-0000-0000-000000000000",
   "osd_data": "\/ceph\/osd.0\/",
   "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
   "osd_journal_size": "5120",
   "osd_max_write_size": "90",
   "osd_max_pgls": "1024",
   "osd_client_message_size_cap": "524288000",
   "osd_client_message_cap": "100",
   "osd_pg_bits": "6",
   "osd_pgp_bits": "6",
   "osd_crush_chooseleaf_type": "1",
   "osd_min_rep": "1",
   "osd_max_rep": "10",
   "osd_pool_default_crush_rule": "0",
   "osd_pool_default_size": "2",
   "osd_pool_default_min_size": "0",
   "osd_pool_default_pg_num": "8",
   "osd_pool_default_pgp_num": "8",
   "osd_pool_default_flags": "0",
   "osd_map_dedup": "true",
   "osd_map_cache_size": "500",
   "osd_map_message_max": "100",
   "osd_map_share_max_epochs": "100",
   "osd_op_threads": "2",
   "osd_peering_wq_batch_size": "20",
   "osd_op_pq_max_tokens_per_priority": "4194304",
   "osd_op_pq_min_cost": "65536",
   "osd_disk_threads": "1",
   "osd_recovery_threads": "2",
   "osd_recover_clone_overlap": "true",
   "osd_backfill_scan_min": "64",
   "osd_backfill_scan_max": "512",
   "osd_op_thread_timeout": "15",
   "osd_recovery_thread_timeout": "30",
   "osd_snap_trim_thread_timeout": "3600",
   "osd_scrub_thread_timeout": "60",
   "osd_scrub_finalize_thread_timeout": "600",
   "osd_remove_thread_timeout": "3600",
   "osd_command_thread_timeout": "600",
   "osd_age": "0.8",
   "osd_age_time": "0",
   "osd_heartbeat_addr": ":\/0",
   "osd_heartbeat_interval": "6",
   "osd_heartbeat_grace": "20",
   "osd_mon_heartbeat_interval": "30",
   "osd_mon_report_interval_max": "120",
   "osd_mon_report_interval_min": "5",
   "osd_pg_stat_report_interval_max": "500",
   "osd_mon_ack_timeout": "30",
   "osd_min_down_reporters": "1",
   "osd_min_down_reports": "3",
   "osd_default_data_pool_replay_window": "45",
   "osd_preserve_trimmed_log": "false",
   "osd_auto_mark_unfound_lost": "false",
   "osd_recovery_delay_start": "0",
   "osd_recovery_max_active": "5",
   "osd_recovery_max_chunk": "8388608",
   "osd_recovery_forget_lost_objects": "false",
   "osd_max_scrubs": "1",
   "osd_scrub_load_threshold": "0.5",
   "osd_scrub_min_interval": "86400",
   "osd_scrub_max_interval": "604800",
   "osd_deep_scrub_interval": "604800",
   "osd_deep_scrub_stride": "524288",
   "osd_scan_list_ping_tp_interval": "100",
   "osd_auto_weight": "false",
   "osd_class_dir": "\/usr\/lib\/rados-classes",
   "osd_check_for_log_corruption": "false",
   "osd_use_stale_snap": "false",
   "osd_rollback_to_cluster_snap": "",
   "osd_default_notify_timeout": "30",
   "osd_kill_backfill_at": "0",
   "osd_pg_epoch_persisted_max_stale": "200",
   "osd_min_pg_log_entries": "500",
   "osd_max_pg_log_entries": "1500",
   "osd_op_complaint_time": "30",
   "osd_command_max_records": "256",
   "osd_op_log_threshold": "5",
   "osd_verify_sparse_read_holes": "false",
   "osd_debug_drop_ping_probability": "0",
   "osd_debug_drop_ping_duration": "0",
   "osd_debug_drop_pg_create_probability": "0",
   "osd_debug_drop_pg_create_duration": "1",
   "osd_debug_drop_op_probability": "0",
   "osd_debug_op_order": "false",
   "osd_debug_verify_snaps_on_info": "false",
   "osd_debug_verify_stray_on_activate": "false",
   "osd_debug_skip_full_check_in_backfill_reservation": "false",
   "osd_op_history_size": "20",
   "osd_op_history_duration": "600",
   "osd_target_transaction_size": "30",
   "osd_failsafe_full_ratio": "0.97",
   "osd_failsafe_nearfull_ratio": "0.9",
   "osd_leveldb_write_buffer_size": "0",
   "osd_leveldb_cache_size": "0",
   "osd_leveldb_block_size": "0",
   "osd_leveldb_bloom_size": "0",
   "osd_leveldb_max_open_files": "0",
   "osd_leveldb_compression": "true",
   "osd_leveldb_paranoid": "false",
   "osd_leveldb_log": "",
   "osd_client_op_priority": "63",
   "osd_recovery_op_priority": "50",
   "osd_mon_shutdown_timeout": "5",
   "filestore": "false",
   "filestore_index_retry_probability": "0",
   "filestore_debug_inject_read_err": "false",
   "filestore_debug_omap_check": "false",
   "filestore_xattr_use_omap": "false",
   "filestore_max_inline_xattr_size": "512",
   "filestore_max_inline_xattrs": "2",
   "filestore_max_sync_interval": "5",
   "filestore_min_sync_interval": "0.01",
   "filestore_btrfs_snap": "true",
   "filestore_btrfs_clone_range": "true",
   "filestore_fsync_flushes_journal_data": "false",
   "filestore_fiemap": "false",
   "filestore_flusher": "true",
   "filestore_flusher_max_fds": "512",
   "filestore_flush_min": "65536",
   "filestore_sync_flush": "false",
   "filestore_journal_parallel": "false",
   "filestore_journal_writeahead": "false",
   "filestore_journal_trailing": "false",
   "filestore_queue_max_ops": "500",
   "filestore_queue_max_bytes": "104857600",
   "filestore_queue_committing_max_ops": "5000",
   "filestore_queue_committing_max_bytes": "104857600",
   "filestore_op_threads": "2",
   "filestore_op_thread_timeout": "60",
   "filestore_op_thread_suicide_timeout": "180",
   "filestore_commit_timeout": "600",
   "filestore_fiemap_threshold": "4096",
   "filestore_merge_threshold": "10",
   "filestore_split_multiple": "2",
   "filestore_update_to": "1000",
   "filestore_blackhole": "false",
   "filestore_dump_file": "",
   "filestore_kill_at": "0",
   "filestore_inject_stall": "0",
   "filestore_fail_eio": "true",
   "filestore_replica_fadvise": "true",
   "filestore_debug_verify_split": "false",
   "journal_dio": "true",
   "journal_aio": "true",
   "journal_force_aio": "false",
   "journal_max_corrupt_search": "10485760",
   "journal_block_align": "true",
   "journal_write_header_frequency": "0",
   "journal_max_write_bytes": "10485760",
   "journal_max_write_entries": "100",
   "journal_queue_max_ops": "300",
   "journal_queue_max_bytes": "33554432",
   "journal_align_min_size": "65536",
   "journal_replay_from": "0",
   "journal_zero_on_create": "false",
   "journal_ignore_corruption": "false",
   "rbd_cache": "false",
   "rbd_cache_writethrough_until_flush": "false",
   "rbd_cache_size": "33554432",
   "rbd_cache_max_dirty": "25165824",
   "rbd_cache_target_dirty": "16777216",
   "rbd_cache_max_dirty_age": "1",
   "rbd_cache_block_writes_upfront": "false",
   "rbd_concurrent_management_ops": "10",
   "rbd_default_format": "1",
   "rbd_default_order": "22",
   "rbd_default_stripe_count": "1",
   "rbd_default_stripe_unit": "4194304",
   "rbd_default_features": "3",
   "nss_db_path": "",
   "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
   "rgw_enable_apis": "s3, swift, swift_auth, admin",
   "rgw_cache_enabled": "true",
   "rgw_cache_lru_size": "10000",
   "rgw_socket_path": "",
   "rgw_host": "",
   "rgw_port": "",
   "rgw_dns_name": "",
   "rgw_script_uri": "",
   "rgw_request_uri": "",
   "rgw_swift_url": "",
   "rgw_swift_url_prefix": "swift",
   "rgw_swift_auth_url": "",
   "rgw_swift_auth_entry": "auth",
   "rgw_keystone_url": "",
   "rgw_keystone_admin_token": "",
   "rgw_keystone_accepted_roles": "Member, admin",
   "rgw_keystone_token_cache_size": "10000",
   "rgw_keystone_revocation_interval": "900",
   "rgw_admin_entry": "admin",
   "rgw_enforce_swift_acls": "true",
   "rgw_swift_token_expiration": "86400",
   "rgw_print_continue": "true",
   "rgw_remote_addr_param": "REMOTE_ADDR",
   "rgw_op_thread_timeout": "600",
   "rgw_op_thread_suicide_timeout": "0",
   "rgw_thread_pool_size": "100",
   "rgw_num_control_oids": "8",
   "rgw_zone_root_pool": ".rgw.root",
   "rgw_log_nonexistent_bucket": "false",
   "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
   "rgw_log_object_name_utc": "false",
   "rgw_usage_max_shards": "32",
   "rgw_usage_max_user_shards": "1",
   "rgw_enable_ops_log": "false",
   "rgw_enable_usage_log": "false",
   "rgw_ops_log_rados": "true",
   "rgw_ops_log_socket_path": "",
   "rgw_ops_log_data_backlog": "5242880",
   "rgw_usage_log_flush_threshold": "1024",
   "rgw_usage_log_tick_interval": "30",
   "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
   "rgw_intent_log_object_name_utc": "false",
   "rgw_init_timeout": "300",
   "rgw_mime_types_file": "\/etc\/mime.types",
   "rgw_gc_max_objs": "32",
   "rgw_gc_obj_min_wait": "7200",
   "rgw_gc_processor_max_time": "3600",
   "rgw_gc_processor_period": "3600",
   "rgw_s3_success_create_obj_status": "0",
   "rgw_resolve_cname": "false",
   "rgw_obj_stripe_size": "4194304",
   "rgw_extended_http_attrs": "",
   "rgw_exit_timeout_secs": "120",
   "rgw_get_obj_window_size": "16777216",
   "rgw_get_obj_max_req_size": "4194304",
   "rgw_relaxed_s3_bucket_names": "false",
   "rgw_list_buckets_max_chunk": "1000",
   "mutex_perf_counter": "false",
   "internal_safe_to_start_threads": "true"}



Stefan

> -Sam
>
> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> Mike we already have the async patch running. Yes it helps but only helps it
>> does not solve. It just hides the issue ...
>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>
>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>
>>> - ceph osd set noout
>>>
>>> - Reboot one of the nodes hosting OSDs
>>>       - VMs mounted from RBD volumes work properly
>>>
>>> - I see the OSD's boot messages as they re-join the cluster
>>>
>>> - Start seeing active+recovery_wait, peering, and active+recovering
>>>       - VMs mounted from RBD volumes become unresponsive.
>>>
>>> - Recovery completes
>>>       - VMs mounted from RBD volumes regain responsiveness
>>>
>>> - ceph osd unset noout
>>>
>>> Would joshd's async patch for qemu help here, or is there something else
>>> going on?
>>>
>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>
>>> Thanks,
>>>
>>> Mike Dawson
>>> Co-Founder & Director of Cloud Architecture
>>> Cloudapt LLC
>>> 6330 East 75th Street, Suite 170
>>> Indianapolis, IN 46250
>>>
>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>
>>>> Can you reproduce and attach the ceph.log from before you stop the osd
>>>> until after you have started the osd and it has recovered?
>>>> -Sam
>>>>
>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>> <s.priebe@profihost.ag> wrote:
>>>>>
>>>>> Hi,
>>>>>
>>>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>>>> it seem to hang for around 2-4 minutes and then recovery seems to start
>>>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>>>> get a lot of slow request messages an hanging VMs.
>>>>>
>>>>> What i noticed today is that if i leave the OSD off as long as ceph
>>>>> starts to backfill - the recovery and "re" backfilling wents absolutely
>>>>> smooth without any issues and no slow request messages at all.
>>>>>
>>>>> Does anybody have an idea why?
>>>>>
>>>>> Greets,
>>>>> Stefan
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>>> the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>
>>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-02  7:44         ` Stefan Priebe
@ 2013-08-02 17:35           ` Samuel Just
  2013-08-02 18:16             ` Stefan Priebe
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-02 17:35 UTC (permalink / raw)
  To: Stefan Priebe; +Cc: Mike Dawson, ceph-devel

You might try turning osd_max_backfills to 2 or 1.
-Sam

On Fri, Aug 2, 2013 at 12:44 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd settings?
>
>> sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
>
> Sure.
>
>
>
> { "name": "osd.0",
>   "cluster": "ceph",
>   "none": "0\/5",
>   "lockdep": "0\/0",
>   "context": "0\/0",
>   "crush": "0\/0",
>   "mds": "0\/0",
>   "mds_balancer": "0\/0",
>   "mds_locker": "0\/0",
>   "mds_log": "0\/0",
>   "mds_log_expire": "0\/0",
>   "mds_migrator": "0\/0",
>   "buffer": "0\/0",
>   "timer": "0\/0",
>   "filer": "0\/0",
>   "striper": "0\/1",
>   "objecter": "0\/0",
>   "rados": "0\/0",
>   "rbd": "0\/0",
>   "journaler": "0\/0",
>   "objectcacher": "0\/0",
>   "client": "0\/0",
>   "osd": "0\/0",
>   "optracker": "0\/0",
>   "objclass": "0\/0",
>   "filestore": "0\/0",
>   "journal": "0\/0",
>   "ms": "0\/0",
>   "mon": "0\/0",
>   "monc": "0\/0",
>   "paxos": "0\/0",
>   "tp": "0\/0",
>   "auth": "0\/0",
>   "crypto": "1\/5",
>   "finisher": "0\/0",
>   "heartbeatmap": "0\/0",
>   "perfcounter": "0\/0",
>   "rgw": "0\/0",
>   "hadoop": "0\/0",
>   "javaclient": "1\/5",
>   "asok": "0\/0",
>   "throttle": "0\/0",
>   "host": "cloud1-1268",
>   "fsid": "00000000-0000-0000-0000-000000000000",
>   "public_addr": "10.255.0.90:0\/0",
>   "cluster_addr": "10.255.0.90:0\/0",
>   "public_network": "10.255.0.1\/24",
>   "cluster_network": "10.255.0.1\/24",
>   "num_client": "1",
>   "monmap": "",
>   "mon_host": "",
>   "lockdep": "false",
>   "run_dir": "\/var\/run\/ceph",
>   "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
>   "daemonize": "true",
>   "pid_file": "\/var\/run\/ceph\/osd.0.pid",
>   "chdir": "\/",
>   "max_open_files": "0",
>   "fatal_signal_handlers": "true",
>   "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
>   "log_max_new": "1000",
>   "log_max_recent": "10000",
>   "log_to_stderr": "false",
>   "err_to_stderr": "true",
>   "log_to_syslog": "false",
>   "err_to_syslog": "false",
>   "log_flush_on_exit": "true",
>   "log_stop_at_utilization": "0.97",
>   "clog_to_monitors": "true",
>   "clog_to_syslog": "false",
>   "clog_to_syslog_level": "info",
>   "clog_to_syslog_facility": "daemon",
>   "mon_cluster_log_to_syslog": "false",
>   "mon_cluster_log_to_syslog_level": "info",
>   "mon_cluster_log_to_syslog_facility": "daemon",
>   "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
>   "key": "",
>   "keyfile": "",
>   "keyring": "\/etc\/ceph\/osd.0.keyring",
>   "heartbeat_interval": "5",
>   "heartbeat_file": "",
>   "heartbeat_inject_failure": "0",
>   "perf": "true",
>   "ms_tcp_nodelay": "true",
>   "ms_tcp_rcvbuf": "0",
>   "ms_initial_backoff": "0.2",
>   "ms_max_backoff": "15",
>   "ms_nocrc": "false",
>   "ms_die_on_bad_msg": "false",
>   "ms_die_on_unhandled_msg": "false",
>   "ms_dispatch_throttle_bytes": "104857600",
>   "ms_bind_ipv6": "false",
>   "ms_bind_port_min": "6800",
>   "ms_bind_port_max": "7100",
>   "ms_rwthread_stack_bytes": "1048576",
>   "ms_tcp_read_timeout": "900",
>   "ms_pq_max_tokens_per_priority": "4194304",
>   "ms_pq_min_cost": "65536",
>   "ms_inject_socket_failures": "0",
>   "ms_inject_delay_type": "",
>   "ms_inject_delay_max": "1",
>   "ms_inject_delay_probability": "0",
>   "ms_inject_internal_delays": "0",
>   "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
>   "mon_initial_members": "",
>   "mon_sync_fs_threshold": "5",
>   "mon_compact_on_start": "false",
>   "mon_compact_on_bootstrap": "false",
>   "mon_compact_on_trim": "true",
>   "mon_tick_interval": "5",
>   "mon_subscribe_interval": "300",
>   "mon_osd_laggy_halflife": "3600",
>   "mon_osd_laggy_weight": "0.3",
>   "mon_osd_adjust_heartbeat_grace": "true",
>   "mon_osd_adjust_down_out_interval": "true",
>   "mon_osd_auto_mark_in": "false",
>   "mon_osd_auto_mark_auto_out_in": "true",
>   "mon_osd_auto_mark_new_in": "true",
>   "mon_osd_down_out_interval": "300",
>   "mon_osd_down_out_subtree_limit": "rack",
>   "mon_osd_min_up_ratio": "0.3",
>   "mon_osd_min_in_ratio": "0.3",
>   "mon_stat_smooth_intervals": "2",
>   "mon_lease": "5",
>   "mon_lease_renew_interval": "3",
>   "mon_lease_ack_timeout": "10",
>   "mon_clock_drift_allowed": "0.05",
>   "mon_clock_drift_warn_backoff": "5",
>   "mon_timecheck_interval": "300",
>   "mon_accept_timeout": "10",
>   "mon_pg_create_interval": "30",
>   "mon_pg_stuck_threshold": "300",
>   "mon_osd_full_ratio": "0.95",
>   "mon_osd_nearfull_ratio": "0.85",
>   "mon_globalid_prealloc": "100",
>   "mon_osd_report_timeout": "900",
>   "mon_force_standby_active": "true",
>   "mon_min_osdmap_epochs": "500",
>   "mon_max_pgmap_epochs": "500",
>   "mon_max_log_epochs": "500",
>   "mon_max_osd": "10000",
>   "mon_probe_timeout": "2",
>   "mon_slurp_timeout": "10",
>   "mon_slurp_bytes": "262144",
>   "mon_client_bytes": "104857600",
>   "mon_daemon_bytes": "419430400",
>   "mon_max_log_entries_per_event": "4096",
>   "mon_health_data_update_interval": "60",
>   "mon_data_avail_crit": "5",
>   "mon_data_avail_warn": "30",
>   "mon_config_key_max_entry_size": "4096",
>   "mon_sync_trim_timeout": "30",
>   "mon_sync_heartbeat_timeout": "30",
>   "mon_sync_heartbeat_interval": "5",
>   "mon_sync_backoff_timeout": "30",
>   "mon_sync_timeout": "30",
>   "mon_sync_max_retries": "5",
>   "mon_sync_max_payload_size": "1048576",
>   "mon_sync_debug": "false",
>   "mon_sync_debug_leader": "-1",
>   "mon_sync_debug_provider": "-1",
>   "mon_sync_debug_provider_fallback": "-1",
>   "mon_debug_dump_transactions": "false",
>   "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
>   "mon_sync_leader_kill_at": "0",
>   "mon_sync_provider_kill_at": "0",
>   "mon_sync_requester_kill_at": "0",
>   "mon_leveldb_write_buffer_size": "33554432",
>   "mon_leveldb_cache_size": "268435456",
>   "mon_leveldb_block_size": "65536",
>   "mon_leveldb_bloom_size": "0",
>   "mon_leveldb_max_open_files": "0",
>   "mon_leveldb_compression": "false",
>   "mon_leveldb_paranoid": "false",
>   "mon_leveldb_log": "",
>   "paxos_stash_full_interval": "25",
>   "paxos_max_join_drift": "100",
>   "paxos_propose_interval": "1",
>   "paxos_min_wait": "0.05",
>   "paxos_min": "500",
>   "paxos_trim_min": "500",
>   "paxos_trim_max": "1000",
>   "paxos_trim_disabled_max_versions": "108000",
>   "paxos_service_trim_min": "500",
>   "paxos_service_trim_max": "1000",
>   "clock_offset": "0",
>   "auth_cluster_required": "none",
>   "auth_service_required": "none",
>   "auth_client_required": "none",
>   "auth_supported": "none",
>   "cephx_require_signatures": "false",
>   "cephx_cluster_require_signatures": "false",
>   "cephx_service_require_signatures": "false",
>   "cephx_sign_messages": "true",
>   "auth_mon_ticket_ttl": "43200",
>   "auth_service_ticket_ttl": "3600",
>   "auth_debug": "false",
>   "mon_client_hunt_interval": "3",
>   "mon_client_ping_interval": "10",
>   "mon_client_max_log_entries_per_message": "1000",
>   "mon_max_pool_pg_num": "65536",
>   "mon_pool_quota_warn_threshold": "0",
>   "mon_pool_quota_crit_threshold": "0",
>   "client_cache_size": "16384",
>   "client_cache_mid": "0.75",
>   "client_use_random_mds": "false",
>   "client_mount_timeout": "300",
>   "client_tick_interval": "1",
>   "client_trace": "",
>   "client_readahead_min": "131072",
>   "client_readahead_max_bytes": "0",
>   "client_readahead_max_periods": "4",
>   "client_snapdir": ".snap",
>   "client_mountpoint": "\/",
>   "client_notify_timeout": "10",
>   "client_caps_release_delay": "5",
>   "client_oc": "true",
>   "client_oc_size": "209715200",
>   "client_oc_max_dirty": "104857600",
>   "client_oc_target_dirty": "8388608",
>   "client_oc_max_dirty_age": "5",
>   "client_oc_max_objects": "1000",
>   "client_debug_force_sync_read": "false",
>   "client_debug_inject_tick_delay": "0",
>   "fuse_use_invalidate_cb": "false",
>   "fuse_allow_other": "true",
>   "fuse_default_permissions": "true",
>   "fuse_big_writes": "true",
>   "fuse_atomic_o_trunc": "true",
>   "fuse_debug": "false",
>   "objecter_tick_interval": "5",
>   "objecter_timeout": "10",
>   "objecter_inflight_op_bytes": "104857600",
>   "objecter_inflight_ops": "1024",
>   "journaler_allow_split_entries": "true",
>   "journaler_write_head_interval": "15",
>   "journaler_prefetch_periods": "10",
>   "journaler_prezero_periods": "5",
>   "journaler_batch_interval": "0.001",
>   "journaler_batch_max": "0",
>   "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
>   "mds_max_file_size": "1099511627776",
>   "mds_cache_size": "100000",
>   "mds_cache_mid": "0.7",
>   "mds_mem_max": "1048576",
>   "mds_dir_commit_ratio": "0.5",
>   "mds_dir_max_commit_size": "90",
>   "mds_decay_halflife": "5",
>   "mds_beacon_interval": "4",
>   "mds_beacon_grace": "15",
>   "mds_enforce_unique_name": "true",
>   "mds_blacklist_interval": "1440",
>   "mds_session_timeout": "60",
>   "mds_session_autoclose": "300",
>   "mds_reconnect_timeout": "45",
>   "mds_tick_interval": "5",
>   "mds_dirstat_min_interval": "1",
>   "mds_scatter_nudge_interval": "5",
>   "mds_client_prealloc_inos": "1000",
>   "mds_early_reply": "true",
>   "mds_use_tmap": "true",
>   "mds_default_dir_hash": "2",
>   "mds_log": "true",
>   "mds_log_skip_corrupt_events": "false",
>   "mds_log_max_events": "-1",
>   "mds_log_segment_size": "0",
>   "mds_log_max_segments": "30",
>   "mds_log_max_expiring": "20",
>   "mds_bal_sample_interval": "3",
>   "mds_bal_replicate_threshold": "8000",
>   "mds_bal_unreplicate_threshold": "0",
>   "mds_bal_frag": "false",
>   "mds_bal_split_size": "10000",
>   "mds_bal_split_rd": "25000",
>   "mds_bal_split_wr": "10000",
>   "mds_bal_split_bits": "3",
>   "mds_bal_merge_size": "50",
>   "mds_bal_merge_rd": "1000",
>   "mds_bal_merge_wr": "1000",
>   "mds_bal_interval": "10",
>   "mds_bal_fragment_interval": "5",
>   "mds_bal_idle_threshold": "0",
>   "mds_bal_max": "-1",
>   "mds_bal_max_until": "-1",
>   "mds_bal_mode": "0",
>   "mds_bal_min_rebalance": "0.1",
>   "mds_bal_min_start": "0.2",
>   "mds_bal_need_min": "0.8",
>   "mds_bal_need_max": "1.2",
>   "mds_bal_midchunk": "0.3",
>   "mds_bal_minchunk": "0.001",
>   "mds_bal_target_removal_min": "5",
>   "mds_bal_target_removal_max": "10",
>   "mds_replay_interval": "1",
>   "mds_shutdown_check": "0",
>   "mds_thrash_exports": "0",
>   "mds_thrash_fragments": "0",
>   "mds_dump_cache_on_map": "false",
>   "mds_dump_cache_after_rejoin": "false",
>   "mds_verify_scatter": "false",
>   "mds_debug_scatterstat": "false",
>   "mds_debug_frag": "false",
>   "mds_debug_auth_pins": "false",
>   "mds_debug_subtrees": "false",
>   "mds_kill_mdstable_at": "0",
>   "mds_kill_export_at": "0",
>   "mds_kill_import_at": "0",
>   "mds_kill_link_at": "0",
>   "mds_kill_rename_at": "0",
>   "mds_kill_openc_at": "0",
>   "mds_kill_journal_at": "0",
>   "mds_kill_journal_expire_at": "0",
>   "mds_kill_journal_replay_at": "0",
>   "mds_inject_traceless_reply_probability": "0",
>   "mds_wipe_sessions": "false",
>   "mds_wipe_ino_prealloc": "false",
>   "mds_skip_ino": "0",
>   "max_mds": "1",
>   "mds_standby_for_name": "",
>   "mds_standby_for_rank": "-1",
>   "mds_standby_replay": "false",
>   "osd_auto_upgrade_tmap": "true",
>   "osd_tmapput_sets_uses_tmap": "false",
>   "osd_max_backfills": "5",
>   "osd_backfill_full_ratio": "0.85",
>   "osd_backfill_retry_interval": "10",
>   "osd_uuid": "00000000-0000-0000-0000-000000000000",
>   "osd_data": "\/ceph\/osd.0\/",
>   "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
>   "osd_journal_size": "5120",
>   "osd_max_write_size": "90",
>   "osd_max_pgls": "1024",
>   "osd_client_message_size_cap": "524288000",
>   "osd_client_message_cap": "100",
>   "osd_pg_bits": "6",
>   "osd_pgp_bits": "6",
>   "osd_crush_chooseleaf_type": "1",
>   "osd_min_rep": "1",
>   "osd_max_rep": "10",
>   "osd_pool_default_crush_rule": "0",
>   "osd_pool_default_size": "2",
>   "osd_pool_default_min_size": "0",
>   "osd_pool_default_pg_num": "8",
>   "osd_pool_default_pgp_num": "8",
>   "osd_pool_default_flags": "0",
>   "osd_map_dedup": "true",
>   "osd_map_cache_size": "500",
>   "osd_map_message_max": "100",
>   "osd_map_share_max_epochs": "100",
>   "osd_op_threads": "2",
>   "osd_peering_wq_batch_size": "20",
>   "osd_op_pq_max_tokens_per_priority": "4194304",
>   "osd_op_pq_min_cost": "65536",
>   "osd_disk_threads": "1",
>   "osd_recovery_threads": "2",
>   "osd_recover_clone_overlap": "true",
>   "osd_backfill_scan_min": "64",
>   "osd_backfill_scan_max": "512",
>   "osd_op_thread_timeout": "15",
>   "osd_recovery_thread_timeout": "30",
>   "osd_snap_trim_thread_timeout": "3600",
>   "osd_scrub_thread_timeout": "60",
>   "osd_scrub_finalize_thread_timeout": "600",
>   "osd_remove_thread_timeout": "3600",
>   "osd_command_thread_timeout": "600",
>   "osd_age": "0.8",
>   "osd_age_time": "0",
>   "osd_heartbeat_addr": ":\/0",
>   "osd_heartbeat_interval": "6",
>   "osd_heartbeat_grace": "20",
>   "osd_mon_heartbeat_interval": "30",
>   "osd_mon_report_interval_max": "120",
>   "osd_mon_report_interval_min": "5",
>   "osd_pg_stat_report_interval_max": "500",
>   "osd_mon_ack_timeout": "30",
>   "osd_min_down_reporters": "1",
>   "osd_min_down_reports": "3",
>   "osd_default_data_pool_replay_window": "45",
>   "osd_preserve_trimmed_log": "false",
>   "osd_auto_mark_unfound_lost": "false",
>   "osd_recovery_delay_start": "0",
>   "osd_recovery_max_active": "5",
>   "osd_recovery_max_chunk": "8388608",
>   "osd_recovery_forget_lost_objects": "false",
>   "osd_max_scrubs": "1",
>   "osd_scrub_load_threshold": "0.5",
>   "osd_scrub_min_interval": "86400",
>   "osd_scrub_max_interval": "604800",
>   "osd_deep_scrub_interval": "604800",
>   "osd_deep_scrub_stride": "524288",
>   "osd_scan_list_ping_tp_interval": "100",
>   "osd_auto_weight": "false",
>   "osd_class_dir": "\/usr\/lib\/rados-classes",
>   "osd_check_for_log_corruption": "false",
>   "osd_use_stale_snap": "false",
>   "osd_rollback_to_cluster_snap": "",
>   "osd_default_notify_timeout": "30",
>   "osd_kill_backfill_at": "0",
>   "osd_pg_epoch_persisted_max_stale": "200",
>   "osd_min_pg_log_entries": "500",
>   "osd_max_pg_log_entries": "1500",
>   "osd_op_complaint_time": "30",
>   "osd_command_max_records": "256",
>   "osd_op_log_threshold": "5",
>   "osd_verify_sparse_read_holes": "false",
>   "osd_debug_drop_ping_probability": "0",
>   "osd_debug_drop_ping_duration": "0",
>   "osd_debug_drop_pg_create_probability": "0",
>   "osd_debug_drop_pg_create_duration": "1",
>   "osd_debug_drop_op_probability": "0",
>   "osd_debug_op_order": "false",
>   "osd_debug_verify_snaps_on_info": "false",
>   "osd_debug_verify_stray_on_activate": "false",
>   "osd_debug_skip_full_check_in_backfill_reservation": "false",
>   "osd_op_history_size": "20",
>   "osd_op_history_duration": "600",
>   "osd_target_transaction_size": "30",
>   "osd_failsafe_full_ratio": "0.97",
>   "osd_failsafe_nearfull_ratio": "0.9",
>   "osd_leveldb_write_buffer_size": "0",
>   "osd_leveldb_cache_size": "0",
>   "osd_leveldb_block_size": "0",
>   "osd_leveldb_bloom_size": "0",
>   "osd_leveldb_max_open_files": "0",
>   "osd_leveldb_compression": "true",
>   "osd_leveldb_paranoid": "false",
>   "osd_leveldb_log": "",
>   "osd_client_op_priority": "63",
>   "osd_recovery_op_priority": "50",
>   "osd_mon_shutdown_timeout": "5",
>   "filestore": "false",
>   "filestore_index_retry_probability": "0",
>   "filestore_debug_inject_read_err": "false",
>   "filestore_debug_omap_check": "false",
>   "filestore_xattr_use_omap": "false",
>   "filestore_max_inline_xattr_size": "512",
>   "filestore_max_inline_xattrs": "2",
>   "filestore_max_sync_interval": "5",
>   "filestore_min_sync_interval": "0.01",
>   "filestore_btrfs_snap": "true",
>   "filestore_btrfs_clone_range": "true",
>   "filestore_fsync_flushes_journal_data": "false",
>   "filestore_fiemap": "false",
>   "filestore_flusher": "true",
>   "filestore_flusher_max_fds": "512",
>   "filestore_flush_min": "65536",
>   "filestore_sync_flush": "false",
>   "filestore_journal_parallel": "false",
>   "filestore_journal_writeahead": "false",
>   "filestore_journal_trailing": "false",
>   "filestore_queue_max_ops": "500",
>   "filestore_queue_max_bytes": "104857600",
>   "filestore_queue_committing_max_ops": "5000",
>   "filestore_queue_committing_max_bytes": "104857600",
>   "filestore_op_threads": "2",
>   "filestore_op_thread_timeout": "60",
>   "filestore_op_thread_suicide_timeout": "180",
>   "filestore_commit_timeout": "600",
>   "filestore_fiemap_threshold": "4096",
>   "filestore_merge_threshold": "10",
>   "filestore_split_multiple": "2",
>   "filestore_update_to": "1000",
>   "filestore_blackhole": "false",
>   "filestore_dump_file": "",
>   "filestore_kill_at": "0",
>   "filestore_inject_stall": "0",
>   "filestore_fail_eio": "true",
>   "filestore_replica_fadvise": "true",
>   "filestore_debug_verify_split": "false",
>   "journal_dio": "true",
>   "journal_aio": "true",
>   "journal_force_aio": "false",
>   "journal_max_corrupt_search": "10485760",
>   "journal_block_align": "true",
>   "journal_write_header_frequency": "0",
>   "journal_max_write_bytes": "10485760",
>   "journal_max_write_entries": "100",
>   "journal_queue_max_ops": "300",
>   "journal_queue_max_bytes": "33554432",
>   "journal_align_min_size": "65536",
>   "journal_replay_from": "0",
>   "journal_zero_on_create": "false",
>   "journal_ignore_corruption": "false",
>   "rbd_cache": "false",
>   "rbd_cache_writethrough_until_flush": "false",
>   "rbd_cache_size": "33554432",
>   "rbd_cache_max_dirty": "25165824",
>   "rbd_cache_target_dirty": "16777216",
>   "rbd_cache_max_dirty_age": "1",
>   "rbd_cache_block_writes_upfront": "false",
>   "rbd_concurrent_management_ops": "10",
>   "rbd_default_format": "1",
>   "rbd_default_order": "22",
>   "rbd_default_stripe_count": "1",
>   "rbd_default_stripe_unit": "4194304",
>   "rbd_default_features": "3",
>   "nss_db_path": "",
>   "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
>   "rgw_enable_apis": "s3, swift, swift_auth, admin",
>   "rgw_cache_enabled": "true",
>   "rgw_cache_lru_size": "10000",
>   "rgw_socket_path": "",
>   "rgw_host": "",
>   "rgw_port": "",
>   "rgw_dns_name": "",
>   "rgw_script_uri": "",
>   "rgw_request_uri": "",
>   "rgw_swift_url": "",
>   "rgw_swift_url_prefix": "swift",
>   "rgw_swift_auth_url": "",
>   "rgw_swift_auth_entry": "auth",
>   "rgw_keystone_url": "",
>   "rgw_keystone_admin_token": "",
>   "rgw_keystone_accepted_roles": "Member, admin",
>   "rgw_keystone_token_cache_size": "10000",
>   "rgw_keystone_revocation_interval": "900",
>   "rgw_admin_entry": "admin",
>   "rgw_enforce_swift_acls": "true",
>   "rgw_swift_token_expiration": "86400",
>   "rgw_print_continue": "true",
>   "rgw_remote_addr_param": "REMOTE_ADDR",
>   "rgw_op_thread_timeout": "600",
>   "rgw_op_thread_suicide_timeout": "0",
>   "rgw_thread_pool_size": "100",
>   "rgw_num_control_oids": "8",
>   "rgw_zone_root_pool": ".rgw.root",
>   "rgw_log_nonexistent_bucket": "false",
>   "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
>   "rgw_log_object_name_utc": "false",
>   "rgw_usage_max_shards": "32",
>   "rgw_usage_max_user_shards": "1",
>   "rgw_enable_ops_log": "false",
>   "rgw_enable_usage_log": "false",
>   "rgw_ops_log_rados": "true",
>   "rgw_ops_log_socket_path": "",
>   "rgw_ops_log_data_backlog": "5242880",
>   "rgw_usage_log_flush_threshold": "1024",
>   "rgw_usage_log_tick_interval": "30",
>   "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
>   "rgw_intent_log_object_name_utc": "false",
>   "rgw_init_timeout": "300",
>   "rgw_mime_types_file": "\/etc\/mime.types",
>   "rgw_gc_max_objs": "32",
>   "rgw_gc_obj_min_wait": "7200",
>   "rgw_gc_processor_max_time": "3600",
>   "rgw_gc_processor_period": "3600",
>   "rgw_s3_success_create_obj_status": "0",
>   "rgw_resolve_cname": "false",
>   "rgw_obj_stripe_size": "4194304",
>   "rgw_extended_http_attrs": "",
>   "rgw_exit_timeout_secs": "120",
>   "rgw_get_obj_window_size": "16777216",
>   "rgw_get_obj_max_req_size": "4194304",
>   "rgw_relaxed_s3_bucket_names": "false",
>   "rgw_list_buckets_max_chunk": "1000",
>   "mutex_perf_counter": "false",
>   "internal_safe_to_start_threads": "true"}
>
>
>
> Stefan
>
>
>> -Sam
>>
>> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe <s.priebe@profihost.ag>
>> wrote:
>>>
>>> Mike we already have the async patch running. Yes it helps but only helps
>>> it
>>> does not solve. It just hides the issue ...
>>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>>
>>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>>
>>>> - ceph osd set noout
>>>>
>>>> - Reboot one of the nodes hosting OSDs
>>>>       - VMs mounted from RBD volumes work properly
>>>>
>>>> - I see the OSD's boot messages as they re-join the cluster
>>>>
>>>> - Start seeing active+recovery_wait, peering, and active+recovering
>>>>       - VMs mounted from RBD volumes become unresponsive.
>>>>
>>>> - Recovery completes
>>>>       - VMs mounted from RBD volumes regain responsiveness
>>>>
>>>> - ceph osd unset noout
>>>>
>>>> Would joshd's async patch for qemu help here, or is there something else
>>>> going on?
>>>>
>>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>>
>>>> Thanks,
>>>>
>>>> Mike Dawson
>>>> Co-Founder & Director of Cloud Architecture
>>>> Cloudapt LLC
>>>> 6330 East 75th Street, Suite 170
>>>> Indianapolis, IN 46250
>>>>
>>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>>
>>>>>
>>>>> Can you reproduce and attach the ceph.log from before you stop the osd
>>>>> until after you have started the osd and it has recovered?
>>>>> -Sam
>>>>>
>>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>>>>> it seem to hang for around 2-4 minutes and then recovery seems to
>>>>>> start
>>>>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>>>>> get a lot of slow request messages an hanging VMs.
>>>>>>
>>>>>> What i noticed today is that if i leave the OSD off as long as ceph
>>>>>> starts to backfill - the recovery and "re" backfilling wents
>>>>>> absolutely
>>>>>> smooth without any issues and no slow request messages at all.
>>>>>>
>>>>>> Does anybody have an idea why?
>>>>>>
>>>>>> Greets,
>>>>>> Stefan
>>>>>> --
>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>> in
>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>>>
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>> in
>>>>> the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-01 18:38   ` Samuel Just
@ 2013-08-02 17:56     ` Andrey Korolyov
  0 siblings, 0 replies; 40+ messages in thread
From: Andrey Korolyov @ 2013-08-02 17:56 UTC (permalink / raw)
  To: Samuel Just; +Cc: Stefan Priebe - Profihost AG, ceph-devel

Created #5844.

On Thu, Aug 1, 2013 at 10:38 PM, Samuel Just <sam.just@inktank.com> wrote:
> Is there a bug open for this?  I suspect we don't sufficiently
> throttle the snapshot removal work.
> -Sam
>
> On Thu, Aug 1, 2013 at 7:50 AM, Andrey Korolyov <andrey@xdel.ru> wrote:
>> Second this. Also for long-lasting snapshot problem and related
>> performance issues I may say that cuttlefish improved things greatly,
>> but creation/deletion of large snapshot (hundreds of gigabytes of
>> commited data) still can bring down cluster for a minutes, despite
>> usage of every possible optimization.
>>
>> On Thu, Aug 1, 2013 at 12:22 PM, Stefan Priebe - Profihost AG
>> <s.priebe@profihost.ag> wrote:
>>> Hi,
>>>
>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>> it seem to hang for around 2-4 minutes and then recovery seems to start
>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>> get a lot of slow request messages an hanging VMs.
>>>
>>> What i noticed today is that if i leave the OSD off as long as ceph
>>> starts to backfill - the recovery and "re" backfilling wents absolutely
>>> smooth without any issues and no slow request messages at all.
>>>
>>> Does anybody have an idea why?
>>>
>>> Greets,
>>> Stefan
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-02 17:35           ` Samuel Just
@ 2013-08-02 18:16             ` Stefan Priebe
  2013-08-02 18:21               ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-02 18:16 UTC (permalink / raw)
  To: Samuel Just; +Cc: Mike Dawson, ceph-devel

I already tried both values this makes no difference. The drives are not 
the bottleneck.

Am 02.08.2013 19:35, schrieb Samuel Just:
> You might try turning osd_max_backfills to 2 or 1.
> -Sam
>
> On Fri, Aug 2, 2013 at 12:44 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd settings?
>>
>>> sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
>>
>> Sure.
>>
>>
>>
>> { "name": "osd.0",
>>    "cluster": "ceph",
>>    "none": "0\/5",
>>    "lockdep": "0\/0",
>>    "context": "0\/0",
>>    "crush": "0\/0",
>>    "mds": "0\/0",
>>    "mds_balancer": "0\/0",
>>    "mds_locker": "0\/0",
>>    "mds_log": "0\/0",
>>    "mds_log_expire": "0\/0",
>>    "mds_migrator": "0\/0",
>>    "buffer": "0\/0",
>>    "timer": "0\/0",
>>    "filer": "0\/0",
>>    "striper": "0\/1",
>>    "objecter": "0\/0",
>>    "rados": "0\/0",
>>    "rbd": "0\/0",
>>    "journaler": "0\/0",
>>    "objectcacher": "0\/0",
>>    "client": "0\/0",
>>    "osd": "0\/0",
>>    "optracker": "0\/0",
>>    "objclass": "0\/0",
>>    "filestore": "0\/0",
>>    "journal": "0\/0",
>>    "ms": "0\/0",
>>    "mon": "0\/0",
>>    "monc": "0\/0",
>>    "paxos": "0\/0",
>>    "tp": "0\/0",
>>    "auth": "0\/0",
>>    "crypto": "1\/5",
>>    "finisher": "0\/0",
>>    "heartbeatmap": "0\/0",
>>    "perfcounter": "0\/0",
>>    "rgw": "0\/0",
>>    "hadoop": "0\/0",
>>    "javaclient": "1\/5",
>>    "asok": "0\/0",
>>    "throttle": "0\/0",
>>    "host": "cloud1-1268",
>>    "fsid": "00000000-0000-0000-0000-000000000000",
>>    "public_addr": "10.255.0.90:0\/0",
>>    "cluster_addr": "10.255.0.90:0\/0",
>>    "public_network": "10.255.0.1\/24",
>>    "cluster_network": "10.255.0.1\/24",
>>    "num_client": "1",
>>    "monmap": "",
>>    "mon_host": "",
>>    "lockdep": "false",
>>    "run_dir": "\/var\/run\/ceph",
>>    "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
>>    "daemonize": "true",
>>    "pid_file": "\/var\/run\/ceph\/osd.0.pid",
>>    "chdir": "\/",
>>    "max_open_files": "0",
>>    "fatal_signal_handlers": "true",
>>    "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
>>    "log_max_new": "1000",
>>    "log_max_recent": "10000",
>>    "log_to_stderr": "false",
>>    "err_to_stderr": "true",
>>    "log_to_syslog": "false",
>>    "err_to_syslog": "false",
>>    "log_flush_on_exit": "true",
>>    "log_stop_at_utilization": "0.97",
>>    "clog_to_monitors": "true",
>>    "clog_to_syslog": "false",
>>    "clog_to_syslog_level": "info",
>>    "clog_to_syslog_facility": "daemon",
>>    "mon_cluster_log_to_syslog": "false",
>>    "mon_cluster_log_to_syslog_level": "info",
>>    "mon_cluster_log_to_syslog_facility": "daemon",
>>    "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
>>    "key": "",
>>    "keyfile": "",
>>    "keyring": "\/etc\/ceph\/osd.0.keyring",
>>    "heartbeat_interval": "5",
>>    "heartbeat_file": "",
>>    "heartbeat_inject_failure": "0",
>>    "perf": "true",
>>    "ms_tcp_nodelay": "true",
>>    "ms_tcp_rcvbuf": "0",
>>    "ms_initial_backoff": "0.2",
>>    "ms_max_backoff": "15",
>>    "ms_nocrc": "false",
>>    "ms_die_on_bad_msg": "false",
>>    "ms_die_on_unhandled_msg": "false",
>>    "ms_dispatch_throttle_bytes": "104857600",
>>    "ms_bind_ipv6": "false",
>>    "ms_bind_port_min": "6800",
>>    "ms_bind_port_max": "7100",
>>    "ms_rwthread_stack_bytes": "1048576",
>>    "ms_tcp_read_timeout": "900",
>>    "ms_pq_max_tokens_per_priority": "4194304",
>>    "ms_pq_min_cost": "65536",
>>    "ms_inject_socket_failures": "0",
>>    "ms_inject_delay_type": "",
>>    "ms_inject_delay_max": "1",
>>    "ms_inject_delay_probability": "0",
>>    "ms_inject_internal_delays": "0",
>>    "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
>>    "mon_initial_members": "",
>>    "mon_sync_fs_threshold": "5",
>>    "mon_compact_on_start": "false",
>>    "mon_compact_on_bootstrap": "false",
>>    "mon_compact_on_trim": "true",
>>    "mon_tick_interval": "5",
>>    "mon_subscribe_interval": "300",
>>    "mon_osd_laggy_halflife": "3600",
>>    "mon_osd_laggy_weight": "0.3",
>>    "mon_osd_adjust_heartbeat_grace": "true",
>>    "mon_osd_adjust_down_out_interval": "true",
>>    "mon_osd_auto_mark_in": "false",
>>    "mon_osd_auto_mark_auto_out_in": "true",
>>    "mon_osd_auto_mark_new_in": "true",
>>    "mon_osd_down_out_interval": "300",
>>    "mon_osd_down_out_subtree_limit": "rack",
>>    "mon_osd_min_up_ratio": "0.3",
>>    "mon_osd_min_in_ratio": "0.3",
>>    "mon_stat_smooth_intervals": "2",
>>    "mon_lease": "5",
>>    "mon_lease_renew_interval": "3",
>>    "mon_lease_ack_timeout": "10",
>>    "mon_clock_drift_allowed": "0.05",
>>    "mon_clock_drift_warn_backoff": "5",
>>    "mon_timecheck_interval": "300",
>>    "mon_accept_timeout": "10",
>>    "mon_pg_create_interval": "30",
>>    "mon_pg_stuck_threshold": "300",
>>    "mon_osd_full_ratio": "0.95",
>>    "mon_osd_nearfull_ratio": "0.85",
>>    "mon_globalid_prealloc": "100",
>>    "mon_osd_report_timeout": "900",
>>    "mon_force_standby_active": "true",
>>    "mon_min_osdmap_epochs": "500",
>>    "mon_max_pgmap_epochs": "500",
>>    "mon_max_log_epochs": "500",
>>    "mon_max_osd": "10000",
>>    "mon_probe_timeout": "2",
>>    "mon_slurp_timeout": "10",
>>    "mon_slurp_bytes": "262144",
>>    "mon_client_bytes": "104857600",
>>    "mon_daemon_bytes": "419430400",
>>    "mon_max_log_entries_per_event": "4096",
>>    "mon_health_data_update_interval": "60",
>>    "mon_data_avail_crit": "5",
>>    "mon_data_avail_warn": "30",
>>    "mon_config_key_max_entry_size": "4096",
>>    "mon_sync_trim_timeout": "30",
>>    "mon_sync_heartbeat_timeout": "30",
>>    "mon_sync_heartbeat_interval": "5",
>>    "mon_sync_backoff_timeout": "30",
>>    "mon_sync_timeout": "30",
>>    "mon_sync_max_retries": "5",
>>    "mon_sync_max_payload_size": "1048576",
>>    "mon_sync_debug": "false",
>>    "mon_sync_debug_leader": "-1",
>>    "mon_sync_debug_provider": "-1",
>>    "mon_sync_debug_provider_fallback": "-1",
>>    "mon_debug_dump_transactions": "false",
>>    "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
>>    "mon_sync_leader_kill_at": "0",
>>    "mon_sync_provider_kill_at": "0",
>>    "mon_sync_requester_kill_at": "0",
>>    "mon_leveldb_write_buffer_size": "33554432",
>>    "mon_leveldb_cache_size": "268435456",
>>    "mon_leveldb_block_size": "65536",
>>    "mon_leveldb_bloom_size": "0",
>>    "mon_leveldb_max_open_files": "0",
>>    "mon_leveldb_compression": "false",
>>    "mon_leveldb_paranoid": "false",
>>    "mon_leveldb_log": "",
>>    "paxos_stash_full_interval": "25",
>>    "paxos_max_join_drift": "100",
>>    "paxos_propose_interval": "1",
>>    "paxos_min_wait": "0.05",
>>    "paxos_min": "500",
>>    "paxos_trim_min": "500",
>>    "paxos_trim_max": "1000",
>>    "paxos_trim_disabled_max_versions": "108000",
>>    "paxos_service_trim_min": "500",
>>    "paxos_service_trim_max": "1000",
>>    "clock_offset": "0",
>>    "auth_cluster_required": "none",
>>    "auth_service_required": "none",
>>    "auth_client_required": "none",
>>    "auth_supported": "none",
>>    "cephx_require_signatures": "false",
>>    "cephx_cluster_require_signatures": "false",
>>    "cephx_service_require_signatures": "false",
>>    "cephx_sign_messages": "true",
>>    "auth_mon_ticket_ttl": "43200",
>>    "auth_service_ticket_ttl": "3600",
>>    "auth_debug": "false",
>>    "mon_client_hunt_interval": "3",
>>    "mon_client_ping_interval": "10",
>>    "mon_client_max_log_entries_per_message": "1000",
>>    "mon_max_pool_pg_num": "65536",
>>    "mon_pool_quota_warn_threshold": "0",
>>    "mon_pool_quota_crit_threshold": "0",
>>    "client_cache_size": "16384",
>>    "client_cache_mid": "0.75",
>>    "client_use_random_mds": "false",
>>    "client_mount_timeout": "300",
>>    "client_tick_interval": "1",
>>    "client_trace": "",
>>    "client_readahead_min": "131072",
>>    "client_readahead_max_bytes": "0",
>>    "client_readahead_max_periods": "4",
>>    "client_snapdir": ".snap",
>>    "client_mountpoint": "\/",
>>    "client_notify_timeout": "10",
>>    "client_caps_release_delay": "5",
>>    "client_oc": "true",
>>    "client_oc_size": "209715200",
>>    "client_oc_max_dirty": "104857600",
>>    "client_oc_target_dirty": "8388608",
>>    "client_oc_max_dirty_age": "5",
>>    "client_oc_max_objects": "1000",
>>    "client_debug_force_sync_read": "false",
>>    "client_debug_inject_tick_delay": "0",
>>    "fuse_use_invalidate_cb": "false",
>>    "fuse_allow_other": "true",
>>    "fuse_default_permissions": "true",
>>    "fuse_big_writes": "true",
>>    "fuse_atomic_o_trunc": "true",
>>    "fuse_debug": "false",
>>    "objecter_tick_interval": "5",
>>    "objecter_timeout": "10",
>>    "objecter_inflight_op_bytes": "104857600",
>>    "objecter_inflight_ops": "1024",
>>    "journaler_allow_split_entries": "true",
>>    "journaler_write_head_interval": "15",
>>    "journaler_prefetch_periods": "10",
>>    "journaler_prezero_periods": "5",
>>    "journaler_batch_interval": "0.001",
>>    "journaler_batch_max": "0",
>>    "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
>>    "mds_max_file_size": "1099511627776",
>>    "mds_cache_size": "100000",
>>    "mds_cache_mid": "0.7",
>>    "mds_mem_max": "1048576",
>>    "mds_dir_commit_ratio": "0.5",
>>    "mds_dir_max_commit_size": "90",
>>    "mds_decay_halflife": "5",
>>    "mds_beacon_interval": "4",
>>    "mds_beacon_grace": "15",
>>    "mds_enforce_unique_name": "true",
>>    "mds_blacklist_interval": "1440",
>>    "mds_session_timeout": "60",
>>    "mds_session_autoclose": "300",
>>    "mds_reconnect_timeout": "45",
>>    "mds_tick_interval": "5",
>>    "mds_dirstat_min_interval": "1",
>>    "mds_scatter_nudge_interval": "5",
>>    "mds_client_prealloc_inos": "1000",
>>    "mds_early_reply": "true",
>>    "mds_use_tmap": "true",
>>    "mds_default_dir_hash": "2",
>>    "mds_log": "true",
>>    "mds_log_skip_corrupt_events": "false",
>>    "mds_log_max_events": "-1",
>>    "mds_log_segment_size": "0",
>>    "mds_log_max_segments": "30",
>>    "mds_log_max_expiring": "20",
>>    "mds_bal_sample_interval": "3",
>>    "mds_bal_replicate_threshold": "8000",
>>    "mds_bal_unreplicate_threshold": "0",
>>    "mds_bal_frag": "false",
>>    "mds_bal_split_size": "10000",
>>    "mds_bal_split_rd": "25000",
>>    "mds_bal_split_wr": "10000",
>>    "mds_bal_split_bits": "3",
>>    "mds_bal_merge_size": "50",
>>    "mds_bal_merge_rd": "1000",
>>    "mds_bal_merge_wr": "1000",
>>    "mds_bal_interval": "10",
>>    "mds_bal_fragment_interval": "5",
>>    "mds_bal_idle_threshold": "0",
>>    "mds_bal_max": "-1",
>>    "mds_bal_max_until": "-1",
>>    "mds_bal_mode": "0",
>>    "mds_bal_min_rebalance": "0.1",
>>    "mds_bal_min_start": "0.2",
>>    "mds_bal_need_min": "0.8",
>>    "mds_bal_need_max": "1.2",
>>    "mds_bal_midchunk": "0.3",
>>    "mds_bal_minchunk": "0.001",
>>    "mds_bal_target_removal_min": "5",
>>    "mds_bal_target_removal_max": "10",
>>    "mds_replay_interval": "1",
>>    "mds_shutdown_check": "0",
>>    "mds_thrash_exports": "0",
>>    "mds_thrash_fragments": "0",
>>    "mds_dump_cache_on_map": "false",
>>    "mds_dump_cache_after_rejoin": "false",
>>    "mds_verify_scatter": "false",
>>    "mds_debug_scatterstat": "false",
>>    "mds_debug_frag": "false",
>>    "mds_debug_auth_pins": "false",
>>    "mds_debug_subtrees": "false",
>>    "mds_kill_mdstable_at": "0",
>>    "mds_kill_export_at": "0",
>>    "mds_kill_import_at": "0",
>>    "mds_kill_link_at": "0",
>>    "mds_kill_rename_at": "0",
>>    "mds_kill_openc_at": "0",
>>    "mds_kill_journal_at": "0",
>>    "mds_kill_journal_expire_at": "0",
>>    "mds_kill_journal_replay_at": "0",
>>    "mds_inject_traceless_reply_probability": "0",
>>    "mds_wipe_sessions": "false",
>>    "mds_wipe_ino_prealloc": "false",
>>    "mds_skip_ino": "0",
>>    "max_mds": "1",
>>    "mds_standby_for_name": "",
>>    "mds_standby_for_rank": "-1",
>>    "mds_standby_replay": "false",
>>    "osd_auto_upgrade_tmap": "true",
>>    "osd_tmapput_sets_uses_tmap": "false",
>>    "osd_max_backfills": "5",
>>    "osd_backfill_full_ratio": "0.85",
>>    "osd_backfill_retry_interval": "10",
>>    "osd_uuid": "00000000-0000-0000-0000-000000000000",
>>    "osd_data": "\/ceph\/osd.0\/",
>>    "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
>>    "osd_journal_size": "5120",
>>    "osd_max_write_size": "90",
>>    "osd_max_pgls": "1024",
>>    "osd_client_message_size_cap": "524288000",
>>    "osd_client_message_cap": "100",
>>    "osd_pg_bits": "6",
>>    "osd_pgp_bits": "6",
>>    "osd_crush_chooseleaf_type": "1",
>>    "osd_min_rep": "1",
>>    "osd_max_rep": "10",
>>    "osd_pool_default_crush_rule": "0",
>>    "osd_pool_default_size": "2",
>>    "osd_pool_default_min_size": "0",
>>    "osd_pool_default_pg_num": "8",
>>    "osd_pool_default_pgp_num": "8",
>>    "osd_pool_default_flags": "0",
>>    "osd_map_dedup": "true",
>>    "osd_map_cache_size": "500",
>>    "osd_map_message_max": "100",
>>    "osd_map_share_max_epochs": "100",
>>    "osd_op_threads": "2",
>>    "osd_peering_wq_batch_size": "20",
>>    "osd_op_pq_max_tokens_per_priority": "4194304",
>>    "osd_op_pq_min_cost": "65536",
>>    "osd_disk_threads": "1",
>>    "osd_recovery_threads": "2",
>>    "osd_recover_clone_overlap": "true",
>>    "osd_backfill_scan_min": "64",
>>    "osd_backfill_scan_max": "512",
>>    "osd_op_thread_timeout": "15",
>>    "osd_recovery_thread_timeout": "30",
>>    "osd_snap_trim_thread_timeout": "3600",
>>    "osd_scrub_thread_timeout": "60",
>>    "osd_scrub_finalize_thread_timeout": "600",
>>    "osd_remove_thread_timeout": "3600",
>>    "osd_command_thread_timeout": "600",
>>    "osd_age": "0.8",
>>    "osd_age_time": "0",
>>    "osd_heartbeat_addr": ":\/0",
>>    "osd_heartbeat_interval": "6",
>>    "osd_heartbeat_grace": "20",
>>    "osd_mon_heartbeat_interval": "30",
>>    "osd_mon_report_interval_max": "120",
>>    "osd_mon_report_interval_min": "5",
>>    "osd_pg_stat_report_interval_max": "500",
>>    "osd_mon_ack_timeout": "30",
>>    "osd_min_down_reporters": "1",
>>    "osd_min_down_reports": "3",
>>    "osd_default_data_pool_replay_window": "45",
>>    "osd_preserve_trimmed_log": "false",
>>    "osd_auto_mark_unfound_lost": "false",
>>    "osd_recovery_delay_start": "0",
>>    "osd_recovery_max_active": "5",
>>    "osd_recovery_max_chunk": "8388608",
>>    "osd_recovery_forget_lost_objects": "false",
>>    "osd_max_scrubs": "1",
>>    "osd_scrub_load_threshold": "0.5",
>>    "osd_scrub_min_interval": "86400",
>>    "osd_scrub_max_interval": "604800",
>>    "osd_deep_scrub_interval": "604800",
>>    "osd_deep_scrub_stride": "524288",
>>    "osd_scan_list_ping_tp_interval": "100",
>>    "osd_auto_weight": "false",
>>    "osd_class_dir": "\/usr\/lib\/rados-classes",
>>    "osd_check_for_log_corruption": "false",
>>    "osd_use_stale_snap": "false",
>>    "osd_rollback_to_cluster_snap": "",
>>    "osd_default_notify_timeout": "30",
>>    "osd_kill_backfill_at": "0",
>>    "osd_pg_epoch_persisted_max_stale": "200",
>>    "osd_min_pg_log_entries": "500",
>>    "osd_max_pg_log_entries": "1500",
>>    "osd_op_complaint_time": "30",
>>    "osd_command_max_records": "256",
>>    "osd_op_log_threshold": "5",
>>    "osd_verify_sparse_read_holes": "false",
>>    "osd_debug_drop_ping_probability": "0",
>>    "osd_debug_drop_ping_duration": "0",
>>    "osd_debug_drop_pg_create_probability": "0",
>>    "osd_debug_drop_pg_create_duration": "1",
>>    "osd_debug_drop_op_probability": "0",
>>    "osd_debug_op_order": "false",
>>    "osd_debug_verify_snaps_on_info": "false",
>>    "osd_debug_verify_stray_on_activate": "false",
>>    "osd_debug_skip_full_check_in_backfill_reservation": "false",
>>    "osd_op_history_size": "20",
>>    "osd_op_history_duration": "600",
>>    "osd_target_transaction_size": "30",
>>    "osd_failsafe_full_ratio": "0.97",
>>    "osd_failsafe_nearfull_ratio": "0.9",
>>    "osd_leveldb_write_buffer_size": "0",
>>    "osd_leveldb_cache_size": "0",
>>    "osd_leveldb_block_size": "0",
>>    "osd_leveldb_bloom_size": "0",
>>    "osd_leveldb_max_open_files": "0",
>>    "osd_leveldb_compression": "true",
>>    "osd_leveldb_paranoid": "false",
>>    "osd_leveldb_log": "",
>>    "osd_client_op_priority": "63",
>>    "osd_recovery_op_priority": "50",
>>    "osd_mon_shutdown_timeout": "5",
>>    "filestore": "false",
>>    "filestore_index_retry_probability": "0",
>>    "filestore_debug_inject_read_err": "false",
>>    "filestore_debug_omap_check": "false",
>>    "filestore_xattr_use_omap": "false",
>>    "filestore_max_inline_xattr_size": "512",
>>    "filestore_max_inline_xattrs": "2",
>>    "filestore_max_sync_interval": "5",
>>    "filestore_min_sync_interval": "0.01",
>>    "filestore_btrfs_snap": "true",
>>    "filestore_btrfs_clone_range": "true",
>>    "filestore_fsync_flushes_journal_data": "false",
>>    "filestore_fiemap": "false",
>>    "filestore_flusher": "true",
>>    "filestore_flusher_max_fds": "512",
>>    "filestore_flush_min": "65536",
>>    "filestore_sync_flush": "false",
>>    "filestore_journal_parallel": "false",
>>    "filestore_journal_writeahead": "false",
>>    "filestore_journal_trailing": "false",
>>    "filestore_queue_max_ops": "500",
>>    "filestore_queue_max_bytes": "104857600",
>>    "filestore_queue_committing_max_ops": "5000",
>>    "filestore_queue_committing_max_bytes": "104857600",
>>    "filestore_op_threads": "2",
>>    "filestore_op_thread_timeout": "60",
>>    "filestore_op_thread_suicide_timeout": "180",
>>    "filestore_commit_timeout": "600",
>>    "filestore_fiemap_threshold": "4096",
>>    "filestore_merge_threshold": "10",
>>    "filestore_split_multiple": "2",
>>    "filestore_update_to": "1000",
>>    "filestore_blackhole": "false",
>>    "filestore_dump_file": "",
>>    "filestore_kill_at": "0",
>>    "filestore_inject_stall": "0",
>>    "filestore_fail_eio": "true",
>>    "filestore_replica_fadvise": "true",
>>    "filestore_debug_verify_split": "false",
>>    "journal_dio": "true",
>>    "journal_aio": "true",
>>    "journal_force_aio": "false",
>>    "journal_max_corrupt_search": "10485760",
>>    "journal_block_align": "true",
>>    "journal_write_header_frequency": "0",
>>    "journal_max_write_bytes": "10485760",
>>    "journal_max_write_entries": "100",
>>    "journal_queue_max_ops": "300",
>>    "journal_queue_max_bytes": "33554432",
>>    "journal_align_min_size": "65536",
>>    "journal_replay_from": "0",
>>    "journal_zero_on_create": "false",
>>    "journal_ignore_corruption": "false",
>>    "rbd_cache": "false",
>>    "rbd_cache_writethrough_until_flush": "false",
>>    "rbd_cache_size": "33554432",
>>    "rbd_cache_max_dirty": "25165824",
>>    "rbd_cache_target_dirty": "16777216",
>>    "rbd_cache_max_dirty_age": "1",
>>    "rbd_cache_block_writes_upfront": "false",
>>    "rbd_concurrent_management_ops": "10",
>>    "rbd_default_format": "1",
>>    "rbd_default_order": "22",
>>    "rbd_default_stripe_count": "1",
>>    "rbd_default_stripe_unit": "4194304",
>>    "rbd_default_features": "3",
>>    "nss_db_path": "",
>>    "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
>>    "rgw_enable_apis": "s3, swift, swift_auth, admin",
>>    "rgw_cache_enabled": "true",
>>    "rgw_cache_lru_size": "10000",
>>    "rgw_socket_path": "",
>>    "rgw_host": "",
>>    "rgw_port": "",
>>    "rgw_dns_name": "",
>>    "rgw_script_uri": "",
>>    "rgw_request_uri": "",
>>    "rgw_swift_url": "",
>>    "rgw_swift_url_prefix": "swift",
>>    "rgw_swift_auth_url": "",
>>    "rgw_swift_auth_entry": "auth",
>>    "rgw_keystone_url": "",
>>    "rgw_keystone_admin_token": "",
>>    "rgw_keystone_accepted_roles": "Member, admin",
>>    "rgw_keystone_token_cache_size": "10000",
>>    "rgw_keystone_revocation_interval": "900",
>>    "rgw_admin_entry": "admin",
>>    "rgw_enforce_swift_acls": "true",
>>    "rgw_swift_token_expiration": "86400",
>>    "rgw_print_continue": "true",
>>    "rgw_remote_addr_param": "REMOTE_ADDR",
>>    "rgw_op_thread_timeout": "600",
>>    "rgw_op_thread_suicide_timeout": "0",
>>    "rgw_thread_pool_size": "100",
>>    "rgw_num_control_oids": "8",
>>    "rgw_zone_root_pool": ".rgw.root",
>>    "rgw_log_nonexistent_bucket": "false",
>>    "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
>>    "rgw_log_object_name_utc": "false",
>>    "rgw_usage_max_shards": "32",
>>    "rgw_usage_max_user_shards": "1",
>>    "rgw_enable_ops_log": "false",
>>    "rgw_enable_usage_log": "false",
>>    "rgw_ops_log_rados": "true",
>>    "rgw_ops_log_socket_path": "",
>>    "rgw_ops_log_data_backlog": "5242880",
>>    "rgw_usage_log_flush_threshold": "1024",
>>    "rgw_usage_log_tick_interval": "30",
>>    "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
>>    "rgw_intent_log_object_name_utc": "false",
>>    "rgw_init_timeout": "300",
>>    "rgw_mime_types_file": "\/etc\/mime.types",
>>    "rgw_gc_max_objs": "32",
>>    "rgw_gc_obj_min_wait": "7200",
>>    "rgw_gc_processor_max_time": "3600",
>>    "rgw_gc_processor_period": "3600",
>>    "rgw_s3_success_create_obj_status": "0",
>>    "rgw_resolve_cname": "false",
>>    "rgw_obj_stripe_size": "4194304",
>>    "rgw_extended_http_attrs": "",
>>    "rgw_exit_timeout_secs": "120",
>>    "rgw_get_obj_window_size": "16777216",
>>    "rgw_get_obj_max_req_size": "4194304",
>>    "rgw_relaxed_s3_bucket_names": "false",
>>    "rgw_list_buckets_max_chunk": "1000",
>>    "mutex_perf_counter": "false",
>>    "internal_safe_to_start_threads": "true"}
>>
>>
>>
>> Stefan
>>
>>
>>> -Sam
>>>
>>> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe <s.priebe@profihost.ag>
>>> wrote:
>>>>
>>>> Mike we already have the async patch running. Yes it helps but only helps
>>>> it
>>>> does not solve. It just hides the issue ...
>>>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>>>
>>>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>>>
>>>>> - ceph osd set noout
>>>>>
>>>>> - Reboot one of the nodes hosting OSDs
>>>>>        - VMs mounted from RBD volumes work properly
>>>>>
>>>>> - I see the OSD's boot messages as they re-join the cluster
>>>>>
>>>>> - Start seeing active+recovery_wait, peering, and active+recovering
>>>>>        - VMs mounted from RBD volumes become unresponsive.
>>>>>
>>>>> - Recovery completes
>>>>>        - VMs mounted from RBD volumes regain responsiveness
>>>>>
>>>>> - ceph osd unset noout
>>>>>
>>>>> Would joshd's async patch for qemu help here, or is there something else
>>>>> going on?
>>>>>
>>>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>>>
>>>>> Thanks,
>>>>>
>>>>> Mike Dawson
>>>>> Co-Founder & Director of Cloud Architecture
>>>>> Cloudapt LLC
>>>>> 6330 East 75th Street, Suite 170
>>>>> Indianapolis, IN 46250
>>>>>
>>>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>>>
>>>>>>
>>>>>> Can you reproduce and attach the ceph.log from before you stop the osd
>>>>>> until after you have started the osd and it has recovered?
>>>>>> -Sam
>>>>>>
>>>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>
>>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> i still have recovery issues with cuttlefish. After the OSD comes back
>>>>>>> it seem to hang for around 2-4 minutes and then recovery seems to
>>>>>>> start
>>>>>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7. I
>>>>>>> get a lot of slow request messages an hanging VMs.
>>>>>>>
>>>>>>> What i noticed today is that if i leave the OSD off as long as ceph
>>>>>>> starts to backfill - the recovery and "re" backfilling wents
>>>>>>> absolutely
>>>>>>> smooth without any issues and no slow request messages at all.
>>>>>>>
>>>>>>> Does anybody have an idea why?
>>>>>>>
>>>>>>> Greets,
>>>>>>> Stefan
>>>>>>> --
>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>>> in
>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>>>
>>>>>> --
>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>> in
>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-02 18:16             ` Stefan Priebe
@ 2013-08-02 18:21               ` Samuel Just
  2013-08-02 18:46                 ` Stefan Priebe
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-02 18:21 UTC (permalink / raw)
  To: Stefan Priebe; +Cc: Mike Dawson, ceph-devel

Also, you have osd_recovery_op_priority at 50.  That is close to the
priority of client IO.  You want it below 10 (defaults to 10), perhaps
at 1.  You can also adjust down osd_recovery_max_active.
-Sam

On Fri, Aug 2, 2013 at 11:16 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> I already tried both values this makes no difference. The drives are not the
> bottleneck.
>
> Am 02.08.2013 19:35, schrieb Samuel Just:
>
>> You might try turning osd_max_backfills to 2 or 1.
>> -Sam
>>
>> On Fri, Aug 2, 2013 at 12:44 AM, Stefan Priebe <s.priebe@profihost.ag>
>> wrote:
>>>
>>> Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd
>>> settings?
>>>
>>>> sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
>>>
>>>
>>> Sure.
>>>
>>>
>>>
>>> { "name": "osd.0",
>>>    "cluster": "ceph",
>>>    "none": "0\/5",
>>>    "lockdep": "0\/0",
>>>    "context": "0\/0",
>>>    "crush": "0\/0",
>>>    "mds": "0\/0",
>>>    "mds_balancer": "0\/0",
>>>    "mds_locker": "0\/0",
>>>    "mds_log": "0\/0",
>>>    "mds_log_expire": "0\/0",
>>>    "mds_migrator": "0\/0",
>>>    "buffer": "0\/0",
>>>    "timer": "0\/0",
>>>    "filer": "0\/0",
>>>    "striper": "0\/1",
>>>    "objecter": "0\/0",
>>>    "rados": "0\/0",
>>>    "rbd": "0\/0",
>>>    "journaler": "0\/0",
>>>    "objectcacher": "0\/0",
>>>    "client": "0\/0",
>>>    "osd": "0\/0",
>>>    "optracker": "0\/0",
>>>    "objclass": "0\/0",
>>>    "filestore": "0\/0",
>>>    "journal": "0\/0",
>>>    "ms": "0\/0",
>>>    "mon": "0\/0",
>>>    "monc": "0\/0",
>>>    "paxos": "0\/0",
>>>    "tp": "0\/0",
>>>    "auth": "0\/0",
>>>    "crypto": "1\/5",
>>>    "finisher": "0\/0",
>>>    "heartbeatmap": "0\/0",
>>>    "perfcounter": "0\/0",
>>>    "rgw": "0\/0",
>>>    "hadoop": "0\/0",
>>>    "javaclient": "1\/5",
>>>    "asok": "0\/0",
>>>    "throttle": "0\/0",
>>>    "host": "cloud1-1268",
>>>    "fsid": "00000000-0000-0000-0000-000000000000",
>>>    "public_addr": "10.255.0.90:0\/0",
>>>    "cluster_addr": "10.255.0.90:0\/0",
>>>    "public_network": "10.255.0.1\/24",
>>>    "cluster_network": "10.255.0.1\/24",
>>>    "num_client": "1",
>>>    "monmap": "",
>>>    "mon_host": "",
>>>    "lockdep": "false",
>>>    "run_dir": "\/var\/run\/ceph",
>>>    "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
>>>    "daemonize": "true",
>>>    "pid_file": "\/var\/run\/ceph\/osd.0.pid",
>>>    "chdir": "\/",
>>>    "max_open_files": "0",
>>>    "fatal_signal_handlers": "true",
>>>    "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
>>>    "log_max_new": "1000",
>>>    "log_max_recent": "10000",
>>>    "log_to_stderr": "false",
>>>    "err_to_stderr": "true",
>>>    "log_to_syslog": "false",
>>>    "err_to_syslog": "false",
>>>    "log_flush_on_exit": "true",
>>>    "log_stop_at_utilization": "0.97",
>>>    "clog_to_monitors": "true",
>>>    "clog_to_syslog": "false",
>>>    "clog_to_syslog_level": "info",
>>>    "clog_to_syslog_facility": "daemon",
>>>    "mon_cluster_log_to_syslog": "false",
>>>    "mon_cluster_log_to_syslog_level": "info",
>>>    "mon_cluster_log_to_syslog_facility": "daemon",
>>>    "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
>>>    "key": "",
>>>    "keyfile": "",
>>>    "keyring": "\/etc\/ceph\/osd.0.keyring",
>>>    "heartbeat_interval": "5",
>>>    "heartbeat_file": "",
>>>    "heartbeat_inject_failure": "0",
>>>    "perf": "true",
>>>    "ms_tcp_nodelay": "true",
>>>    "ms_tcp_rcvbuf": "0",
>>>    "ms_initial_backoff": "0.2",
>>>    "ms_max_backoff": "15",
>>>    "ms_nocrc": "false",
>>>    "ms_die_on_bad_msg": "false",
>>>    "ms_die_on_unhandled_msg": "false",
>>>    "ms_dispatch_throttle_bytes": "104857600",
>>>    "ms_bind_ipv6": "false",
>>>    "ms_bind_port_min": "6800",
>>>    "ms_bind_port_max": "7100",
>>>    "ms_rwthread_stack_bytes": "1048576",
>>>    "ms_tcp_read_timeout": "900",
>>>    "ms_pq_max_tokens_per_priority": "4194304",
>>>    "ms_pq_min_cost": "65536",
>>>    "ms_inject_socket_failures": "0",
>>>    "ms_inject_delay_type": "",
>>>    "ms_inject_delay_max": "1",
>>>    "ms_inject_delay_probability": "0",
>>>    "ms_inject_internal_delays": "0",
>>>    "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
>>>    "mon_initial_members": "",
>>>    "mon_sync_fs_threshold": "5",
>>>    "mon_compact_on_start": "false",
>>>    "mon_compact_on_bootstrap": "false",
>>>    "mon_compact_on_trim": "true",
>>>    "mon_tick_interval": "5",
>>>    "mon_subscribe_interval": "300",
>>>    "mon_osd_laggy_halflife": "3600",
>>>    "mon_osd_laggy_weight": "0.3",
>>>    "mon_osd_adjust_heartbeat_grace": "true",
>>>    "mon_osd_adjust_down_out_interval": "true",
>>>    "mon_osd_auto_mark_in": "false",
>>>    "mon_osd_auto_mark_auto_out_in": "true",
>>>    "mon_osd_auto_mark_new_in": "true",
>>>    "mon_osd_down_out_interval": "300",
>>>    "mon_osd_down_out_subtree_limit": "rack",
>>>    "mon_osd_min_up_ratio": "0.3",
>>>    "mon_osd_min_in_ratio": "0.3",
>>>    "mon_stat_smooth_intervals": "2",
>>>    "mon_lease": "5",
>>>    "mon_lease_renew_interval": "3",
>>>    "mon_lease_ack_timeout": "10",
>>>    "mon_clock_drift_allowed": "0.05",
>>>    "mon_clock_drift_warn_backoff": "5",
>>>    "mon_timecheck_interval": "300",
>>>    "mon_accept_timeout": "10",
>>>    "mon_pg_create_interval": "30",
>>>    "mon_pg_stuck_threshold": "300",
>>>    "mon_osd_full_ratio": "0.95",
>>>    "mon_osd_nearfull_ratio": "0.85",
>>>    "mon_globalid_prealloc": "100",
>>>    "mon_osd_report_timeout": "900",
>>>    "mon_force_standby_active": "true",
>>>    "mon_min_osdmap_epochs": "500",
>>>    "mon_max_pgmap_epochs": "500",
>>>    "mon_max_log_epochs": "500",
>>>    "mon_max_osd": "10000",
>>>    "mon_probe_timeout": "2",
>>>    "mon_slurp_timeout": "10",
>>>    "mon_slurp_bytes": "262144",
>>>    "mon_client_bytes": "104857600",
>>>    "mon_daemon_bytes": "419430400",
>>>    "mon_max_log_entries_per_event": "4096",
>>>    "mon_health_data_update_interval": "60",
>>>    "mon_data_avail_crit": "5",
>>>    "mon_data_avail_warn": "30",
>>>    "mon_config_key_max_entry_size": "4096",
>>>    "mon_sync_trim_timeout": "30",
>>>    "mon_sync_heartbeat_timeout": "30",
>>>    "mon_sync_heartbeat_interval": "5",
>>>    "mon_sync_backoff_timeout": "30",
>>>    "mon_sync_timeout": "30",
>>>    "mon_sync_max_retries": "5",
>>>    "mon_sync_max_payload_size": "1048576",
>>>    "mon_sync_debug": "false",
>>>    "mon_sync_debug_leader": "-1",
>>>    "mon_sync_debug_provider": "-1",
>>>    "mon_sync_debug_provider_fallback": "-1",
>>>    "mon_debug_dump_transactions": "false",
>>>    "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
>>>    "mon_sync_leader_kill_at": "0",
>>>    "mon_sync_provider_kill_at": "0",
>>>    "mon_sync_requester_kill_at": "0",
>>>    "mon_leveldb_write_buffer_size": "33554432",
>>>    "mon_leveldb_cache_size": "268435456",
>>>    "mon_leveldb_block_size": "65536",
>>>    "mon_leveldb_bloom_size": "0",
>>>    "mon_leveldb_max_open_files": "0",
>>>    "mon_leveldb_compression": "false",
>>>    "mon_leveldb_paranoid": "false",
>>>    "mon_leveldb_log": "",
>>>    "paxos_stash_full_interval": "25",
>>>    "paxos_max_join_drift": "100",
>>>    "paxos_propose_interval": "1",
>>>    "paxos_min_wait": "0.05",
>>>    "paxos_min": "500",
>>>    "paxos_trim_min": "500",
>>>    "paxos_trim_max": "1000",
>>>    "paxos_trim_disabled_max_versions": "108000",
>>>    "paxos_service_trim_min": "500",
>>>    "paxos_service_trim_max": "1000",
>>>    "clock_offset": "0",
>>>    "auth_cluster_required": "none",
>>>    "auth_service_required": "none",
>>>    "auth_client_required": "none",
>>>    "auth_supported": "none",
>>>    "cephx_require_signatures": "false",
>>>    "cephx_cluster_require_signatures": "false",
>>>    "cephx_service_require_signatures": "false",
>>>    "cephx_sign_messages": "true",
>>>    "auth_mon_ticket_ttl": "43200",
>>>    "auth_service_ticket_ttl": "3600",
>>>    "auth_debug": "false",
>>>    "mon_client_hunt_interval": "3",
>>>    "mon_client_ping_interval": "10",
>>>    "mon_client_max_log_entries_per_message": "1000",
>>>    "mon_max_pool_pg_num": "65536",
>>>    "mon_pool_quota_warn_threshold": "0",
>>>    "mon_pool_quota_crit_threshold": "0",
>>>    "client_cache_size": "16384",
>>>    "client_cache_mid": "0.75",
>>>    "client_use_random_mds": "false",
>>>    "client_mount_timeout": "300",
>>>    "client_tick_interval": "1",
>>>    "client_trace": "",
>>>    "client_readahead_min": "131072",
>>>    "client_readahead_max_bytes": "0",
>>>    "client_readahead_max_periods": "4",
>>>    "client_snapdir": ".snap",
>>>    "client_mountpoint": "\/",
>>>    "client_notify_timeout": "10",
>>>    "client_caps_release_delay": "5",
>>>    "client_oc": "true",
>>>    "client_oc_size": "209715200",
>>>    "client_oc_max_dirty": "104857600",
>>>    "client_oc_target_dirty": "8388608",
>>>    "client_oc_max_dirty_age": "5",
>>>    "client_oc_max_objects": "1000",
>>>    "client_debug_force_sync_read": "false",
>>>    "client_debug_inject_tick_delay": "0",
>>>    "fuse_use_invalidate_cb": "false",
>>>    "fuse_allow_other": "true",
>>>    "fuse_default_permissions": "true",
>>>    "fuse_big_writes": "true",
>>>    "fuse_atomic_o_trunc": "true",
>>>    "fuse_debug": "false",
>>>    "objecter_tick_interval": "5",
>>>    "objecter_timeout": "10",
>>>    "objecter_inflight_op_bytes": "104857600",
>>>    "objecter_inflight_ops": "1024",
>>>    "journaler_allow_split_entries": "true",
>>>    "journaler_write_head_interval": "15",
>>>    "journaler_prefetch_periods": "10",
>>>    "journaler_prezero_periods": "5",
>>>    "journaler_batch_interval": "0.001",
>>>    "journaler_batch_max": "0",
>>>    "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
>>>    "mds_max_file_size": "1099511627776",
>>>    "mds_cache_size": "100000",
>>>    "mds_cache_mid": "0.7",
>>>    "mds_mem_max": "1048576",
>>>    "mds_dir_commit_ratio": "0.5",
>>>    "mds_dir_max_commit_size": "90",
>>>    "mds_decay_halflife": "5",
>>>    "mds_beacon_interval": "4",
>>>    "mds_beacon_grace": "15",
>>>    "mds_enforce_unique_name": "true",
>>>    "mds_blacklist_interval": "1440",
>>>    "mds_session_timeout": "60",
>>>    "mds_session_autoclose": "300",
>>>    "mds_reconnect_timeout": "45",
>>>    "mds_tick_interval": "5",
>>>    "mds_dirstat_min_interval": "1",
>>>    "mds_scatter_nudge_interval": "5",
>>>    "mds_client_prealloc_inos": "1000",
>>>    "mds_early_reply": "true",
>>>    "mds_use_tmap": "true",
>>>    "mds_default_dir_hash": "2",
>>>    "mds_log": "true",
>>>    "mds_log_skip_corrupt_events": "false",
>>>    "mds_log_max_events": "-1",
>>>    "mds_log_segment_size": "0",
>>>    "mds_log_max_segments": "30",
>>>    "mds_log_max_expiring": "20",
>>>    "mds_bal_sample_interval": "3",
>>>    "mds_bal_replicate_threshold": "8000",
>>>    "mds_bal_unreplicate_threshold": "0",
>>>    "mds_bal_frag": "false",
>>>    "mds_bal_split_size": "10000",
>>>    "mds_bal_split_rd": "25000",
>>>    "mds_bal_split_wr": "10000",
>>>    "mds_bal_split_bits": "3",
>>>    "mds_bal_merge_size": "50",
>>>    "mds_bal_merge_rd": "1000",
>>>    "mds_bal_merge_wr": "1000",
>>>    "mds_bal_interval": "10",
>>>    "mds_bal_fragment_interval": "5",
>>>    "mds_bal_idle_threshold": "0",
>>>    "mds_bal_max": "-1",
>>>    "mds_bal_max_until": "-1",
>>>    "mds_bal_mode": "0",
>>>    "mds_bal_min_rebalance": "0.1",
>>>    "mds_bal_min_start": "0.2",
>>>    "mds_bal_need_min": "0.8",
>>>    "mds_bal_need_max": "1.2",
>>>    "mds_bal_midchunk": "0.3",
>>>    "mds_bal_minchunk": "0.001",
>>>    "mds_bal_target_removal_min": "5",
>>>    "mds_bal_target_removal_max": "10",
>>>    "mds_replay_interval": "1",
>>>    "mds_shutdown_check": "0",
>>>    "mds_thrash_exports": "0",
>>>    "mds_thrash_fragments": "0",
>>>    "mds_dump_cache_on_map": "false",
>>>    "mds_dump_cache_after_rejoin": "false",
>>>    "mds_verify_scatter": "false",
>>>    "mds_debug_scatterstat": "false",
>>>    "mds_debug_frag": "false",
>>>    "mds_debug_auth_pins": "false",
>>>    "mds_debug_subtrees": "false",
>>>    "mds_kill_mdstable_at": "0",
>>>    "mds_kill_export_at": "0",
>>>    "mds_kill_import_at": "0",
>>>    "mds_kill_link_at": "0",
>>>    "mds_kill_rename_at": "0",
>>>    "mds_kill_openc_at": "0",
>>>    "mds_kill_journal_at": "0",
>>>    "mds_kill_journal_expire_at": "0",
>>>    "mds_kill_journal_replay_at": "0",
>>>    "mds_inject_traceless_reply_probability": "0",
>>>    "mds_wipe_sessions": "false",
>>>    "mds_wipe_ino_prealloc": "false",
>>>    "mds_skip_ino": "0",
>>>    "max_mds": "1",
>>>    "mds_standby_for_name": "",
>>>    "mds_standby_for_rank": "-1",
>>>    "mds_standby_replay": "false",
>>>    "osd_auto_upgrade_tmap": "true",
>>>    "osd_tmapput_sets_uses_tmap": "false",
>>>    "osd_max_backfills": "5",
>>>    "osd_backfill_full_ratio": "0.85",
>>>    "osd_backfill_retry_interval": "10",
>>>    "osd_uuid": "00000000-0000-0000-0000-000000000000",
>>>    "osd_data": "\/ceph\/osd.0\/",
>>>    "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
>>>    "osd_journal_size": "5120",
>>>    "osd_max_write_size": "90",
>>>    "osd_max_pgls": "1024",
>>>    "osd_client_message_size_cap": "524288000",
>>>    "osd_client_message_cap": "100",
>>>    "osd_pg_bits": "6",
>>>    "osd_pgp_bits": "6",
>>>    "osd_crush_chooseleaf_type": "1",
>>>    "osd_min_rep": "1",
>>>    "osd_max_rep": "10",
>>>    "osd_pool_default_crush_rule": "0",
>>>    "osd_pool_default_size": "2",
>>>    "osd_pool_default_min_size": "0",
>>>    "osd_pool_default_pg_num": "8",
>>>    "osd_pool_default_pgp_num": "8",
>>>    "osd_pool_default_flags": "0",
>>>    "osd_map_dedup": "true",
>>>    "osd_map_cache_size": "500",
>>>    "osd_map_message_max": "100",
>>>    "osd_map_share_max_epochs": "100",
>>>    "osd_op_threads": "2",
>>>    "osd_peering_wq_batch_size": "20",
>>>    "osd_op_pq_max_tokens_per_priority": "4194304",
>>>    "osd_op_pq_min_cost": "65536",
>>>    "osd_disk_threads": "1",
>>>    "osd_recovery_threads": "2",
>>>    "osd_recover_clone_overlap": "true",
>>>    "osd_backfill_scan_min": "64",
>>>    "osd_backfill_scan_max": "512",
>>>    "osd_op_thread_timeout": "15",
>>>    "osd_recovery_thread_timeout": "30",
>>>    "osd_snap_trim_thread_timeout": "3600",
>>>    "osd_scrub_thread_timeout": "60",
>>>    "osd_scrub_finalize_thread_timeout": "600",
>>>    "osd_remove_thread_timeout": "3600",
>>>    "osd_command_thread_timeout": "600",
>>>    "osd_age": "0.8",
>>>    "osd_age_time": "0",
>>>    "osd_heartbeat_addr": ":\/0",
>>>    "osd_heartbeat_interval": "6",
>>>    "osd_heartbeat_grace": "20",
>>>    "osd_mon_heartbeat_interval": "30",
>>>    "osd_mon_report_interval_max": "120",
>>>    "osd_mon_report_interval_min": "5",
>>>    "osd_pg_stat_report_interval_max": "500",
>>>    "osd_mon_ack_timeout": "30",
>>>    "osd_min_down_reporters": "1",
>>>    "osd_min_down_reports": "3",
>>>    "osd_default_data_pool_replay_window": "45",
>>>    "osd_preserve_trimmed_log": "false",
>>>    "osd_auto_mark_unfound_lost": "false",
>>>    "osd_recovery_delay_start": "0",
>>>    "osd_recovery_max_active": "5",
>>>    "osd_recovery_max_chunk": "8388608",
>>>    "osd_recovery_forget_lost_objects": "false",
>>>    "osd_max_scrubs": "1",
>>>    "osd_scrub_load_threshold": "0.5",
>>>    "osd_scrub_min_interval": "86400",
>>>    "osd_scrub_max_interval": "604800",
>>>    "osd_deep_scrub_interval": "604800",
>>>    "osd_deep_scrub_stride": "524288",
>>>    "osd_scan_list_ping_tp_interval": "100",
>>>    "osd_auto_weight": "false",
>>>    "osd_class_dir": "\/usr\/lib\/rados-classes",
>>>    "osd_check_for_log_corruption": "false",
>>>    "osd_use_stale_snap": "false",
>>>    "osd_rollback_to_cluster_snap": "",
>>>    "osd_default_notify_timeout": "30",
>>>    "osd_kill_backfill_at": "0",
>>>    "osd_pg_epoch_persisted_max_stale": "200",
>>>    "osd_min_pg_log_entries": "500",
>>>    "osd_max_pg_log_entries": "1500",
>>>    "osd_op_complaint_time": "30",
>>>    "osd_command_max_records": "256",
>>>    "osd_op_log_threshold": "5",
>>>    "osd_verify_sparse_read_holes": "false",
>>>    "osd_debug_drop_ping_probability": "0",
>>>    "osd_debug_drop_ping_duration": "0",
>>>    "osd_debug_drop_pg_create_probability": "0",
>>>    "osd_debug_drop_pg_create_duration": "1",
>>>    "osd_debug_drop_op_probability": "0",
>>>    "osd_debug_op_order": "false",
>>>    "osd_debug_verify_snaps_on_info": "false",
>>>    "osd_debug_verify_stray_on_activate": "false",
>>>    "osd_debug_skip_full_check_in_backfill_reservation": "false",
>>>    "osd_op_history_size": "20",
>>>    "osd_op_history_duration": "600",
>>>    "osd_target_transaction_size": "30",
>>>    "osd_failsafe_full_ratio": "0.97",
>>>    "osd_failsafe_nearfull_ratio": "0.9",
>>>    "osd_leveldb_write_buffer_size": "0",
>>>    "osd_leveldb_cache_size": "0",
>>>    "osd_leveldb_block_size": "0",
>>>    "osd_leveldb_bloom_size": "0",
>>>    "osd_leveldb_max_open_files": "0",
>>>    "osd_leveldb_compression": "true",
>>>    "osd_leveldb_paranoid": "false",
>>>    "osd_leveldb_log": "",
>>>    "osd_client_op_priority": "63",
>>>    "osd_recovery_op_priority": "50",
>>>    "osd_mon_shutdown_timeout": "5",
>>>    "filestore": "false",
>>>    "filestore_index_retry_probability": "0",
>>>    "filestore_debug_inject_read_err": "false",
>>>    "filestore_debug_omap_check": "false",
>>>    "filestore_xattr_use_omap": "false",
>>>    "filestore_max_inline_xattr_size": "512",
>>>    "filestore_max_inline_xattrs": "2",
>>>    "filestore_max_sync_interval": "5",
>>>    "filestore_min_sync_interval": "0.01",
>>>    "filestore_btrfs_snap": "true",
>>>    "filestore_btrfs_clone_range": "true",
>>>    "filestore_fsync_flushes_journal_data": "false",
>>>    "filestore_fiemap": "false",
>>>    "filestore_flusher": "true",
>>>    "filestore_flusher_max_fds": "512",
>>>    "filestore_flush_min": "65536",
>>>    "filestore_sync_flush": "false",
>>>    "filestore_journal_parallel": "false",
>>>    "filestore_journal_writeahead": "false",
>>>    "filestore_journal_trailing": "false",
>>>    "filestore_queue_max_ops": "500",
>>>    "filestore_queue_max_bytes": "104857600",
>>>    "filestore_queue_committing_max_ops": "5000",
>>>    "filestore_queue_committing_max_bytes": "104857600",
>>>    "filestore_op_threads": "2",
>>>    "filestore_op_thread_timeout": "60",
>>>    "filestore_op_thread_suicide_timeout": "180",
>>>    "filestore_commit_timeout": "600",
>>>    "filestore_fiemap_threshold": "4096",
>>>    "filestore_merge_threshold": "10",
>>>    "filestore_split_multiple": "2",
>>>    "filestore_update_to": "1000",
>>>    "filestore_blackhole": "false",
>>>    "filestore_dump_file": "",
>>>    "filestore_kill_at": "0",
>>>    "filestore_inject_stall": "0",
>>>    "filestore_fail_eio": "true",
>>>    "filestore_replica_fadvise": "true",
>>>    "filestore_debug_verify_split": "false",
>>>    "journal_dio": "true",
>>>    "journal_aio": "true",
>>>    "journal_force_aio": "false",
>>>    "journal_max_corrupt_search": "10485760",
>>>    "journal_block_align": "true",
>>>    "journal_write_header_frequency": "0",
>>>    "journal_max_write_bytes": "10485760",
>>>    "journal_max_write_entries": "100",
>>>    "journal_queue_max_ops": "300",
>>>    "journal_queue_max_bytes": "33554432",
>>>    "journal_align_min_size": "65536",
>>>    "journal_replay_from": "0",
>>>    "journal_zero_on_create": "false",
>>>    "journal_ignore_corruption": "false",
>>>    "rbd_cache": "false",
>>>    "rbd_cache_writethrough_until_flush": "false",
>>>    "rbd_cache_size": "33554432",
>>>    "rbd_cache_max_dirty": "25165824",
>>>    "rbd_cache_target_dirty": "16777216",
>>>    "rbd_cache_max_dirty_age": "1",
>>>    "rbd_cache_block_writes_upfront": "false",
>>>    "rbd_concurrent_management_ops": "10",
>>>    "rbd_default_format": "1",
>>>    "rbd_default_order": "22",
>>>    "rbd_default_stripe_count": "1",
>>>    "rbd_default_stripe_unit": "4194304",
>>>    "rbd_default_features": "3",
>>>    "nss_db_path": "",
>>>    "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
>>>    "rgw_enable_apis": "s3, swift, swift_auth, admin",
>>>    "rgw_cache_enabled": "true",
>>>    "rgw_cache_lru_size": "10000",
>>>    "rgw_socket_path": "",
>>>    "rgw_host": "",
>>>    "rgw_port": "",
>>>    "rgw_dns_name": "",
>>>    "rgw_script_uri": "",
>>>    "rgw_request_uri": "",
>>>    "rgw_swift_url": "",
>>>    "rgw_swift_url_prefix": "swift",
>>>    "rgw_swift_auth_url": "",
>>>    "rgw_swift_auth_entry": "auth",
>>>    "rgw_keystone_url": "",
>>>    "rgw_keystone_admin_token": "",
>>>    "rgw_keystone_accepted_roles": "Member, admin",
>>>    "rgw_keystone_token_cache_size": "10000",
>>>    "rgw_keystone_revocation_interval": "900",
>>>    "rgw_admin_entry": "admin",
>>>    "rgw_enforce_swift_acls": "true",
>>>    "rgw_swift_token_expiration": "86400",
>>>    "rgw_print_continue": "true",
>>>    "rgw_remote_addr_param": "REMOTE_ADDR",
>>>    "rgw_op_thread_timeout": "600",
>>>    "rgw_op_thread_suicide_timeout": "0",
>>>    "rgw_thread_pool_size": "100",
>>>    "rgw_num_control_oids": "8",
>>>    "rgw_zone_root_pool": ".rgw.root",
>>>    "rgw_log_nonexistent_bucket": "false",
>>>    "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
>>>    "rgw_log_object_name_utc": "false",
>>>    "rgw_usage_max_shards": "32",
>>>    "rgw_usage_max_user_shards": "1",
>>>    "rgw_enable_ops_log": "false",
>>>    "rgw_enable_usage_log": "false",
>>>    "rgw_ops_log_rados": "true",
>>>    "rgw_ops_log_socket_path": "",
>>>    "rgw_ops_log_data_backlog": "5242880",
>>>    "rgw_usage_log_flush_threshold": "1024",
>>>    "rgw_usage_log_tick_interval": "30",
>>>    "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
>>>    "rgw_intent_log_object_name_utc": "false",
>>>    "rgw_init_timeout": "300",
>>>    "rgw_mime_types_file": "\/etc\/mime.types",
>>>    "rgw_gc_max_objs": "32",
>>>    "rgw_gc_obj_min_wait": "7200",
>>>    "rgw_gc_processor_max_time": "3600",
>>>    "rgw_gc_processor_period": "3600",
>>>    "rgw_s3_success_create_obj_status": "0",
>>>    "rgw_resolve_cname": "false",
>>>    "rgw_obj_stripe_size": "4194304",
>>>    "rgw_extended_http_attrs": "",
>>>    "rgw_exit_timeout_secs": "120",
>>>    "rgw_get_obj_window_size": "16777216",
>>>    "rgw_get_obj_max_req_size": "4194304",
>>>    "rgw_relaxed_s3_bucket_names": "false",
>>>    "rgw_list_buckets_max_chunk": "1000",
>>>    "mutex_perf_counter": "false",
>>>    "internal_safe_to_start_threads": "true"}
>>>
>>>
>>>
>>> Stefan
>>>
>>>
>>>> -Sam
>>>>
>>>> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe <s.priebe@profihost.ag>
>>>> wrote:
>>>>>
>>>>>
>>>>> Mike we already have the async patch running. Yes it helps but only
>>>>> helps
>>>>> it
>>>>> does not solve. It just hides the issue ...
>>>>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>>>>
>>>>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>>>>
>>>>>> - ceph osd set noout
>>>>>>
>>>>>> - Reboot one of the nodes hosting OSDs
>>>>>>        - VMs mounted from RBD volumes work properly
>>>>>>
>>>>>> - I see the OSD's boot messages as they re-join the cluster
>>>>>>
>>>>>> - Start seeing active+recovery_wait, peering, and active+recovering
>>>>>>        - VMs mounted from RBD volumes become unresponsive.
>>>>>>
>>>>>> - Recovery completes
>>>>>>        - VMs mounted from RBD volumes regain responsiveness
>>>>>>
>>>>>> - ceph osd unset noout
>>>>>>
>>>>>> Would joshd's async patch for qemu help here, or is there something
>>>>>> else
>>>>>> going on?
>>>>>>
>>>>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>>>>
>>>>>> Thanks,
>>>>>>
>>>>>> Mike Dawson
>>>>>> Co-Founder & Director of Cloud Architecture
>>>>>> Cloudapt LLC
>>>>>> 6330 East 75th Street, Suite 170
>>>>>> Indianapolis, IN 46250
>>>>>>
>>>>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Can you reproduce and attach the ceph.log from before you stop the
>>>>>>> osd
>>>>>>> until after you have started the osd and it has recovered?
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> i still have recovery issues with cuttlefish. After the OSD comes
>>>>>>>> back
>>>>>>>> it seem to hang for around 2-4 minutes and then recovery seems to
>>>>>>>> start
>>>>>>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7.
>>>>>>>> I
>>>>>>>> get a lot of slow request messages an hanging VMs.
>>>>>>>>
>>>>>>>> What i noticed today is that if i leave the OSD off as long as ceph
>>>>>>>> starts to backfill - the recovery and "re" backfilling wents
>>>>>>>> absolutely
>>>>>>>> smooth without any issues and no slow request messages at all.
>>>>>>>>
>>>>>>>> Does anybody have an idea why?
>>>>>>>>
>>>>>>>> Greets,
>>>>>>>> Stefan
>>>>>>>> --
>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>> ceph-devel"
>>>>>>>> in
>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>>> in
>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>
>>>>>
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>
>>>
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-02 18:21               ` Samuel Just
@ 2013-08-02 18:46                 ` Stefan Priebe
  2013-08-08 14:05                   ` Mike Dawson
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-02 18:46 UTC (permalink / raw)
  To: Samuel Just; +Cc: Mike Dawson, ceph-devel

Hi,

         osd recovery max active = 1
         osd max backfills = 1
         osd recovery op priority = 5

still no difference...

Stefan

Am 02.08.2013 20:21, schrieb Samuel Just:
> Also, you have osd_recovery_op_priority at 50.  That is close to the
> priority of client IO.  You want it below 10 (defaults to 10), perhaps
> at 1.  You can also adjust down osd_recovery_max_active.
> -Sam
>
> On Fri, Aug 2, 2013 at 11:16 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> I already tried both values this makes no difference. The drives are not the
>> bottleneck.
>>
>> Am 02.08.2013 19:35, schrieb Samuel Just:
>>
>>> You might try turning osd_max_backfills to 2 or 1.
>>> -Sam
>>>
>>> On Fri, Aug 2, 2013 at 12:44 AM, Stefan Priebe <s.priebe@profihost.ag>
>>> wrote:
>>>>
>>>> Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd
>>>> settings?
>>>>
>>>>> sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
>>>>
>>>>
>>>> Sure.
>>>>
>>>>
>>>>
>>>> { "name": "osd.0",
>>>>     "cluster": "ceph",
>>>>     "none": "0\/5",
>>>>     "lockdep": "0\/0",
>>>>     "context": "0\/0",
>>>>     "crush": "0\/0",
>>>>     "mds": "0\/0",
>>>>     "mds_balancer": "0\/0",
>>>>     "mds_locker": "0\/0",
>>>>     "mds_log": "0\/0",
>>>>     "mds_log_expire": "0\/0",
>>>>     "mds_migrator": "0\/0",
>>>>     "buffer": "0\/0",
>>>>     "timer": "0\/0",
>>>>     "filer": "0\/0",
>>>>     "striper": "0\/1",
>>>>     "objecter": "0\/0",
>>>>     "rados": "0\/0",
>>>>     "rbd": "0\/0",
>>>>     "journaler": "0\/0",
>>>>     "objectcacher": "0\/0",
>>>>     "client": "0\/0",
>>>>     "osd": "0\/0",
>>>>     "optracker": "0\/0",
>>>>     "objclass": "0\/0",
>>>>     "filestore": "0\/0",
>>>>     "journal": "0\/0",
>>>>     "ms": "0\/0",
>>>>     "mon": "0\/0",
>>>>     "monc": "0\/0",
>>>>     "paxos": "0\/0",
>>>>     "tp": "0\/0",
>>>>     "auth": "0\/0",
>>>>     "crypto": "1\/5",
>>>>     "finisher": "0\/0",
>>>>     "heartbeatmap": "0\/0",
>>>>     "perfcounter": "0\/0",
>>>>     "rgw": "0\/0",
>>>>     "hadoop": "0\/0",
>>>>     "javaclient": "1\/5",
>>>>     "asok": "0\/0",
>>>>     "throttle": "0\/0",
>>>>     "host": "cloud1-1268",
>>>>     "fsid": "00000000-0000-0000-0000-000000000000",
>>>>     "public_addr": "10.255.0.90:0\/0",
>>>>     "cluster_addr": "10.255.0.90:0\/0",
>>>>     "public_network": "10.255.0.1\/24",
>>>>     "cluster_network": "10.255.0.1\/24",
>>>>     "num_client": "1",
>>>>     "monmap": "",
>>>>     "mon_host": "",
>>>>     "lockdep": "false",
>>>>     "run_dir": "\/var\/run\/ceph",
>>>>     "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
>>>>     "daemonize": "true",
>>>>     "pid_file": "\/var\/run\/ceph\/osd.0.pid",
>>>>     "chdir": "\/",
>>>>     "max_open_files": "0",
>>>>     "fatal_signal_handlers": "true",
>>>>     "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
>>>>     "log_max_new": "1000",
>>>>     "log_max_recent": "10000",
>>>>     "log_to_stderr": "false",
>>>>     "err_to_stderr": "true",
>>>>     "log_to_syslog": "false",
>>>>     "err_to_syslog": "false",
>>>>     "log_flush_on_exit": "true",
>>>>     "log_stop_at_utilization": "0.97",
>>>>     "clog_to_monitors": "true",
>>>>     "clog_to_syslog": "false",
>>>>     "clog_to_syslog_level": "info",
>>>>     "clog_to_syslog_facility": "daemon",
>>>>     "mon_cluster_log_to_syslog": "false",
>>>>     "mon_cluster_log_to_syslog_level": "info",
>>>>     "mon_cluster_log_to_syslog_facility": "daemon",
>>>>     "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
>>>>     "key": "",
>>>>     "keyfile": "",
>>>>     "keyring": "\/etc\/ceph\/osd.0.keyring",
>>>>     "heartbeat_interval": "5",
>>>>     "heartbeat_file": "",
>>>>     "heartbeat_inject_failure": "0",
>>>>     "perf": "true",
>>>>     "ms_tcp_nodelay": "true",
>>>>     "ms_tcp_rcvbuf": "0",
>>>>     "ms_initial_backoff": "0.2",
>>>>     "ms_max_backoff": "15",
>>>>     "ms_nocrc": "false",
>>>>     "ms_die_on_bad_msg": "false",
>>>>     "ms_die_on_unhandled_msg": "false",
>>>>     "ms_dispatch_throttle_bytes": "104857600",
>>>>     "ms_bind_ipv6": "false",
>>>>     "ms_bind_port_min": "6800",
>>>>     "ms_bind_port_max": "7100",
>>>>     "ms_rwthread_stack_bytes": "1048576",
>>>>     "ms_tcp_read_timeout": "900",
>>>>     "ms_pq_max_tokens_per_priority": "4194304",
>>>>     "ms_pq_min_cost": "65536",
>>>>     "ms_inject_socket_failures": "0",
>>>>     "ms_inject_delay_type": "",
>>>>     "ms_inject_delay_max": "1",
>>>>     "ms_inject_delay_probability": "0",
>>>>     "ms_inject_internal_delays": "0",
>>>>     "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
>>>>     "mon_initial_members": "",
>>>>     "mon_sync_fs_threshold": "5",
>>>>     "mon_compact_on_start": "false",
>>>>     "mon_compact_on_bootstrap": "false",
>>>>     "mon_compact_on_trim": "true",
>>>>     "mon_tick_interval": "5",
>>>>     "mon_subscribe_interval": "300",
>>>>     "mon_osd_laggy_halflife": "3600",
>>>>     "mon_osd_laggy_weight": "0.3",
>>>>     "mon_osd_adjust_heartbeat_grace": "true",
>>>>     "mon_osd_adjust_down_out_interval": "true",
>>>>     "mon_osd_auto_mark_in": "false",
>>>>     "mon_osd_auto_mark_auto_out_in": "true",
>>>>     "mon_osd_auto_mark_new_in": "true",
>>>>     "mon_osd_down_out_interval": "300",
>>>>     "mon_osd_down_out_subtree_limit": "rack",
>>>>     "mon_osd_min_up_ratio": "0.3",
>>>>     "mon_osd_min_in_ratio": "0.3",
>>>>     "mon_stat_smooth_intervals": "2",
>>>>     "mon_lease": "5",
>>>>     "mon_lease_renew_interval": "3",
>>>>     "mon_lease_ack_timeout": "10",
>>>>     "mon_clock_drift_allowed": "0.05",
>>>>     "mon_clock_drift_warn_backoff": "5",
>>>>     "mon_timecheck_interval": "300",
>>>>     "mon_accept_timeout": "10",
>>>>     "mon_pg_create_interval": "30",
>>>>     "mon_pg_stuck_threshold": "300",
>>>>     "mon_osd_full_ratio": "0.95",
>>>>     "mon_osd_nearfull_ratio": "0.85",
>>>>     "mon_globalid_prealloc": "100",
>>>>     "mon_osd_report_timeout": "900",
>>>>     "mon_force_standby_active": "true",
>>>>     "mon_min_osdmap_epochs": "500",
>>>>     "mon_max_pgmap_epochs": "500",
>>>>     "mon_max_log_epochs": "500",
>>>>     "mon_max_osd": "10000",
>>>>     "mon_probe_timeout": "2",
>>>>     "mon_slurp_timeout": "10",
>>>>     "mon_slurp_bytes": "262144",
>>>>     "mon_client_bytes": "104857600",
>>>>     "mon_daemon_bytes": "419430400",
>>>>     "mon_max_log_entries_per_event": "4096",
>>>>     "mon_health_data_update_interval": "60",
>>>>     "mon_data_avail_crit": "5",
>>>>     "mon_data_avail_warn": "30",
>>>>     "mon_config_key_max_entry_size": "4096",
>>>>     "mon_sync_trim_timeout": "30",
>>>>     "mon_sync_heartbeat_timeout": "30",
>>>>     "mon_sync_heartbeat_interval": "5",
>>>>     "mon_sync_backoff_timeout": "30",
>>>>     "mon_sync_timeout": "30",
>>>>     "mon_sync_max_retries": "5",
>>>>     "mon_sync_max_payload_size": "1048576",
>>>>     "mon_sync_debug": "false",
>>>>     "mon_sync_debug_leader": "-1",
>>>>     "mon_sync_debug_provider": "-1",
>>>>     "mon_sync_debug_provider_fallback": "-1",
>>>>     "mon_debug_dump_transactions": "false",
>>>>     "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
>>>>     "mon_sync_leader_kill_at": "0",
>>>>     "mon_sync_provider_kill_at": "0",
>>>>     "mon_sync_requester_kill_at": "0",
>>>>     "mon_leveldb_write_buffer_size": "33554432",
>>>>     "mon_leveldb_cache_size": "268435456",
>>>>     "mon_leveldb_block_size": "65536",
>>>>     "mon_leveldb_bloom_size": "0",
>>>>     "mon_leveldb_max_open_files": "0",
>>>>     "mon_leveldb_compression": "false",
>>>>     "mon_leveldb_paranoid": "false",
>>>>     "mon_leveldb_log": "",
>>>>     "paxos_stash_full_interval": "25",
>>>>     "paxos_max_join_drift": "100",
>>>>     "paxos_propose_interval": "1",
>>>>     "paxos_min_wait": "0.05",
>>>>     "paxos_min": "500",
>>>>     "paxos_trim_min": "500",
>>>>     "paxos_trim_max": "1000",
>>>>     "paxos_trim_disabled_max_versions": "108000",
>>>>     "paxos_service_trim_min": "500",
>>>>     "paxos_service_trim_max": "1000",
>>>>     "clock_offset": "0",
>>>>     "auth_cluster_required": "none",
>>>>     "auth_service_required": "none",
>>>>     "auth_client_required": "none",
>>>>     "auth_supported": "none",
>>>>     "cephx_require_signatures": "false",
>>>>     "cephx_cluster_require_signatures": "false",
>>>>     "cephx_service_require_signatures": "false",
>>>>     "cephx_sign_messages": "true",
>>>>     "auth_mon_ticket_ttl": "43200",
>>>>     "auth_service_ticket_ttl": "3600",
>>>>     "auth_debug": "false",
>>>>     "mon_client_hunt_interval": "3",
>>>>     "mon_client_ping_interval": "10",
>>>>     "mon_client_max_log_entries_per_message": "1000",
>>>>     "mon_max_pool_pg_num": "65536",
>>>>     "mon_pool_quota_warn_threshold": "0",
>>>>     "mon_pool_quota_crit_threshold": "0",
>>>>     "client_cache_size": "16384",
>>>>     "client_cache_mid": "0.75",
>>>>     "client_use_random_mds": "false",
>>>>     "client_mount_timeout": "300",
>>>>     "client_tick_interval": "1",
>>>>     "client_trace": "",
>>>>     "client_readahead_min": "131072",
>>>>     "client_readahead_max_bytes": "0",
>>>>     "client_readahead_max_periods": "4",
>>>>     "client_snapdir": ".snap",
>>>>     "client_mountpoint": "\/",
>>>>     "client_notify_timeout": "10",
>>>>     "client_caps_release_delay": "5",
>>>>     "client_oc": "true",
>>>>     "client_oc_size": "209715200",
>>>>     "client_oc_max_dirty": "104857600",
>>>>     "client_oc_target_dirty": "8388608",
>>>>     "client_oc_max_dirty_age": "5",
>>>>     "client_oc_max_objects": "1000",
>>>>     "client_debug_force_sync_read": "false",
>>>>     "client_debug_inject_tick_delay": "0",
>>>>     "fuse_use_invalidate_cb": "false",
>>>>     "fuse_allow_other": "true",
>>>>     "fuse_default_permissions": "true",
>>>>     "fuse_big_writes": "true",
>>>>     "fuse_atomic_o_trunc": "true",
>>>>     "fuse_debug": "false",
>>>>     "objecter_tick_interval": "5",
>>>>     "objecter_timeout": "10",
>>>>     "objecter_inflight_op_bytes": "104857600",
>>>>     "objecter_inflight_ops": "1024",
>>>>     "journaler_allow_split_entries": "true",
>>>>     "journaler_write_head_interval": "15",
>>>>     "journaler_prefetch_periods": "10",
>>>>     "journaler_prezero_periods": "5",
>>>>     "journaler_batch_interval": "0.001",
>>>>     "journaler_batch_max": "0",
>>>>     "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
>>>>     "mds_max_file_size": "1099511627776",
>>>>     "mds_cache_size": "100000",
>>>>     "mds_cache_mid": "0.7",
>>>>     "mds_mem_max": "1048576",
>>>>     "mds_dir_commit_ratio": "0.5",
>>>>     "mds_dir_max_commit_size": "90",
>>>>     "mds_decay_halflife": "5",
>>>>     "mds_beacon_interval": "4",
>>>>     "mds_beacon_grace": "15",
>>>>     "mds_enforce_unique_name": "true",
>>>>     "mds_blacklist_interval": "1440",
>>>>     "mds_session_timeout": "60",
>>>>     "mds_session_autoclose": "300",
>>>>     "mds_reconnect_timeout": "45",
>>>>     "mds_tick_interval": "5",
>>>>     "mds_dirstat_min_interval": "1",
>>>>     "mds_scatter_nudge_interval": "5",
>>>>     "mds_client_prealloc_inos": "1000",
>>>>     "mds_early_reply": "true",
>>>>     "mds_use_tmap": "true",
>>>>     "mds_default_dir_hash": "2",
>>>>     "mds_log": "true",
>>>>     "mds_log_skip_corrupt_events": "false",
>>>>     "mds_log_max_events": "-1",
>>>>     "mds_log_segment_size": "0",
>>>>     "mds_log_max_segments": "30",
>>>>     "mds_log_max_expiring": "20",
>>>>     "mds_bal_sample_interval": "3",
>>>>     "mds_bal_replicate_threshold": "8000",
>>>>     "mds_bal_unreplicate_threshold": "0",
>>>>     "mds_bal_frag": "false",
>>>>     "mds_bal_split_size": "10000",
>>>>     "mds_bal_split_rd": "25000",
>>>>     "mds_bal_split_wr": "10000",
>>>>     "mds_bal_split_bits": "3",
>>>>     "mds_bal_merge_size": "50",
>>>>     "mds_bal_merge_rd": "1000",
>>>>     "mds_bal_merge_wr": "1000",
>>>>     "mds_bal_interval": "10",
>>>>     "mds_bal_fragment_interval": "5",
>>>>     "mds_bal_idle_threshold": "0",
>>>>     "mds_bal_max": "-1",
>>>>     "mds_bal_max_until": "-1",
>>>>     "mds_bal_mode": "0",
>>>>     "mds_bal_min_rebalance": "0.1",
>>>>     "mds_bal_min_start": "0.2",
>>>>     "mds_bal_need_min": "0.8",
>>>>     "mds_bal_need_max": "1.2",
>>>>     "mds_bal_midchunk": "0.3",
>>>>     "mds_bal_minchunk": "0.001",
>>>>     "mds_bal_target_removal_min": "5",
>>>>     "mds_bal_target_removal_max": "10",
>>>>     "mds_replay_interval": "1",
>>>>     "mds_shutdown_check": "0",
>>>>     "mds_thrash_exports": "0",
>>>>     "mds_thrash_fragments": "0",
>>>>     "mds_dump_cache_on_map": "false",
>>>>     "mds_dump_cache_after_rejoin": "false",
>>>>     "mds_verify_scatter": "false",
>>>>     "mds_debug_scatterstat": "false",
>>>>     "mds_debug_frag": "false",
>>>>     "mds_debug_auth_pins": "false",
>>>>     "mds_debug_subtrees": "false",
>>>>     "mds_kill_mdstable_at": "0",
>>>>     "mds_kill_export_at": "0",
>>>>     "mds_kill_import_at": "0",
>>>>     "mds_kill_link_at": "0",
>>>>     "mds_kill_rename_at": "0",
>>>>     "mds_kill_openc_at": "0",
>>>>     "mds_kill_journal_at": "0",
>>>>     "mds_kill_journal_expire_at": "0",
>>>>     "mds_kill_journal_replay_at": "0",
>>>>     "mds_inject_traceless_reply_probability": "0",
>>>>     "mds_wipe_sessions": "false",
>>>>     "mds_wipe_ino_prealloc": "false",
>>>>     "mds_skip_ino": "0",
>>>>     "max_mds": "1",
>>>>     "mds_standby_for_name": "",
>>>>     "mds_standby_for_rank": "-1",
>>>>     "mds_standby_replay": "false",
>>>>     "osd_auto_upgrade_tmap": "true",
>>>>     "osd_tmapput_sets_uses_tmap": "false",
>>>>     "osd_max_backfills": "5",
>>>>     "osd_backfill_full_ratio": "0.85",
>>>>     "osd_backfill_retry_interval": "10",
>>>>     "osd_uuid": "00000000-0000-0000-0000-000000000000",
>>>>     "osd_data": "\/ceph\/osd.0\/",
>>>>     "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
>>>>     "osd_journal_size": "5120",
>>>>     "osd_max_write_size": "90",
>>>>     "osd_max_pgls": "1024",
>>>>     "osd_client_message_size_cap": "524288000",
>>>>     "osd_client_message_cap": "100",
>>>>     "osd_pg_bits": "6",
>>>>     "osd_pgp_bits": "6",
>>>>     "osd_crush_chooseleaf_type": "1",
>>>>     "osd_min_rep": "1",
>>>>     "osd_max_rep": "10",
>>>>     "osd_pool_default_crush_rule": "0",
>>>>     "osd_pool_default_size": "2",
>>>>     "osd_pool_default_min_size": "0",
>>>>     "osd_pool_default_pg_num": "8",
>>>>     "osd_pool_default_pgp_num": "8",
>>>>     "osd_pool_default_flags": "0",
>>>>     "osd_map_dedup": "true",
>>>>     "osd_map_cache_size": "500",
>>>>     "osd_map_message_max": "100",
>>>>     "osd_map_share_max_epochs": "100",
>>>>     "osd_op_threads": "2",
>>>>     "osd_peering_wq_batch_size": "20",
>>>>     "osd_op_pq_max_tokens_per_priority": "4194304",
>>>>     "osd_op_pq_min_cost": "65536",
>>>>     "osd_disk_threads": "1",
>>>>     "osd_recovery_threads": "2",
>>>>     "osd_recover_clone_overlap": "true",
>>>>     "osd_backfill_scan_min": "64",
>>>>     "osd_backfill_scan_max": "512",
>>>>     "osd_op_thread_timeout": "15",
>>>>     "osd_recovery_thread_timeout": "30",
>>>>     "osd_snap_trim_thread_timeout": "3600",
>>>>     "osd_scrub_thread_timeout": "60",
>>>>     "osd_scrub_finalize_thread_timeout": "600",
>>>>     "osd_remove_thread_timeout": "3600",
>>>>     "osd_command_thread_timeout": "600",
>>>>     "osd_age": "0.8",
>>>>     "osd_age_time": "0",
>>>>     "osd_heartbeat_addr": ":\/0",
>>>>     "osd_heartbeat_interval": "6",
>>>>     "osd_heartbeat_grace": "20",
>>>>     "osd_mon_heartbeat_interval": "30",
>>>>     "osd_mon_report_interval_max": "120",
>>>>     "osd_mon_report_interval_min": "5",
>>>>     "osd_pg_stat_report_interval_max": "500",
>>>>     "osd_mon_ack_timeout": "30",
>>>>     "osd_min_down_reporters": "1",
>>>>     "osd_min_down_reports": "3",
>>>>     "osd_default_data_pool_replay_window": "45",
>>>>     "osd_preserve_trimmed_log": "false",
>>>>     "osd_auto_mark_unfound_lost": "false",
>>>>     "osd_recovery_delay_start": "0",
>>>>     "osd_recovery_max_active": "5",
>>>>     "osd_recovery_max_chunk": "8388608",
>>>>     "osd_recovery_forget_lost_objects": "false",
>>>>     "osd_max_scrubs": "1",
>>>>     "osd_scrub_load_threshold": "0.5",
>>>>     "osd_scrub_min_interval": "86400",
>>>>     "osd_scrub_max_interval": "604800",
>>>>     "osd_deep_scrub_interval": "604800",
>>>>     "osd_deep_scrub_stride": "524288",
>>>>     "osd_scan_list_ping_tp_interval": "100",
>>>>     "osd_auto_weight": "false",
>>>>     "osd_class_dir": "\/usr\/lib\/rados-classes",
>>>>     "osd_check_for_log_corruption": "false",
>>>>     "osd_use_stale_snap": "false",
>>>>     "osd_rollback_to_cluster_snap": "",
>>>>     "osd_default_notify_timeout": "30",
>>>>     "osd_kill_backfill_at": "0",
>>>>     "osd_pg_epoch_persisted_max_stale": "200",
>>>>     "osd_min_pg_log_entries": "500",
>>>>     "osd_max_pg_log_entries": "1500",
>>>>     "osd_op_complaint_time": "30",
>>>>     "osd_command_max_records": "256",
>>>>     "osd_op_log_threshold": "5",
>>>>     "osd_verify_sparse_read_holes": "false",
>>>>     "osd_debug_drop_ping_probability": "0",
>>>>     "osd_debug_drop_ping_duration": "0",
>>>>     "osd_debug_drop_pg_create_probability": "0",
>>>>     "osd_debug_drop_pg_create_duration": "1",
>>>>     "osd_debug_drop_op_probability": "0",
>>>>     "osd_debug_op_order": "false",
>>>>     "osd_debug_verify_snaps_on_info": "false",
>>>>     "osd_debug_verify_stray_on_activate": "false",
>>>>     "osd_debug_skip_full_check_in_backfill_reservation": "false",
>>>>     "osd_op_history_size": "20",
>>>>     "osd_op_history_duration": "600",
>>>>     "osd_target_transaction_size": "30",
>>>>     "osd_failsafe_full_ratio": "0.97",
>>>>     "osd_failsafe_nearfull_ratio": "0.9",
>>>>     "osd_leveldb_write_buffer_size": "0",
>>>>     "osd_leveldb_cache_size": "0",
>>>>     "osd_leveldb_block_size": "0",
>>>>     "osd_leveldb_bloom_size": "0",
>>>>     "osd_leveldb_max_open_files": "0",
>>>>     "osd_leveldb_compression": "true",
>>>>     "osd_leveldb_paranoid": "false",
>>>>     "osd_leveldb_log": "",
>>>>     "osd_client_op_priority": "63",
>>>>     "osd_recovery_op_priority": "50",
>>>>     "osd_mon_shutdown_timeout": "5",
>>>>     "filestore": "false",
>>>>     "filestore_index_retry_probability": "0",
>>>>     "filestore_debug_inject_read_err": "false",
>>>>     "filestore_debug_omap_check": "false",
>>>>     "filestore_xattr_use_omap": "false",
>>>>     "filestore_max_inline_xattr_size": "512",
>>>>     "filestore_max_inline_xattrs": "2",
>>>>     "filestore_max_sync_interval": "5",
>>>>     "filestore_min_sync_interval": "0.01",
>>>>     "filestore_btrfs_snap": "true",
>>>>     "filestore_btrfs_clone_range": "true",
>>>>     "filestore_fsync_flushes_journal_data": "false",
>>>>     "filestore_fiemap": "false",
>>>>     "filestore_flusher": "true",
>>>>     "filestore_flusher_max_fds": "512",
>>>>     "filestore_flush_min": "65536",
>>>>     "filestore_sync_flush": "false",
>>>>     "filestore_journal_parallel": "false",
>>>>     "filestore_journal_writeahead": "false",
>>>>     "filestore_journal_trailing": "false",
>>>>     "filestore_queue_max_ops": "500",
>>>>     "filestore_queue_max_bytes": "104857600",
>>>>     "filestore_queue_committing_max_ops": "5000",
>>>>     "filestore_queue_committing_max_bytes": "104857600",
>>>>     "filestore_op_threads": "2",
>>>>     "filestore_op_thread_timeout": "60",
>>>>     "filestore_op_thread_suicide_timeout": "180",
>>>>     "filestore_commit_timeout": "600",
>>>>     "filestore_fiemap_threshold": "4096",
>>>>     "filestore_merge_threshold": "10",
>>>>     "filestore_split_multiple": "2",
>>>>     "filestore_update_to": "1000",
>>>>     "filestore_blackhole": "false",
>>>>     "filestore_dump_file": "",
>>>>     "filestore_kill_at": "0",
>>>>     "filestore_inject_stall": "0",
>>>>     "filestore_fail_eio": "true",
>>>>     "filestore_replica_fadvise": "true",
>>>>     "filestore_debug_verify_split": "false",
>>>>     "journal_dio": "true",
>>>>     "journal_aio": "true",
>>>>     "journal_force_aio": "false",
>>>>     "journal_max_corrupt_search": "10485760",
>>>>     "journal_block_align": "true",
>>>>     "journal_write_header_frequency": "0",
>>>>     "journal_max_write_bytes": "10485760",
>>>>     "journal_max_write_entries": "100",
>>>>     "journal_queue_max_ops": "300",
>>>>     "journal_queue_max_bytes": "33554432",
>>>>     "journal_align_min_size": "65536",
>>>>     "journal_replay_from": "0",
>>>>     "journal_zero_on_create": "false",
>>>>     "journal_ignore_corruption": "false",
>>>>     "rbd_cache": "false",
>>>>     "rbd_cache_writethrough_until_flush": "false",
>>>>     "rbd_cache_size": "33554432",
>>>>     "rbd_cache_max_dirty": "25165824",
>>>>     "rbd_cache_target_dirty": "16777216",
>>>>     "rbd_cache_max_dirty_age": "1",
>>>>     "rbd_cache_block_writes_upfront": "false",
>>>>     "rbd_concurrent_management_ops": "10",
>>>>     "rbd_default_format": "1",
>>>>     "rbd_default_order": "22",
>>>>     "rbd_default_stripe_count": "1",
>>>>     "rbd_default_stripe_unit": "4194304",
>>>>     "rbd_default_features": "3",
>>>>     "nss_db_path": "",
>>>>     "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
>>>>     "rgw_enable_apis": "s3, swift, swift_auth, admin",
>>>>     "rgw_cache_enabled": "true",
>>>>     "rgw_cache_lru_size": "10000",
>>>>     "rgw_socket_path": "",
>>>>     "rgw_host": "",
>>>>     "rgw_port": "",
>>>>     "rgw_dns_name": "",
>>>>     "rgw_script_uri": "",
>>>>     "rgw_request_uri": "",
>>>>     "rgw_swift_url": "",
>>>>     "rgw_swift_url_prefix": "swift",
>>>>     "rgw_swift_auth_url": "",
>>>>     "rgw_swift_auth_entry": "auth",
>>>>     "rgw_keystone_url": "",
>>>>     "rgw_keystone_admin_token": "",
>>>>     "rgw_keystone_accepted_roles": "Member, admin",
>>>>     "rgw_keystone_token_cache_size": "10000",
>>>>     "rgw_keystone_revocation_interval": "900",
>>>>     "rgw_admin_entry": "admin",
>>>>     "rgw_enforce_swift_acls": "true",
>>>>     "rgw_swift_token_expiration": "86400",
>>>>     "rgw_print_continue": "true",
>>>>     "rgw_remote_addr_param": "REMOTE_ADDR",
>>>>     "rgw_op_thread_timeout": "600",
>>>>     "rgw_op_thread_suicide_timeout": "0",
>>>>     "rgw_thread_pool_size": "100",
>>>>     "rgw_num_control_oids": "8",
>>>>     "rgw_zone_root_pool": ".rgw.root",
>>>>     "rgw_log_nonexistent_bucket": "false",
>>>>     "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
>>>>     "rgw_log_object_name_utc": "false",
>>>>     "rgw_usage_max_shards": "32",
>>>>     "rgw_usage_max_user_shards": "1",
>>>>     "rgw_enable_ops_log": "false",
>>>>     "rgw_enable_usage_log": "false",
>>>>     "rgw_ops_log_rados": "true",
>>>>     "rgw_ops_log_socket_path": "",
>>>>     "rgw_ops_log_data_backlog": "5242880",
>>>>     "rgw_usage_log_flush_threshold": "1024",
>>>>     "rgw_usage_log_tick_interval": "30",
>>>>     "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
>>>>     "rgw_intent_log_object_name_utc": "false",
>>>>     "rgw_init_timeout": "300",
>>>>     "rgw_mime_types_file": "\/etc\/mime.types",
>>>>     "rgw_gc_max_objs": "32",
>>>>     "rgw_gc_obj_min_wait": "7200",
>>>>     "rgw_gc_processor_max_time": "3600",
>>>>     "rgw_gc_processor_period": "3600",
>>>>     "rgw_s3_success_create_obj_status": "0",
>>>>     "rgw_resolve_cname": "false",
>>>>     "rgw_obj_stripe_size": "4194304",
>>>>     "rgw_extended_http_attrs": "",
>>>>     "rgw_exit_timeout_secs": "120",
>>>>     "rgw_get_obj_window_size": "16777216",
>>>>     "rgw_get_obj_max_req_size": "4194304",
>>>>     "rgw_relaxed_s3_bucket_names": "false",
>>>>     "rgw_list_buckets_max_chunk": "1000",
>>>>     "mutex_perf_counter": "false",
>>>>     "internal_safe_to_start_threads": "true"}
>>>>
>>>>
>>>>
>>>> Stefan
>>>>
>>>>
>>>>> -Sam
>>>>>
>>>>> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe <s.priebe@profihost.ag>
>>>>> wrote:
>>>>>>
>>>>>>
>>>>>> Mike we already have the async patch running. Yes it helps but only
>>>>>> helps
>>>>>> it
>>>>>> does not solve. It just hides the issue ...
>>>>>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>>>>>
>>>>>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>>>>>
>>>>>>> - ceph osd set noout
>>>>>>>
>>>>>>> - Reboot one of the nodes hosting OSDs
>>>>>>>         - VMs mounted from RBD volumes work properly
>>>>>>>
>>>>>>> - I see the OSD's boot messages as they re-join the cluster
>>>>>>>
>>>>>>> - Start seeing active+recovery_wait, peering, and active+recovering
>>>>>>>         - VMs mounted from RBD volumes become unresponsive.
>>>>>>>
>>>>>>> - Recovery completes
>>>>>>>         - VMs mounted from RBD volumes regain responsiveness
>>>>>>>
>>>>>>> - ceph osd unset noout
>>>>>>>
>>>>>>> Would joshd's async patch for qemu help here, or is there something
>>>>>>> else
>>>>>>> going on?
>>>>>>>
>>>>>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>>>>>
>>>>>>> Thanks,
>>>>>>>
>>>>>>> Mike Dawson
>>>>>>> Co-Founder & Director of Cloud Architecture
>>>>>>> Cloudapt LLC
>>>>>>> 6330 East 75th Street, Suite 170
>>>>>>> Indianapolis, IN 46250
>>>>>>>
>>>>>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Can you reproduce and attach the ceph.log from before you stop the
>>>>>>>> osd
>>>>>>>> until after you have started the osd and it has recovered?
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> i still have recovery issues with cuttlefish. After the OSD comes
>>>>>>>>> back
>>>>>>>>> it seem to hang for around 2-4 minutes and then recovery seems to
>>>>>>>>> start
>>>>>>>>> (pgs in recovery_wait start to decrement). This is with ceph 0.61.7.
>>>>>>>>> I
>>>>>>>>> get a lot of slow request messages an hanging VMs.
>>>>>>>>>
>>>>>>>>> What i noticed today is that if i leave the OSD off as long as ceph
>>>>>>>>> starts to backfill - the recovery and "re" backfilling wents
>>>>>>>>> absolutely
>>>>>>>>> smooth without any issues and no slow request messages at all.
>>>>>>>>>
>>>>>>>>> Does anybody have an idea why?
>>>>>>>>>
>>>>>>>>> Greets,
>>>>>>>>> Stefan
>>>>>>>>> --
>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>> ceph-devel"
>>>>>>>>> in
>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>>>> in
>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>>
>>>>>>
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>>> the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>>
>>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-02 18:46                 ` Stefan Priebe
@ 2013-08-08 14:05                   ` Mike Dawson
  2013-08-08 15:43                     ` Oliver Francke
  2013-08-08 18:13                     ` Stefan Priebe
  0 siblings, 2 replies; 40+ messages in thread
From: Mike Dawson @ 2013-08-08 14:05 UTC (permalink / raw)
  To: Stefan Priebe, Samuel Just, josh.durgin, Oliver Francke
  Cc: ceph-devel, Stefan Hajnoczi

Stefan,

I see the same behavior and I theorize it is linked to an issue detailed 
in another thread [0]. Do your VM guests ever hang while your cluster is 
HEALTH_OK like described in that other thread?

[0] http://comments.gmane.org/gmane.comp.file-systems.ceph.user/2982

A few observations:

- The VMs that hang do lots of writes (video surveillance).
- I use rbd and qemu. The problem exists in both qemu 1.4.x and 1.5.2.
- The problem exists with or without joshd's qemu async flush patch.
- Windows VMs seem to be more vulnerable than Linux VMs.
- If I restart the qemu-system-x86_64 process, the guest will come back 
to life.
- A partial workaround seems to be console input (NoVNC or 'virsh 
screenshot'), but restarting qemu-system-x86_64 works better.
- The issue of VMs hanging seems worse with RBD writeback cache enabled
- I typically run virtio, but I believe I've seen it with e1000, too.
- VM guests hang at different times, not all at once on a host (or 
across all hosts).
- I co-mingle VM guests on servers that host ceph OSDs.



Oliver,

If your cluster has to recover/backfill, do your guest VMs hang with 
more frequency than under normal HEALTH_OK conditions, even if you 
prioritize client IO as Sam wrote below?


Sam,

Turning down all the settings you mentioned certainly does slow the 
recover/backfill process, but it doesn't prevent the VM guests backed by 
RBD volumes from hanging. In fact, I often try to prioritize 
recovery/backfill because my guests tend to hang until I get back to 
HEALTH_OK. Given this apparent bug, completing recovery/backfill quicker 
leads to less total outage, it seems.


Josh,

How can I help you investigate if RBD is the common source of both of 
these issues?


Thanks,
Mike Dawson


On 8/2/2013 2:46 PM, Stefan Priebe wrote:
> Hi,
>
>          osd recovery max active = 1
>          osd max backfills = 1
>          osd recovery op priority = 5
>
> still no difference...
>
> Stefan
>
> Am 02.08.2013 20:21, schrieb Samuel Just:
>> Also, you have osd_recovery_op_priority at 50.  That is close to the
>> priority of client IO.  You want it below 10 (defaults to 10), perhaps
>> at 1.  You can also adjust down osd_recovery_max_active.
>> -Sam
>>
>> On Fri, Aug 2, 2013 at 11:16 AM, Stefan Priebe <s.priebe@profihost.ag>
>> wrote:
>>> I already tried both values this makes no difference. The drives are
>>> not the
>>> bottleneck.
>>>
>>> Am 02.08.2013 19:35, schrieb Samuel Just:
>>>
>>>> You might try turning osd_max_backfills to 2 or 1.
>>>> -Sam
>>>>
>>>> On Fri, Aug 2, 2013 at 12:44 AM, Stefan Priebe <s.priebe@profihost.ag>
>>>> wrote:
>>>>>
>>>>> Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd
>>>>> settings?
>>>>>
>>>>>> sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
>>>>>
>>>>>
>>>>> Sure.
>>>>>
>>>>>
>>>>>
>>>>> { "name": "osd.0",
>>>>>     "cluster": "ceph",
>>>>>     "none": "0\/5",
>>>>>     "lockdep": "0\/0",
>>>>>     "context": "0\/0",
>>>>>     "crush": "0\/0",
>>>>>     "mds": "0\/0",
>>>>>     "mds_balancer": "0\/0",
>>>>>     "mds_locker": "0\/0",
>>>>>     "mds_log": "0\/0",
>>>>>     "mds_log_expire": "0\/0",
>>>>>     "mds_migrator": "0\/0",
>>>>>     "buffer": "0\/0",
>>>>>     "timer": "0\/0",
>>>>>     "filer": "0\/0",
>>>>>     "striper": "0\/1",
>>>>>     "objecter": "0\/0",
>>>>>     "rados": "0\/0",
>>>>>     "rbd": "0\/0",
>>>>>     "journaler": "0\/0",
>>>>>     "objectcacher": "0\/0",
>>>>>     "client": "0\/0",
>>>>>     "osd": "0\/0",
>>>>>     "optracker": "0\/0",
>>>>>     "objclass": "0\/0",
>>>>>     "filestore": "0\/0",
>>>>>     "journal": "0\/0",
>>>>>     "ms": "0\/0",
>>>>>     "mon": "0\/0",
>>>>>     "monc": "0\/0",
>>>>>     "paxos": "0\/0",
>>>>>     "tp": "0\/0",
>>>>>     "auth": "0\/0",
>>>>>     "crypto": "1\/5",
>>>>>     "finisher": "0\/0",
>>>>>     "heartbeatmap": "0\/0",
>>>>>     "perfcounter": "0\/0",
>>>>>     "rgw": "0\/0",
>>>>>     "hadoop": "0\/0",
>>>>>     "javaclient": "1\/5",
>>>>>     "asok": "0\/0",
>>>>>     "throttle": "0\/0",
>>>>>     "host": "cloud1-1268",
>>>>>     "fsid": "00000000-0000-0000-0000-000000000000",
>>>>>     "public_addr": "10.255.0.90:0\/0",
>>>>>     "cluster_addr": "10.255.0.90:0\/0",
>>>>>     "public_network": "10.255.0.1\/24",
>>>>>     "cluster_network": "10.255.0.1\/24",
>>>>>     "num_client": "1",
>>>>>     "monmap": "",
>>>>>     "mon_host": "",
>>>>>     "lockdep": "false",
>>>>>     "run_dir": "\/var\/run\/ceph",
>>>>>     "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
>>>>>     "daemonize": "true",
>>>>>     "pid_file": "\/var\/run\/ceph\/osd.0.pid",
>>>>>     "chdir": "\/",
>>>>>     "max_open_files": "0",
>>>>>     "fatal_signal_handlers": "true",
>>>>>     "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
>>>>>     "log_max_new": "1000",
>>>>>     "log_max_recent": "10000",
>>>>>     "log_to_stderr": "false",
>>>>>     "err_to_stderr": "true",
>>>>>     "log_to_syslog": "false",
>>>>>     "err_to_syslog": "false",
>>>>>     "log_flush_on_exit": "true",
>>>>>     "log_stop_at_utilization": "0.97",
>>>>>     "clog_to_monitors": "true",
>>>>>     "clog_to_syslog": "false",
>>>>>     "clog_to_syslog_level": "info",
>>>>>     "clog_to_syslog_facility": "daemon",
>>>>>     "mon_cluster_log_to_syslog": "false",
>>>>>     "mon_cluster_log_to_syslog_level": "info",
>>>>>     "mon_cluster_log_to_syslog_facility": "daemon",
>>>>>     "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
>>>>>     "key": "",
>>>>>     "keyfile": "",
>>>>>     "keyring": "\/etc\/ceph\/osd.0.keyring",
>>>>>     "heartbeat_interval": "5",
>>>>>     "heartbeat_file": "",
>>>>>     "heartbeat_inject_failure": "0",
>>>>>     "perf": "true",
>>>>>     "ms_tcp_nodelay": "true",
>>>>>     "ms_tcp_rcvbuf": "0",
>>>>>     "ms_initial_backoff": "0.2",
>>>>>     "ms_max_backoff": "15",
>>>>>     "ms_nocrc": "false",
>>>>>     "ms_die_on_bad_msg": "false",
>>>>>     "ms_die_on_unhandled_msg": "false",
>>>>>     "ms_dispatch_throttle_bytes": "104857600",
>>>>>     "ms_bind_ipv6": "false",
>>>>>     "ms_bind_port_min": "6800",
>>>>>     "ms_bind_port_max": "7100",
>>>>>     "ms_rwthread_stack_bytes": "1048576",
>>>>>     "ms_tcp_read_timeout": "900",
>>>>>     "ms_pq_max_tokens_per_priority": "4194304",
>>>>>     "ms_pq_min_cost": "65536",
>>>>>     "ms_inject_socket_failures": "0",
>>>>>     "ms_inject_delay_type": "",
>>>>>     "ms_inject_delay_max": "1",
>>>>>     "ms_inject_delay_probability": "0",
>>>>>     "ms_inject_internal_delays": "0",
>>>>>     "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
>>>>>     "mon_initial_members": "",
>>>>>     "mon_sync_fs_threshold": "5",
>>>>>     "mon_compact_on_start": "false",
>>>>>     "mon_compact_on_bootstrap": "false",
>>>>>     "mon_compact_on_trim": "true",
>>>>>     "mon_tick_interval": "5",
>>>>>     "mon_subscribe_interval": "300",
>>>>>     "mon_osd_laggy_halflife": "3600",
>>>>>     "mon_osd_laggy_weight": "0.3",
>>>>>     "mon_osd_adjust_heartbeat_grace": "true",
>>>>>     "mon_osd_adjust_down_out_interval": "true",
>>>>>     "mon_osd_auto_mark_in": "false",
>>>>>     "mon_osd_auto_mark_auto_out_in": "true",
>>>>>     "mon_osd_auto_mark_new_in": "true",
>>>>>     "mon_osd_down_out_interval": "300",
>>>>>     "mon_osd_down_out_subtree_limit": "rack",
>>>>>     "mon_osd_min_up_ratio": "0.3",
>>>>>     "mon_osd_min_in_ratio": "0.3",
>>>>>     "mon_stat_smooth_intervals": "2",
>>>>>     "mon_lease": "5",
>>>>>     "mon_lease_renew_interval": "3",
>>>>>     "mon_lease_ack_timeout": "10",
>>>>>     "mon_clock_drift_allowed": "0.05",
>>>>>     "mon_clock_drift_warn_backoff": "5",
>>>>>     "mon_timecheck_interval": "300",
>>>>>     "mon_accept_timeout": "10",
>>>>>     "mon_pg_create_interval": "30",
>>>>>     "mon_pg_stuck_threshold": "300",
>>>>>     "mon_osd_full_ratio": "0.95",
>>>>>     "mon_osd_nearfull_ratio": "0.85",
>>>>>     "mon_globalid_prealloc": "100",
>>>>>     "mon_osd_report_timeout": "900",
>>>>>     "mon_force_standby_active": "true",
>>>>>     "mon_min_osdmap_epochs": "500",
>>>>>     "mon_max_pgmap_epochs": "500",
>>>>>     "mon_max_log_epochs": "500",
>>>>>     "mon_max_osd": "10000",
>>>>>     "mon_probe_timeout": "2",
>>>>>     "mon_slurp_timeout": "10",
>>>>>     "mon_slurp_bytes": "262144",
>>>>>     "mon_client_bytes": "104857600",
>>>>>     "mon_daemon_bytes": "419430400",
>>>>>     "mon_max_log_entries_per_event": "4096",
>>>>>     "mon_health_data_update_interval": "60",
>>>>>     "mon_data_avail_crit": "5",
>>>>>     "mon_data_avail_warn": "30",
>>>>>     "mon_config_key_max_entry_size": "4096",
>>>>>     "mon_sync_trim_timeout": "30",
>>>>>     "mon_sync_heartbeat_timeout": "30",
>>>>>     "mon_sync_heartbeat_interval": "5",
>>>>>     "mon_sync_backoff_timeout": "30",
>>>>>     "mon_sync_timeout": "30",
>>>>>     "mon_sync_max_retries": "5",
>>>>>     "mon_sync_max_payload_size": "1048576",
>>>>>     "mon_sync_debug": "false",
>>>>>     "mon_sync_debug_leader": "-1",
>>>>>     "mon_sync_debug_provider": "-1",
>>>>>     "mon_sync_debug_provider_fallback": "-1",
>>>>>     "mon_debug_dump_transactions": "false",
>>>>>     "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
>>>>>     "mon_sync_leader_kill_at": "0",
>>>>>     "mon_sync_provider_kill_at": "0",
>>>>>     "mon_sync_requester_kill_at": "0",
>>>>>     "mon_leveldb_write_buffer_size": "33554432",
>>>>>     "mon_leveldb_cache_size": "268435456",
>>>>>     "mon_leveldb_block_size": "65536",
>>>>>     "mon_leveldb_bloom_size": "0",
>>>>>     "mon_leveldb_max_open_files": "0",
>>>>>     "mon_leveldb_compression": "false",
>>>>>     "mon_leveldb_paranoid": "false",
>>>>>     "mon_leveldb_log": "",
>>>>>     "paxos_stash_full_interval": "25",
>>>>>     "paxos_max_join_drift": "100",
>>>>>     "paxos_propose_interval": "1",
>>>>>     "paxos_min_wait": "0.05",
>>>>>     "paxos_min": "500",
>>>>>     "paxos_trim_min": "500",
>>>>>     "paxos_trim_max": "1000",
>>>>>     "paxos_trim_disabled_max_versions": "108000",
>>>>>     "paxos_service_trim_min": "500",
>>>>>     "paxos_service_trim_max": "1000",
>>>>>     "clock_offset": "0",
>>>>>     "auth_cluster_required": "none",
>>>>>     "auth_service_required": "none",
>>>>>     "auth_client_required": "none",
>>>>>     "auth_supported": "none",
>>>>>     "cephx_require_signatures": "false",
>>>>>     "cephx_cluster_require_signatures": "false",
>>>>>     "cephx_service_require_signatures": "false",
>>>>>     "cephx_sign_messages": "true",
>>>>>     "auth_mon_ticket_ttl": "43200",
>>>>>     "auth_service_ticket_ttl": "3600",
>>>>>     "auth_debug": "false",
>>>>>     "mon_client_hunt_interval": "3",
>>>>>     "mon_client_ping_interval": "10",
>>>>>     "mon_client_max_log_entries_per_message": "1000",
>>>>>     "mon_max_pool_pg_num": "65536",
>>>>>     "mon_pool_quota_warn_threshold": "0",
>>>>>     "mon_pool_quota_crit_threshold": "0",
>>>>>     "client_cache_size": "16384",
>>>>>     "client_cache_mid": "0.75",
>>>>>     "client_use_random_mds": "false",
>>>>>     "client_mount_timeout": "300",
>>>>>     "client_tick_interval": "1",
>>>>>     "client_trace": "",
>>>>>     "client_readahead_min": "131072",
>>>>>     "client_readahead_max_bytes": "0",
>>>>>     "client_readahead_max_periods": "4",
>>>>>     "client_snapdir": ".snap",
>>>>>     "client_mountpoint": "\/",
>>>>>     "client_notify_timeout": "10",
>>>>>     "client_caps_release_delay": "5",
>>>>>     "client_oc": "true",
>>>>>     "client_oc_size": "209715200",
>>>>>     "client_oc_max_dirty": "104857600",
>>>>>     "client_oc_target_dirty": "8388608",
>>>>>     "client_oc_max_dirty_age": "5",
>>>>>     "client_oc_max_objects": "1000",
>>>>>     "client_debug_force_sync_read": "false",
>>>>>     "client_debug_inject_tick_delay": "0",
>>>>>     "fuse_use_invalidate_cb": "false",
>>>>>     "fuse_allow_other": "true",
>>>>>     "fuse_default_permissions": "true",
>>>>>     "fuse_big_writes": "true",
>>>>>     "fuse_atomic_o_trunc": "true",
>>>>>     "fuse_debug": "false",
>>>>>     "objecter_tick_interval": "5",
>>>>>     "objecter_timeout": "10",
>>>>>     "objecter_inflight_op_bytes": "104857600",
>>>>>     "objecter_inflight_ops": "1024",
>>>>>     "journaler_allow_split_entries": "true",
>>>>>     "journaler_write_head_interval": "15",
>>>>>     "journaler_prefetch_periods": "10",
>>>>>     "journaler_prezero_periods": "5",
>>>>>     "journaler_batch_interval": "0.001",
>>>>>     "journaler_batch_max": "0",
>>>>>     "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
>>>>>     "mds_max_file_size": "1099511627776",
>>>>>     "mds_cache_size": "100000",
>>>>>     "mds_cache_mid": "0.7",
>>>>>     "mds_mem_max": "1048576",
>>>>>     "mds_dir_commit_ratio": "0.5",
>>>>>     "mds_dir_max_commit_size": "90",
>>>>>     "mds_decay_halflife": "5",
>>>>>     "mds_beacon_interval": "4",
>>>>>     "mds_beacon_grace": "15",
>>>>>     "mds_enforce_unique_name": "true",
>>>>>     "mds_blacklist_interval": "1440",
>>>>>     "mds_session_timeout": "60",
>>>>>     "mds_session_autoclose": "300",
>>>>>     "mds_reconnect_timeout": "45",
>>>>>     "mds_tick_interval": "5",
>>>>>     "mds_dirstat_min_interval": "1",
>>>>>     "mds_scatter_nudge_interval": "5",
>>>>>     "mds_client_prealloc_inos": "1000",
>>>>>     "mds_early_reply": "true",
>>>>>     "mds_use_tmap": "true",
>>>>>     "mds_default_dir_hash": "2",
>>>>>     "mds_log": "true",
>>>>>     "mds_log_skip_corrupt_events": "false",
>>>>>     "mds_log_max_events": "-1",
>>>>>     "mds_log_segment_size": "0",
>>>>>     "mds_log_max_segments": "30",
>>>>>     "mds_log_max_expiring": "20",
>>>>>     "mds_bal_sample_interval": "3",
>>>>>     "mds_bal_replicate_threshold": "8000",
>>>>>     "mds_bal_unreplicate_threshold": "0",
>>>>>     "mds_bal_frag": "false",
>>>>>     "mds_bal_split_size": "10000",
>>>>>     "mds_bal_split_rd": "25000",
>>>>>     "mds_bal_split_wr": "10000",
>>>>>     "mds_bal_split_bits": "3",
>>>>>     "mds_bal_merge_size": "50",
>>>>>     "mds_bal_merge_rd": "1000",
>>>>>     "mds_bal_merge_wr": "1000",
>>>>>     "mds_bal_interval": "10",
>>>>>     "mds_bal_fragment_interval": "5",
>>>>>     "mds_bal_idle_threshold": "0",
>>>>>     "mds_bal_max": "-1",
>>>>>     "mds_bal_max_until": "-1",
>>>>>     "mds_bal_mode": "0",
>>>>>     "mds_bal_min_rebalance": "0.1",
>>>>>     "mds_bal_min_start": "0.2",
>>>>>     "mds_bal_need_min": "0.8",
>>>>>     "mds_bal_need_max": "1.2",
>>>>>     "mds_bal_midchunk": "0.3",
>>>>>     "mds_bal_minchunk": "0.001",
>>>>>     "mds_bal_target_removal_min": "5",
>>>>>     "mds_bal_target_removal_max": "10",
>>>>>     "mds_replay_interval": "1",
>>>>>     "mds_shutdown_check": "0",
>>>>>     "mds_thrash_exports": "0",
>>>>>     "mds_thrash_fragments": "0",
>>>>>     "mds_dump_cache_on_map": "false",
>>>>>     "mds_dump_cache_after_rejoin": "false",
>>>>>     "mds_verify_scatter": "false",
>>>>>     "mds_debug_scatterstat": "false",
>>>>>     "mds_debug_frag": "false",
>>>>>     "mds_debug_auth_pins": "false",
>>>>>     "mds_debug_subtrees": "false",
>>>>>     "mds_kill_mdstable_at": "0",
>>>>>     "mds_kill_export_at": "0",
>>>>>     "mds_kill_import_at": "0",
>>>>>     "mds_kill_link_at": "0",
>>>>>     "mds_kill_rename_at": "0",
>>>>>     "mds_kill_openc_at": "0",
>>>>>     "mds_kill_journal_at": "0",
>>>>>     "mds_kill_journal_expire_at": "0",
>>>>>     "mds_kill_journal_replay_at": "0",
>>>>>     "mds_inject_traceless_reply_probability": "0",
>>>>>     "mds_wipe_sessions": "false",
>>>>>     "mds_wipe_ino_prealloc": "false",
>>>>>     "mds_skip_ino": "0",
>>>>>     "max_mds": "1",
>>>>>     "mds_standby_for_name": "",
>>>>>     "mds_standby_for_rank": "-1",
>>>>>     "mds_standby_replay": "false",
>>>>>     "osd_auto_upgrade_tmap": "true",
>>>>>     "osd_tmapput_sets_uses_tmap": "false",
>>>>>     "osd_max_backfills": "5",
>>>>>     "osd_backfill_full_ratio": "0.85",
>>>>>     "osd_backfill_retry_interval": "10",
>>>>>     "osd_uuid": "00000000-0000-0000-0000-000000000000",
>>>>>     "osd_data": "\/ceph\/osd.0\/",
>>>>>     "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
>>>>>     "osd_journal_size": "5120",
>>>>>     "osd_max_write_size": "90",
>>>>>     "osd_max_pgls": "1024",
>>>>>     "osd_client_message_size_cap": "524288000",
>>>>>     "osd_client_message_cap": "100",
>>>>>     "osd_pg_bits": "6",
>>>>>     "osd_pgp_bits": "6",
>>>>>     "osd_crush_chooseleaf_type": "1",
>>>>>     "osd_min_rep": "1",
>>>>>     "osd_max_rep": "10",
>>>>>     "osd_pool_default_crush_rule": "0",
>>>>>     "osd_pool_default_size": "2",
>>>>>     "osd_pool_default_min_size": "0",
>>>>>     "osd_pool_default_pg_num": "8",
>>>>>     "osd_pool_default_pgp_num": "8",
>>>>>     "osd_pool_default_flags": "0",
>>>>>     "osd_map_dedup": "true",
>>>>>     "osd_map_cache_size": "500",
>>>>>     "osd_map_message_max": "100",
>>>>>     "osd_map_share_max_epochs": "100",
>>>>>     "osd_op_threads": "2",
>>>>>     "osd_peering_wq_batch_size": "20",
>>>>>     "osd_op_pq_max_tokens_per_priority": "4194304",
>>>>>     "osd_op_pq_min_cost": "65536",
>>>>>     "osd_disk_threads": "1",
>>>>>     "osd_recovery_threads": "2",
>>>>>     "osd_recover_clone_overlap": "true",
>>>>>     "osd_backfill_scan_min": "64",
>>>>>     "osd_backfill_scan_max": "512",
>>>>>     "osd_op_thread_timeout": "15",
>>>>>     "osd_recovery_thread_timeout": "30",
>>>>>     "osd_snap_trim_thread_timeout": "3600",
>>>>>     "osd_scrub_thread_timeout": "60",
>>>>>     "osd_scrub_finalize_thread_timeout": "600",
>>>>>     "osd_remove_thread_timeout": "3600",
>>>>>     "osd_command_thread_timeout": "600",
>>>>>     "osd_age": "0.8",
>>>>>     "osd_age_time": "0",
>>>>>     "osd_heartbeat_addr": ":\/0",
>>>>>     "osd_heartbeat_interval": "6",
>>>>>     "osd_heartbeat_grace": "20",
>>>>>     "osd_mon_heartbeat_interval": "30",
>>>>>     "osd_mon_report_interval_max": "120",
>>>>>     "osd_mon_report_interval_min": "5",
>>>>>     "osd_pg_stat_report_interval_max": "500",
>>>>>     "osd_mon_ack_timeout": "30",
>>>>>     "osd_min_down_reporters": "1",
>>>>>     "osd_min_down_reports": "3",
>>>>>     "osd_default_data_pool_replay_window": "45",
>>>>>     "osd_preserve_trimmed_log": "false",
>>>>>     "osd_auto_mark_unfound_lost": "false",
>>>>>     "osd_recovery_delay_start": "0",
>>>>>     "osd_recovery_max_active": "5",
>>>>>     "osd_recovery_max_chunk": "8388608",
>>>>>     "osd_recovery_forget_lost_objects": "false",
>>>>>     "osd_max_scrubs": "1",
>>>>>     "osd_scrub_load_threshold": "0.5",
>>>>>     "osd_scrub_min_interval": "86400",
>>>>>     "osd_scrub_max_interval": "604800",
>>>>>     "osd_deep_scrub_interval": "604800",
>>>>>     "osd_deep_scrub_stride": "524288",
>>>>>     "osd_scan_list_ping_tp_interval": "100",
>>>>>     "osd_auto_weight": "false",
>>>>>     "osd_class_dir": "\/usr\/lib\/rados-classes",
>>>>>     "osd_check_for_log_corruption": "false",
>>>>>     "osd_use_stale_snap": "false",
>>>>>     "osd_rollback_to_cluster_snap": "",
>>>>>     "osd_default_notify_timeout": "30",
>>>>>     "osd_kill_backfill_at": "0",
>>>>>     "osd_pg_epoch_persisted_max_stale": "200",
>>>>>     "osd_min_pg_log_entries": "500",
>>>>>     "osd_max_pg_log_entries": "1500",
>>>>>     "osd_op_complaint_time": "30",
>>>>>     "osd_command_max_records": "256",
>>>>>     "osd_op_log_threshold": "5",
>>>>>     "osd_verify_sparse_read_holes": "false",
>>>>>     "osd_debug_drop_ping_probability": "0",
>>>>>     "osd_debug_drop_ping_duration": "0",
>>>>>     "osd_debug_drop_pg_create_probability": "0",
>>>>>     "osd_debug_drop_pg_create_duration": "1",
>>>>>     "osd_debug_drop_op_probability": "0",
>>>>>     "osd_debug_op_order": "false",
>>>>>     "osd_debug_verify_snaps_on_info": "false",
>>>>>     "osd_debug_verify_stray_on_activate": "false",
>>>>>     "osd_debug_skip_full_check_in_backfill_reservation": "false",
>>>>>     "osd_op_history_size": "20",
>>>>>     "osd_op_history_duration": "600",
>>>>>     "osd_target_transaction_size": "30",
>>>>>     "osd_failsafe_full_ratio": "0.97",
>>>>>     "osd_failsafe_nearfull_ratio": "0.9",
>>>>>     "osd_leveldb_write_buffer_size": "0",
>>>>>     "osd_leveldb_cache_size": "0",
>>>>>     "osd_leveldb_block_size": "0",
>>>>>     "osd_leveldb_bloom_size": "0",
>>>>>     "osd_leveldb_max_open_files": "0",
>>>>>     "osd_leveldb_compression": "true",
>>>>>     "osd_leveldb_paranoid": "false",
>>>>>     "osd_leveldb_log": "",
>>>>>     "osd_client_op_priority": "63",
>>>>>     "osd_recovery_op_priority": "50",
>>>>>     "osd_mon_shutdown_timeout": "5",
>>>>>     "filestore": "false",
>>>>>     "filestore_index_retry_probability": "0",
>>>>>     "filestore_debug_inject_read_err": "false",
>>>>>     "filestore_debug_omap_check": "false",
>>>>>     "filestore_xattr_use_omap": "false",
>>>>>     "filestore_max_inline_xattr_size": "512",
>>>>>     "filestore_max_inline_xattrs": "2",
>>>>>     "filestore_max_sync_interval": "5",
>>>>>     "filestore_min_sync_interval": "0.01",
>>>>>     "filestore_btrfs_snap": "true",
>>>>>     "filestore_btrfs_clone_range": "true",
>>>>>     "filestore_fsync_flushes_journal_data": "false",
>>>>>     "filestore_fiemap": "false",
>>>>>     "filestore_flusher": "true",
>>>>>     "filestore_flusher_max_fds": "512",
>>>>>     "filestore_flush_min": "65536",
>>>>>     "filestore_sync_flush": "false",
>>>>>     "filestore_journal_parallel": "false",
>>>>>     "filestore_journal_writeahead": "false",
>>>>>     "filestore_journal_trailing": "false",
>>>>>     "filestore_queue_max_ops": "500",
>>>>>     "filestore_queue_max_bytes": "104857600",
>>>>>     "filestore_queue_committing_max_ops": "5000",
>>>>>     "filestore_queue_committing_max_bytes": "104857600",
>>>>>     "filestore_op_threads": "2",
>>>>>     "filestore_op_thread_timeout": "60",
>>>>>     "filestore_op_thread_suicide_timeout": "180",
>>>>>     "filestore_commit_timeout": "600",
>>>>>     "filestore_fiemap_threshold": "4096",
>>>>>     "filestore_merge_threshold": "10",
>>>>>     "filestore_split_multiple": "2",
>>>>>     "filestore_update_to": "1000",
>>>>>     "filestore_blackhole": "false",
>>>>>     "filestore_dump_file": "",
>>>>>     "filestore_kill_at": "0",
>>>>>     "filestore_inject_stall": "0",
>>>>>     "filestore_fail_eio": "true",
>>>>>     "filestore_replica_fadvise": "true",
>>>>>     "filestore_debug_verify_split": "false",
>>>>>     "journal_dio": "true",
>>>>>     "journal_aio": "true",
>>>>>     "journal_force_aio": "false",
>>>>>     "journal_max_corrupt_search": "10485760",
>>>>>     "journal_block_align": "true",
>>>>>     "journal_write_header_frequency": "0",
>>>>>     "journal_max_write_bytes": "10485760",
>>>>>     "journal_max_write_entries": "100",
>>>>>     "journal_queue_max_ops": "300",
>>>>>     "journal_queue_max_bytes": "33554432",
>>>>>     "journal_align_min_size": "65536",
>>>>>     "journal_replay_from": "0",
>>>>>     "journal_zero_on_create": "false",
>>>>>     "journal_ignore_corruption": "false",
>>>>>     "rbd_cache": "false",
>>>>>     "rbd_cache_writethrough_until_flush": "false",
>>>>>     "rbd_cache_size": "33554432",
>>>>>     "rbd_cache_max_dirty": "25165824",
>>>>>     "rbd_cache_target_dirty": "16777216",
>>>>>     "rbd_cache_max_dirty_age": "1",
>>>>>     "rbd_cache_block_writes_upfront": "false",
>>>>>     "rbd_concurrent_management_ops": "10",
>>>>>     "rbd_default_format": "1",
>>>>>     "rbd_default_order": "22",
>>>>>     "rbd_default_stripe_count": "1",
>>>>>     "rbd_default_stripe_unit": "4194304",
>>>>>     "rbd_default_features": "3",
>>>>>     "nss_db_path": "",
>>>>>     "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
>>>>>     "rgw_enable_apis": "s3, swift, swift_auth, admin",
>>>>>     "rgw_cache_enabled": "true",
>>>>>     "rgw_cache_lru_size": "10000",
>>>>>     "rgw_socket_path": "",
>>>>>     "rgw_host": "",
>>>>>     "rgw_port": "",
>>>>>     "rgw_dns_name": "",
>>>>>     "rgw_script_uri": "",
>>>>>     "rgw_request_uri": "",
>>>>>     "rgw_swift_url": "",
>>>>>     "rgw_swift_url_prefix": "swift",
>>>>>     "rgw_swift_auth_url": "",
>>>>>     "rgw_swift_auth_entry": "auth",
>>>>>     "rgw_keystone_url": "",
>>>>>     "rgw_keystone_admin_token": "",
>>>>>     "rgw_keystone_accepted_roles": "Member, admin",
>>>>>     "rgw_keystone_token_cache_size": "10000",
>>>>>     "rgw_keystone_revocation_interval": "900",
>>>>>     "rgw_admin_entry": "admin",
>>>>>     "rgw_enforce_swift_acls": "true",
>>>>>     "rgw_swift_token_expiration": "86400",
>>>>>     "rgw_print_continue": "true",
>>>>>     "rgw_remote_addr_param": "REMOTE_ADDR",
>>>>>     "rgw_op_thread_timeout": "600",
>>>>>     "rgw_op_thread_suicide_timeout": "0",
>>>>>     "rgw_thread_pool_size": "100",
>>>>>     "rgw_num_control_oids": "8",
>>>>>     "rgw_zone_root_pool": ".rgw.root",
>>>>>     "rgw_log_nonexistent_bucket": "false",
>>>>>     "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
>>>>>     "rgw_log_object_name_utc": "false",
>>>>>     "rgw_usage_max_shards": "32",
>>>>>     "rgw_usage_max_user_shards": "1",
>>>>>     "rgw_enable_ops_log": "false",
>>>>>     "rgw_enable_usage_log": "false",
>>>>>     "rgw_ops_log_rados": "true",
>>>>>     "rgw_ops_log_socket_path": "",
>>>>>     "rgw_ops_log_data_backlog": "5242880",
>>>>>     "rgw_usage_log_flush_threshold": "1024",
>>>>>     "rgw_usage_log_tick_interval": "30",
>>>>>     "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
>>>>>     "rgw_intent_log_object_name_utc": "false",
>>>>>     "rgw_init_timeout": "300",
>>>>>     "rgw_mime_types_file": "\/etc\/mime.types",
>>>>>     "rgw_gc_max_objs": "32",
>>>>>     "rgw_gc_obj_min_wait": "7200",
>>>>>     "rgw_gc_processor_max_time": "3600",
>>>>>     "rgw_gc_processor_period": "3600",
>>>>>     "rgw_s3_success_create_obj_status": "0",
>>>>>     "rgw_resolve_cname": "false",
>>>>>     "rgw_obj_stripe_size": "4194304",
>>>>>     "rgw_extended_http_attrs": "",
>>>>>     "rgw_exit_timeout_secs": "120",
>>>>>     "rgw_get_obj_window_size": "16777216",
>>>>>     "rgw_get_obj_max_req_size": "4194304",
>>>>>     "rgw_relaxed_s3_bucket_names": "false",
>>>>>     "rgw_list_buckets_max_chunk": "1000",
>>>>>     "mutex_perf_counter": "false",
>>>>>     "internal_safe_to_start_threads": "true"}
>>>>>
>>>>>
>>>>>
>>>>> Stefan
>>>>>
>>>>>
>>>>>> -Sam
>>>>>>
>>>>>> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe
>>>>>> <s.priebe@profihost.ag>
>>>>>> wrote:
>>>>>>>
>>>>>>>
>>>>>>> Mike we already have the async patch running. Yes it helps but only
>>>>>>> helps
>>>>>>> it
>>>>>>> does not solve. It just hides the issue ...
>>>>>>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>>>>>>
>>>>>>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>>>>>>
>>>>>>>> - ceph osd set noout
>>>>>>>>
>>>>>>>> - Reboot one of the nodes hosting OSDs
>>>>>>>>         - VMs mounted from RBD volumes work properly
>>>>>>>>
>>>>>>>> - I see the OSD's boot messages as they re-join the cluster
>>>>>>>>
>>>>>>>> - Start seeing active+recovery_wait, peering, and active+recovering
>>>>>>>>         - VMs mounted from RBD volumes become unresponsive.
>>>>>>>>
>>>>>>>> - Recovery completes
>>>>>>>>         - VMs mounted from RBD volumes regain responsiveness
>>>>>>>>
>>>>>>>> - ceph osd unset noout
>>>>>>>>
>>>>>>>> Would joshd's async patch for qemu help here, or is there something
>>>>>>>> else
>>>>>>>> going on?
>>>>>>>>
>>>>>>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>>
>>>>>>>> Mike Dawson
>>>>>>>> Co-Founder & Director of Cloud Architecture
>>>>>>>> Cloudapt LLC
>>>>>>>> 6330 East 75th Street, Suite 170
>>>>>>>> Indianapolis, IN 46250
>>>>>>>>
>>>>>>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Can you reproduce and attach the ceph.log from before you stop the
>>>>>>>>> osd
>>>>>>>>> until after you have started the osd and it has recovered?
>>>>>>>>> -Sam
>>>>>>>>>
>>>>>>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> i still have recovery issues with cuttlefish. After the OSD comes
>>>>>>>>>> back
>>>>>>>>>> it seem to hang for around 2-4 minutes and then recovery seems to
>>>>>>>>>> start
>>>>>>>>>> (pgs in recovery_wait start to decrement). This is with ceph
>>>>>>>>>> 0.61.7.
>>>>>>>>>> I
>>>>>>>>>> get a lot of slow request messages an hanging VMs.
>>>>>>>>>>
>>>>>>>>>> What i noticed today is that if i leave the OSD off as long as
>>>>>>>>>> ceph
>>>>>>>>>> starts to backfill - the recovery and "re" backfilling wents
>>>>>>>>>> absolutely
>>>>>>>>>> smooth without any issues and no slow request messages at all.
>>>>>>>>>>
>>>>>>>>>> Does anybody have an idea why?
>>>>>>>>>>
>>>>>>>>>> Greets,
>>>>>>>>>> Stefan
>>>>>>>>>> --
>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>> ceph-devel"
>>>>>>>>>> in
>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>> More majordomo info at
>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>> ceph-devel"
>>>>>>>>> in
>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>>>
>>>>>>>
>>>>>> --
>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>> ceph-devel" in
>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>>
>>>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-08 14:05                   ` Mike Dawson
@ 2013-08-08 15:43                     ` Oliver Francke
  2013-08-08 18:13                     ` Stefan Priebe
  1 sibling, 0 replies; 40+ messages in thread
From: Oliver Francke @ 2013-08-08 15:43 UTC (permalink / raw)
  To: Mike Dawson
  Cc: Stefan Priebe, Samuel Just, josh.durgin, ceph-devel, Stefan Hajnoczi

Hi Mike,

On 08/08/2013 04:05 PM, Mike Dawson wrote:
> Stefan,
>
> I see the same behavior and I theorize it is linked to an issue 
> detailed in another thread [0]. Do your VM guests ever hang while your 
> cluster is HEALTH_OK like described in that other thread?
>
> [0] http://comments.gmane.org/gmane.comp.file-systems.ceph.user/2982
>
> A few observations:
>
> - The VMs that hang do lots of writes (video surveillance).
> - I use rbd and qemu. The problem exists in both qemu 1.4.x and 1.5.2.
> - The problem exists with or without joshd's qemu async flush patch.
> - Windows VMs seem to be more vulnerable than Linux VMs.
> - If I restart the qemu-system-x86_64 process, the guest will come 
> back to life.
> - A partial workaround seems to be console input (NoVNC or 'virsh 
> screenshot'), but restarting qemu-system-x86_64 works better.
> - The issue of VMs hanging seems worse with RBD writeback cache enabled
> - I typically run virtio, but I believe I've seen it with e1000, too.
> - VM guests hang at different times, not all at once on a host (or 
> across all hosts).
> - I co-mingle VM guests on servers that host ceph OSDs.
>
>
>
> Oliver,
>
> If your cluster has to recover/backfill, do your guest VMs hang with 
> more frequency than under normal HEALTH_OK conditions, even if you 
> prioritize client IO as Sam wrote below?

well, at least I can confirm, that with the stupid "while 
install/remove-loop" alone and a "rbd cp some/stuff some/other" and 
ongoing remapping/backfilling the 120-secs problem occurs, too.
though with a spew-test inside the VM the occurance happens 
earlier/sooner/quicker.
That now in a LAB-environment to let alone production ;)

Oliver.

>
>
> Sam,
>
> Turning down all the settings you mentioned certainly does slow the 
> recover/backfill process, but it doesn't prevent the VM guests backed 
> by RBD volumes from hanging. In fact, I often try to prioritize 
> recovery/backfill because my guests tend to hang until I get back to 
> HEALTH_OK. Given this apparent bug, completing recovery/backfill 
> quicker leads to less total outage, it seems.
>
>
> Josh,
>
> How can I help you investigate if RBD is the common source of both of 
> these issues?
>
>
> Thanks,
> Mike Dawson
>
>
> On 8/2/2013 2:46 PM, Stefan Priebe wrote:
>> Hi,
>>
>>          osd recovery max active = 1
>>          osd max backfills = 1
>>          osd recovery op priority = 5
>>
>> still no difference...
>>
>> Stefan
>>
>> Am 02.08.2013 20:21, schrieb Samuel Just:
>>> Also, you have osd_recovery_op_priority at 50.  That is close to the
>>> priority of client IO.  You want it below 10 (defaults to 10), perhaps
>>> at 1.  You can also adjust down osd_recovery_max_active.
>>> -Sam
>>>
>>> On Fri, Aug 2, 2013 at 11:16 AM, Stefan Priebe <s.priebe@profihost.ag>
>>> wrote:
>>>> I already tried both values this makes no difference. The drives are
>>>> not the
>>>> bottleneck.
>>>>
>>>> Am 02.08.2013 19:35, schrieb Samuel Just:
>>>>
>>>>> You might try turning osd_max_backfills to 2 or 1.
>>>>> -Sam
>>>>>
>>>>> On Fri, Aug 2, 2013 at 12:44 AM, Stefan Priebe 
>>>>> <s.priebe@profihost.ag>
>>>>> wrote:
>>>>>>
>>>>>> Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd
>>>>>> settings?
>>>>>>
>>>>>>> sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
>>>>>>
>>>>>>
>>>>>> Sure.
>>>>>>
>>>>>>
>>>>>>
>>>>>> { "name": "osd.0",
>>>>>>     "cluster": "ceph",
>>>>>>     "none": "0\/5",
>>>>>>     "lockdep": "0\/0",
>>>>>>     "context": "0\/0",
>>>>>>     "crush": "0\/0",
>>>>>>     "mds": "0\/0",
>>>>>>     "mds_balancer": "0\/0",
>>>>>>     "mds_locker": "0\/0",
>>>>>>     "mds_log": "0\/0",
>>>>>>     "mds_log_expire": "0\/0",
>>>>>>     "mds_migrator": "0\/0",
>>>>>>     "buffer": "0\/0",
>>>>>>     "timer": "0\/0",
>>>>>>     "filer": "0\/0",
>>>>>>     "striper": "0\/1",
>>>>>>     "objecter": "0\/0",
>>>>>>     "rados": "0\/0",
>>>>>>     "rbd": "0\/0",
>>>>>>     "journaler": "0\/0",
>>>>>>     "objectcacher": "0\/0",
>>>>>>     "client": "0\/0",
>>>>>>     "osd": "0\/0",
>>>>>>     "optracker": "0\/0",
>>>>>>     "objclass": "0\/0",
>>>>>>     "filestore": "0\/0",
>>>>>>     "journal": "0\/0",
>>>>>>     "ms": "0\/0",
>>>>>>     "mon": "0\/0",
>>>>>>     "monc": "0\/0",
>>>>>>     "paxos": "0\/0",
>>>>>>     "tp": "0\/0",
>>>>>>     "auth": "0\/0",
>>>>>>     "crypto": "1\/5",
>>>>>>     "finisher": "0\/0",
>>>>>>     "heartbeatmap": "0\/0",
>>>>>>     "perfcounter": "0\/0",
>>>>>>     "rgw": "0\/0",
>>>>>>     "hadoop": "0\/0",
>>>>>>     "javaclient": "1\/5",
>>>>>>     "asok": "0\/0",
>>>>>>     "throttle": "0\/0",
>>>>>>     "host": "cloud1-1268",
>>>>>>     "fsid": "00000000-0000-0000-0000-000000000000",
>>>>>>     "public_addr": "10.255.0.90:0\/0",
>>>>>>     "cluster_addr": "10.255.0.90:0\/0",
>>>>>>     "public_network": "10.255.0.1\/24",
>>>>>>     "cluster_network": "10.255.0.1\/24",
>>>>>>     "num_client": "1",
>>>>>>     "monmap": "",
>>>>>>     "mon_host": "",
>>>>>>     "lockdep": "false",
>>>>>>     "run_dir": "\/var\/run\/ceph",
>>>>>>     "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
>>>>>>     "daemonize": "true",
>>>>>>     "pid_file": "\/var\/run\/ceph\/osd.0.pid",
>>>>>>     "chdir": "\/",
>>>>>>     "max_open_files": "0",
>>>>>>     "fatal_signal_handlers": "true",
>>>>>>     "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
>>>>>>     "log_max_new": "1000",
>>>>>>     "log_max_recent": "10000",
>>>>>>     "log_to_stderr": "false",
>>>>>>     "err_to_stderr": "true",
>>>>>>     "log_to_syslog": "false",
>>>>>>     "err_to_syslog": "false",
>>>>>>     "log_flush_on_exit": "true",
>>>>>>     "log_stop_at_utilization": "0.97",
>>>>>>     "clog_to_monitors": "true",
>>>>>>     "clog_to_syslog": "false",
>>>>>>     "clog_to_syslog_level": "info",
>>>>>>     "clog_to_syslog_facility": "daemon",
>>>>>>     "mon_cluster_log_to_syslog": "false",
>>>>>>     "mon_cluster_log_to_syslog_level": "info",
>>>>>>     "mon_cluster_log_to_syslog_facility": "daemon",
>>>>>>     "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
>>>>>>     "key": "",
>>>>>>     "keyfile": "",
>>>>>>     "keyring": "\/etc\/ceph\/osd.0.keyring",
>>>>>>     "heartbeat_interval": "5",
>>>>>>     "heartbeat_file": "",
>>>>>>     "heartbeat_inject_failure": "0",
>>>>>>     "perf": "true",
>>>>>>     "ms_tcp_nodelay": "true",
>>>>>>     "ms_tcp_rcvbuf": "0",
>>>>>>     "ms_initial_backoff": "0.2",
>>>>>>     "ms_max_backoff": "15",
>>>>>>     "ms_nocrc": "false",
>>>>>>     "ms_die_on_bad_msg": "false",
>>>>>>     "ms_die_on_unhandled_msg": "false",
>>>>>>     "ms_dispatch_throttle_bytes": "104857600",
>>>>>>     "ms_bind_ipv6": "false",
>>>>>>     "ms_bind_port_min": "6800",
>>>>>>     "ms_bind_port_max": "7100",
>>>>>>     "ms_rwthread_stack_bytes": "1048576",
>>>>>>     "ms_tcp_read_timeout": "900",
>>>>>>     "ms_pq_max_tokens_per_priority": "4194304",
>>>>>>     "ms_pq_min_cost": "65536",
>>>>>>     "ms_inject_socket_failures": "0",
>>>>>>     "ms_inject_delay_type": "",
>>>>>>     "ms_inject_delay_max": "1",
>>>>>>     "ms_inject_delay_probability": "0",
>>>>>>     "ms_inject_internal_delays": "0",
>>>>>>     "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
>>>>>>     "mon_initial_members": "",
>>>>>>     "mon_sync_fs_threshold": "5",
>>>>>>     "mon_compact_on_start": "false",
>>>>>>     "mon_compact_on_bootstrap": "false",
>>>>>>     "mon_compact_on_trim": "true",
>>>>>>     "mon_tick_interval": "5",
>>>>>>     "mon_subscribe_interval": "300",
>>>>>>     "mon_osd_laggy_halflife": "3600",
>>>>>>     "mon_osd_laggy_weight": "0.3",
>>>>>>     "mon_osd_adjust_heartbeat_grace": "true",
>>>>>>     "mon_osd_adjust_down_out_interval": "true",
>>>>>>     "mon_osd_auto_mark_in": "false",
>>>>>>     "mon_osd_auto_mark_auto_out_in": "true",
>>>>>>     "mon_osd_auto_mark_new_in": "true",
>>>>>>     "mon_osd_down_out_interval": "300",
>>>>>>     "mon_osd_down_out_subtree_limit": "rack",
>>>>>>     "mon_osd_min_up_ratio": "0.3",
>>>>>>     "mon_osd_min_in_ratio": "0.3",
>>>>>>     "mon_stat_smooth_intervals": "2",
>>>>>>     "mon_lease": "5",
>>>>>>     "mon_lease_renew_interval": "3",
>>>>>>     "mon_lease_ack_timeout": "10",
>>>>>>     "mon_clock_drift_allowed": "0.05",
>>>>>>     "mon_clock_drift_warn_backoff": "5",
>>>>>>     "mon_timecheck_interval": "300",
>>>>>>     "mon_accept_timeout": "10",
>>>>>>     "mon_pg_create_interval": "30",
>>>>>>     "mon_pg_stuck_threshold": "300",
>>>>>>     "mon_osd_full_ratio": "0.95",
>>>>>>     "mon_osd_nearfull_ratio": "0.85",
>>>>>>     "mon_globalid_prealloc": "100",
>>>>>>     "mon_osd_report_timeout": "900",
>>>>>>     "mon_force_standby_active": "true",
>>>>>>     "mon_min_osdmap_epochs": "500",
>>>>>>     "mon_max_pgmap_epochs": "500",
>>>>>>     "mon_max_log_epochs": "500",
>>>>>>     "mon_max_osd": "10000",
>>>>>>     "mon_probe_timeout": "2",
>>>>>>     "mon_slurp_timeout": "10",
>>>>>>     "mon_slurp_bytes": "262144",
>>>>>>     "mon_client_bytes": "104857600",
>>>>>>     "mon_daemon_bytes": "419430400",
>>>>>>     "mon_max_log_entries_per_event": "4096",
>>>>>>     "mon_health_data_update_interval": "60",
>>>>>>     "mon_data_avail_crit": "5",
>>>>>>     "mon_data_avail_warn": "30",
>>>>>>     "mon_config_key_max_entry_size": "4096",
>>>>>>     "mon_sync_trim_timeout": "30",
>>>>>>     "mon_sync_heartbeat_timeout": "30",
>>>>>>     "mon_sync_heartbeat_interval": "5",
>>>>>>     "mon_sync_backoff_timeout": "30",
>>>>>>     "mon_sync_timeout": "30",
>>>>>>     "mon_sync_max_retries": "5",
>>>>>>     "mon_sync_max_payload_size": "1048576",
>>>>>>     "mon_sync_debug": "false",
>>>>>>     "mon_sync_debug_leader": "-1",
>>>>>>     "mon_sync_debug_provider": "-1",
>>>>>>     "mon_sync_debug_provider_fallback": "-1",
>>>>>>     "mon_debug_dump_transactions": "false",
>>>>>>     "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
>>>>>>     "mon_sync_leader_kill_at": "0",
>>>>>>     "mon_sync_provider_kill_at": "0",
>>>>>>     "mon_sync_requester_kill_at": "0",
>>>>>>     "mon_leveldb_write_buffer_size": "33554432",
>>>>>>     "mon_leveldb_cache_size": "268435456",
>>>>>>     "mon_leveldb_block_size": "65536",
>>>>>>     "mon_leveldb_bloom_size": "0",
>>>>>>     "mon_leveldb_max_open_files": "0",
>>>>>>     "mon_leveldb_compression": "false",
>>>>>>     "mon_leveldb_paranoid": "false",
>>>>>>     "mon_leveldb_log": "",
>>>>>>     "paxos_stash_full_interval": "25",
>>>>>>     "paxos_max_join_drift": "100",
>>>>>>     "paxos_propose_interval": "1",
>>>>>>     "paxos_min_wait": "0.05",
>>>>>>     "paxos_min": "500",
>>>>>>     "paxos_trim_min": "500",
>>>>>>     "paxos_trim_max": "1000",
>>>>>>     "paxos_trim_disabled_max_versions": "108000",
>>>>>>     "paxos_service_trim_min": "500",
>>>>>>     "paxos_service_trim_max": "1000",
>>>>>>     "clock_offset": "0",
>>>>>>     "auth_cluster_required": "none",
>>>>>>     "auth_service_required": "none",
>>>>>>     "auth_client_required": "none",
>>>>>>     "auth_supported": "none",
>>>>>>     "cephx_require_signatures": "false",
>>>>>>     "cephx_cluster_require_signatures": "false",
>>>>>>     "cephx_service_require_signatures": "false",
>>>>>>     "cephx_sign_messages": "true",
>>>>>>     "auth_mon_ticket_ttl": "43200",
>>>>>>     "auth_service_ticket_ttl": "3600",
>>>>>>     "auth_debug": "false",
>>>>>>     "mon_client_hunt_interval": "3",
>>>>>>     "mon_client_ping_interval": "10",
>>>>>>     "mon_client_max_log_entries_per_message": "1000",
>>>>>>     "mon_max_pool_pg_num": "65536",
>>>>>>     "mon_pool_quota_warn_threshold": "0",
>>>>>>     "mon_pool_quota_crit_threshold": "0",
>>>>>>     "client_cache_size": "16384",
>>>>>>     "client_cache_mid": "0.75",
>>>>>>     "client_use_random_mds": "false",
>>>>>>     "client_mount_timeout": "300",
>>>>>>     "client_tick_interval": "1",
>>>>>>     "client_trace": "",
>>>>>>     "client_readahead_min": "131072",
>>>>>>     "client_readahead_max_bytes": "0",
>>>>>>     "client_readahead_max_periods": "4",
>>>>>>     "client_snapdir": ".snap",
>>>>>>     "client_mountpoint": "\/",
>>>>>>     "client_notify_timeout": "10",
>>>>>>     "client_caps_release_delay": "5",
>>>>>>     "client_oc": "true",
>>>>>>     "client_oc_size": "209715200",
>>>>>>     "client_oc_max_dirty": "104857600",
>>>>>>     "client_oc_target_dirty": "8388608",
>>>>>>     "client_oc_max_dirty_age": "5",
>>>>>>     "client_oc_max_objects": "1000",
>>>>>>     "client_debug_force_sync_read": "false",
>>>>>>     "client_debug_inject_tick_delay": "0",
>>>>>>     "fuse_use_invalidate_cb": "false",
>>>>>>     "fuse_allow_other": "true",
>>>>>>     "fuse_default_permissions": "true",
>>>>>>     "fuse_big_writes": "true",
>>>>>>     "fuse_atomic_o_trunc": "true",
>>>>>>     "fuse_debug": "false",
>>>>>>     "objecter_tick_interval": "5",
>>>>>>     "objecter_timeout": "10",
>>>>>>     "objecter_inflight_op_bytes": "104857600",
>>>>>>     "objecter_inflight_ops": "1024",
>>>>>>     "journaler_allow_split_entries": "true",
>>>>>>     "journaler_write_head_interval": "15",
>>>>>>     "journaler_prefetch_periods": "10",
>>>>>>     "journaler_prezero_periods": "5",
>>>>>>     "journaler_batch_interval": "0.001",
>>>>>>     "journaler_batch_max": "0",
>>>>>>     "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
>>>>>>     "mds_max_file_size": "1099511627776",
>>>>>>     "mds_cache_size": "100000",
>>>>>>     "mds_cache_mid": "0.7",
>>>>>>     "mds_mem_max": "1048576",
>>>>>>     "mds_dir_commit_ratio": "0.5",
>>>>>>     "mds_dir_max_commit_size": "90",
>>>>>>     "mds_decay_halflife": "5",
>>>>>>     "mds_beacon_interval": "4",
>>>>>>     "mds_beacon_grace": "15",
>>>>>>     "mds_enforce_unique_name": "true",
>>>>>>     "mds_blacklist_interval": "1440",
>>>>>>     "mds_session_timeout": "60",
>>>>>>     "mds_session_autoclose": "300",
>>>>>>     "mds_reconnect_timeout": "45",
>>>>>>     "mds_tick_interval": "5",
>>>>>>     "mds_dirstat_min_interval": "1",
>>>>>>     "mds_scatter_nudge_interval": "5",
>>>>>>     "mds_client_prealloc_inos": "1000",
>>>>>>     "mds_early_reply": "true",
>>>>>>     "mds_use_tmap": "true",
>>>>>>     "mds_default_dir_hash": "2",
>>>>>>     "mds_log": "true",
>>>>>>     "mds_log_skip_corrupt_events": "false",
>>>>>>     "mds_log_max_events": "-1",
>>>>>>     "mds_log_segment_size": "0",
>>>>>>     "mds_log_max_segments": "30",
>>>>>>     "mds_log_max_expiring": "20",
>>>>>>     "mds_bal_sample_interval": "3",
>>>>>>     "mds_bal_replicate_threshold": "8000",
>>>>>>     "mds_bal_unreplicate_threshold": "0",
>>>>>>     "mds_bal_frag": "false",
>>>>>>     "mds_bal_split_size": "10000",
>>>>>>     "mds_bal_split_rd": "25000",
>>>>>>     "mds_bal_split_wr": "10000",
>>>>>>     "mds_bal_split_bits": "3",
>>>>>>     "mds_bal_merge_size": "50",
>>>>>>     "mds_bal_merge_rd": "1000",
>>>>>>     "mds_bal_merge_wr": "1000",
>>>>>>     "mds_bal_interval": "10",
>>>>>>     "mds_bal_fragment_interval": "5",
>>>>>>     "mds_bal_idle_threshold": "0",
>>>>>>     "mds_bal_max": "-1",
>>>>>>     "mds_bal_max_until": "-1",
>>>>>>     "mds_bal_mode": "0",
>>>>>>     "mds_bal_min_rebalance": "0.1",
>>>>>>     "mds_bal_min_start": "0.2",
>>>>>>     "mds_bal_need_min": "0.8",
>>>>>>     "mds_bal_need_max": "1.2",
>>>>>>     "mds_bal_midchunk": "0.3",
>>>>>>     "mds_bal_minchunk": "0.001",
>>>>>>     "mds_bal_target_removal_min": "5",
>>>>>>     "mds_bal_target_removal_max": "10",
>>>>>>     "mds_replay_interval": "1",
>>>>>>     "mds_shutdown_check": "0",
>>>>>>     "mds_thrash_exports": "0",
>>>>>>     "mds_thrash_fragments": "0",
>>>>>>     "mds_dump_cache_on_map": "false",
>>>>>>     "mds_dump_cache_after_rejoin": "false",
>>>>>>     "mds_verify_scatter": "false",
>>>>>>     "mds_debug_scatterstat": "false",
>>>>>>     "mds_debug_frag": "false",
>>>>>>     "mds_debug_auth_pins": "false",
>>>>>>     "mds_debug_subtrees": "false",
>>>>>>     "mds_kill_mdstable_at": "0",
>>>>>>     "mds_kill_export_at": "0",
>>>>>>     "mds_kill_import_at": "0",
>>>>>>     "mds_kill_link_at": "0",
>>>>>>     "mds_kill_rename_at": "0",
>>>>>>     "mds_kill_openc_at": "0",
>>>>>>     "mds_kill_journal_at": "0",
>>>>>>     "mds_kill_journal_expire_at": "0",
>>>>>>     "mds_kill_journal_replay_at": "0",
>>>>>>     "mds_inject_traceless_reply_probability": "0",
>>>>>>     "mds_wipe_sessions": "false",
>>>>>>     "mds_wipe_ino_prealloc": "false",
>>>>>>     "mds_skip_ino": "0",
>>>>>>     "max_mds": "1",
>>>>>>     "mds_standby_for_name": "",
>>>>>>     "mds_standby_for_rank": "-1",
>>>>>>     "mds_standby_replay": "false",
>>>>>>     "osd_auto_upgrade_tmap": "true",
>>>>>>     "osd_tmapput_sets_uses_tmap": "false",
>>>>>>     "osd_max_backfills": "5",
>>>>>>     "osd_backfill_full_ratio": "0.85",
>>>>>>     "osd_backfill_retry_interval": "10",
>>>>>>     "osd_uuid": "00000000-0000-0000-0000-000000000000",
>>>>>>     "osd_data": "\/ceph\/osd.0\/",
>>>>>>     "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
>>>>>>     "osd_journal_size": "5120",
>>>>>>     "osd_max_write_size": "90",
>>>>>>     "osd_max_pgls": "1024",
>>>>>>     "osd_client_message_size_cap": "524288000",
>>>>>>     "osd_client_message_cap": "100",
>>>>>>     "osd_pg_bits": "6",
>>>>>>     "osd_pgp_bits": "6",
>>>>>>     "osd_crush_chooseleaf_type": "1",
>>>>>>     "osd_min_rep": "1",
>>>>>>     "osd_max_rep": "10",
>>>>>>     "osd_pool_default_crush_rule": "0",
>>>>>>     "osd_pool_default_size": "2",
>>>>>>     "osd_pool_default_min_size": "0",
>>>>>>     "osd_pool_default_pg_num": "8",
>>>>>>     "osd_pool_default_pgp_num": "8",
>>>>>>     "osd_pool_default_flags": "0",
>>>>>>     "osd_map_dedup": "true",
>>>>>>     "osd_map_cache_size": "500",
>>>>>>     "osd_map_message_max": "100",
>>>>>>     "osd_map_share_max_epochs": "100",
>>>>>>     "osd_op_threads": "2",
>>>>>>     "osd_peering_wq_batch_size": "20",
>>>>>>     "osd_op_pq_max_tokens_per_priority": "4194304",
>>>>>>     "osd_op_pq_min_cost": "65536",
>>>>>>     "osd_disk_threads": "1",
>>>>>>     "osd_recovery_threads": "2",
>>>>>>     "osd_recover_clone_overlap": "true",
>>>>>>     "osd_backfill_scan_min": "64",
>>>>>>     "osd_backfill_scan_max": "512",
>>>>>>     "osd_op_thread_timeout": "15",
>>>>>>     "osd_recovery_thread_timeout": "30",
>>>>>>     "osd_snap_trim_thread_timeout": "3600",
>>>>>>     "osd_scrub_thread_timeout": "60",
>>>>>>     "osd_scrub_finalize_thread_timeout": "600",
>>>>>>     "osd_remove_thread_timeout": "3600",
>>>>>>     "osd_command_thread_timeout": "600",
>>>>>>     "osd_age": "0.8",
>>>>>>     "osd_age_time": "0",
>>>>>>     "osd_heartbeat_addr": ":\/0",
>>>>>>     "osd_heartbeat_interval": "6",
>>>>>>     "osd_heartbeat_grace": "20",
>>>>>>     "osd_mon_heartbeat_interval": "30",
>>>>>>     "osd_mon_report_interval_max": "120",
>>>>>>     "osd_mon_report_interval_min": "5",
>>>>>>     "osd_pg_stat_report_interval_max": "500",
>>>>>>     "osd_mon_ack_timeout": "30",
>>>>>>     "osd_min_down_reporters": "1",
>>>>>>     "osd_min_down_reports": "3",
>>>>>>     "osd_default_data_pool_replay_window": "45",
>>>>>>     "osd_preserve_trimmed_log": "false",
>>>>>>     "osd_auto_mark_unfound_lost": "false",
>>>>>>     "osd_recovery_delay_start": "0",
>>>>>>     "osd_recovery_max_active": "5",
>>>>>>     "osd_recovery_max_chunk": "8388608",
>>>>>>     "osd_recovery_forget_lost_objects": "false",
>>>>>>     "osd_max_scrubs": "1",
>>>>>>     "osd_scrub_load_threshold": "0.5",
>>>>>>     "osd_scrub_min_interval": "86400",
>>>>>>     "osd_scrub_max_interval": "604800",
>>>>>>     "osd_deep_scrub_interval": "604800",
>>>>>>     "osd_deep_scrub_stride": "524288",
>>>>>>     "osd_scan_list_ping_tp_interval": "100",
>>>>>>     "osd_auto_weight": "false",
>>>>>>     "osd_class_dir": "\/usr\/lib\/rados-classes",
>>>>>>     "osd_check_for_log_corruption": "false",
>>>>>>     "osd_use_stale_snap": "false",
>>>>>>     "osd_rollback_to_cluster_snap": "",
>>>>>>     "osd_default_notify_timeout": "30",
>>>>>>     "osd_kill_backfill_at": "0",
>>>>>>     "osd_pg_epoch_persisted_max_stale": "200",
>>>>>>     "osd_min_pg_log_entries": "500",
>>>>>>     "osd_max_pg_log_entries": "1500",
>>>>>>     "osd_op_complaint_time": "30",
>>>>>>     "osd_command_max_records": "256",
>>>>>>     "osd_op_log_threshold": "5",
>>>>>>     "osd_verify_sparse_read_holes": "false",
>>>>>>     "osd_debug_drop_ping_probability": "0",
>>>>>>     "osd_debug_drop_ping_duration": "0",
>>>>>>     "osd_debug_drop_pg_create_probability": "0",
>>>>>>     "osd_debug_drop_pg_create_duration": "1",
>>>>>>     "osd_debug_drop_op_probability": "0",
>>>>>>     "osd_debug_op_order": "false",
>>>>>>     "osd_debug_verify_snaps_on_info": "false",
>>>>>>     "osd_debug_verify_stray_on_activate": "false",
>>>>>>     "osd_debug_skip_full_check_in_backfill_reservation": "false",
>>>>>>     "osd_op_history_size": "20",
>>>>>>     "osd_op_history_duration": "600",
>>>>>>     "osd_target_transaction_size": "30",
>>>>>>     "osd_failsafe_full_ratio": "0.97",
>>>>>>     "osd_failsafe_nearfull_ratio": "0.9",
>>>>>>     "osd_leveldb_write_buffer_size": "0",
>>>>>>     "osd_leveldb_cache_size": "0",
>>>>>>     "osd_leveldb_block_size": "0",
>>>>>>     "osd_leveldb_bloom_size": "0",
>>>>>>     "osd_leveldb_max_open_files": "0",
>>>>>>     "osd_leveldb_compression": "true",
>>>>>>     "osd_leveldb_paranoid": "false",
>>>>>>     "osd_leveldb_log": "",
>>>>>>     "osd_client_op_priority": "63",
>>>>>>     "osd_recovery_op_priority": "50",
>>>>>>     "osd_mon_shutdown_timeout": "5",
>>>>>>     "filestore": "false",
>>>>>>     "filestore_index_retry_probability": "0",
>>>>>>     "filestore_debug_inject_read_err": "false",
>>>>>>     "filestore_debug_omap_check": "false",
>>>>>>     "filestore_xattr_use_omap": "false",
>>>>>>     "filestore_max_inline_xattr_size": "512",
>>>>>>     "filestore_max_inline_xattrs": "2",
>>>>>>     "filestore_max_sync_interval": "5",
>>>>>>     "filestore_min_sync_interval": "0.01",
>>>>>>     "filestore_btrfs_snap": "true",
>>>>>>     "filestore_btrfs_clone_range": "true",
>>>>>>     "filestore_fsync_flushes_journal_data": "false",
>>>>>>     "filestore_fiemap": "false",
>>>>>>     "filestore_flusher": "true",
>>>>>>     "filestore_flusher_max_fds": "512",
>>>>>>     "filestore_flush_min": "65536",
>>>>>>     "filestore_sync_flush": "false",
>>>>>>     "filestore_journal_parallel": "false",
>>>>>>     "filestore_journal_writeahead": "false",
>>>>>>     "filestore_journal_trailing": "false",
>>>>>>     "filestore_queue_max_ops": "500",
>>>>>>     "filestore_queue_max_bytes": "104857600",
>>>>>>     "filestore_queue_committing_max_ops": "5000",
>>>>>>     "filestore_queue_committing_max_bytes": "104857600",
>>>>>>     "filestore_op_threads": "2",
>>>>>>     "filestore_op_thread_timeout": "60",
>>>>>>     "filestore_op_thread_suicide_timeout": "180",
>>>>>>     "filestore_commit_timeout": "600",
>>>>>>     "filestore_fiemap_threshold": "4096",
>>>>>>     "filestore_merge_threshold": "10",
>>>>>>     "filestore_split_multiple": "2",
>>>>>>     "filestore_update_to": "1000",
>>>>>>     "filestore_blackhole": "false",
>>>>>>     "filestore_dump_file": "",
>>>>>>     "filestore_kill_at": "0",
>>>>>>     "filestore_inject_stall": "0",
>>>>>>     "filestore_fail_eio": "true",
>>>>>>     "filestore_replica_fadvise": "true",
>>>>>>     "filestore_debug_verify_split": "false",
>>>>>>     "journal_dio": "true",
>>>>>>     "journal_aio": "true",
>>>>>>     "journal_force_aio": "false",
>>>>>>     "journal_max_corrupt_search": "10485760",
>>>>>>     "journal_block_align": "true",
>>>>>>     "journal_write_header_frequency": "0",
>>>>>>     "journal_max_write_bytes": "10485760",
>>>>>>     "journal_max_write_entries": "100",
>>>>>>     "journal_queue_max_ops": "300",
>>>>>>     "journal_queue_max_bytes": "33554432",
>>>>>>     "journal_align_min_size": "65536",
>>>>>>     "journal_replay_from": "0",
>>>>>>     "journal_zero_on_create": "false",
>>>>>>     "journal_ignore_corruption": "false",
>>>>>>     "rbd_cache": "false",
>>>>>>     "rbd_cache_writethrough_until_flush": "false",
>>>>>>     "rbd_cache_size": "33554432",
>>>>>>     "rbd_cache_max_dirty": "25165824",
>>>>>>     "rbd_cache_target_dirty": "16777216",
>>>>>>     "rbd_cache_max_dirty_age": "1",
>>>>>>     "rbd_cache_block_writes_upfront": "false",
>>>>>>     "rbd_concurrent_management_ops": "10",
>>>>>>     "rbd_default_format": "1",
>>>>>>     "rbd_default_order": "22",
>>>>>>     "rbd_default_stripe_count": "1",
>>>>>>     "rbd_default_stripe_unit": "4194304",
>>>>>>     "rbd_default_features": "3",
>>>>>>     "nss_db_path": "",
>>>>>>     "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
>>>>>>     "rgw_enable_apis": "s3, swift, swift_auth, admin",
>>>>>>     "rgw_cache_enabled": "true",
>>>>>>     "rgw_cache_lru_size": "10000",
>>>>>>     "rgw_socket_path": "",
>>>>>>     "rgw_host": "",
>>>>>>     "rgw_port": "",
>>>>>>     "rgw_dns_name": "",
>>>>>>     "rgw_script_uri": "",
>>>>>>     "rgw_request_uri": "",
>>>>>>     "rgw_swift_url": "",
>>>>>>     "rgw_swift_url_prefix": "swift",
>>>>>>     "rgw_swift_auth_url": "",
>>>>>>     "rgw_swift_auth_entry": "auth",
>>>>>>     "rgw_keystone_url": "",
>>>>>>     "rgw_keystone_admin_token": "",
>>>>>>     "rgw_keystone_accepted_roles": "Member, admin",
>>>>>>     "rgw_keystone_token_cache_size": "10000",
>>>>>>     "rgw_keystone_revocation_interval": "900",
>>>>>>     "rgw_admin_entry": "admin",
>>>>>>     "rgw_enforce_swift_acls": "true",
>>>>>>     "rgw_swift_token_expiration": "86400",
>>>>>>     "rgw_print_continue": "true",
>>>>>>     "rgw_remote_addr_param": "REMOTE_ADDR",
>>>>>>     "rgw_op_thread_timeout": "600",
>>>>>>     "rgw_op_thread_suicide_timeout": "0",
>>>>>>     "rgw_thread_pool_size": "100",
>>>>>>     "rgw_num_control_oids": "8",
>>>>>>     "rgw_zone_root_pool": ".rgw.root",
>>>>>>     "rgw_log_nonexistent_bucket": "false",
>>>>>>     "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
>>>>>>     "rgw_log_object_name_utc": "false",
>>>>>>     "rgw_usage_max_shards": "32",
>>>>>>     "rgw_usage_max_user_shards": "1",
>>>>>>     "rgw_enable_ops_log": "false",
>>>>>>     "rgw_enable_usage_log": "false",
>>>>>>     "rgw_ops_log_rados": "true",
>>>>>>     "rgw_ops_log_socket_path": "",
>>>>>>     "rgw_ops_log_data_backlog": "5242880",
>>>>>>     "rgw_usage_log_flush_threshold": "1024",
>>>>>>     "rgw_usage_log_tick_interval": "30",
>>>>>>     "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
>>>>>>     "rgw_intent_log_object_name_utc": "false",
>>>>>>     "rgw_init_timeout": "300",
>>>>>>     "rgw_mime_types_file": "\/etc\/mime.types",
>>>>>>     "rgw_gc_max_objs": "32",
>>>>>>     "rgw_gc_obj_min_wait": "7200",
>>>>>>     "rgw_gc_processor_max_time": "3600",
>>>>>>     "rgw_gc_processor_period": "3600",
>>>>>>     "rgw_s3_success_create_obj_status": "0",
>>>>>>     "rgw_resolve_cname": "false",
>>>>>>     "rgw_obj_stripe_size": "4194304",
>>>>>>     "rgw_extended_http_attrs": "",
>>>>>>     "rgw_exit_timeout_secs": "120",
>>>>>>     "rgw_get_obj_window_size": "16777216",
>>>>>>     "rgw_get_obj_max_req_size": "4194304",
>>>>>>     "rgw_relaxed_s3_bucket_names": "false",
>>>>>>     "rgw_list_buckets_max_chunk": "1000",
>>>>>>     "mutex_perf_counter": "false",
>>>>>>     "internal_safe_to_start_threads": "true"}
>>>>>>
>>>>>>
>>>>>>
>>>>>> Stefan
>>>>>>
>>>>>>
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe
>>>>>>> <s.priebe@profihost.ag>
>>>>>>> wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>> Mike we already have the async patch running. Yes it helps but 
>>>>>>>> only
>>>>>>>> helps
>>>>>>>> it
>>>>>>>> does not solve. It just hides the issue ...
>>>>>>>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>>>>>>>
>>>>>>>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>>>>>>>
>>>>>>>>> - ceph osd set noout
>>>>>>>>>
>>>>>>>>> - Reboot one of the nodes hosting OSDs
>>>>>>>>>         - VMs mounted from RBD volumes work properly
>>>>>>>>>
>>>>>>>>> - I see the OSD's boot messages as they re-join the cluster
>>>>>>>>>
>>>>>>>>> - Start seeing active+recovery_wait, peering, and 
>>>>>>>>> active+recovering
>>>>>>>>>         - VMs mounted from RBD volumes become unresponsive.
>>>>>>>>>
>>>>>>>>> - Recovery completes
>>>>>>>>>         - VMs mounted from RBD volumes regain responsiveness
>>>>>>>>>
>>>>>>>>> - ceph osd unset noout
>>>>>>>>>
>>>>>>>>> Would joshd's async patch for qemu help here, or is there 
>>>>>>>>> something
>>>>>>>>> else
>>>>>>>>> going on?
>>>>>>>>>
>>>>>>>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>>
>>>>>>>>> Mike Dawson
>>>>>>>>> Co-Founder & Director of Cloud Architecture
>>>>>>>>> Cloudapt LLC
>>>>>>>>> 6330 East 75th Street, Suite 170
>>>>>>>>> Indianapolis, IN 46250
>>>>>>>>>
>>>>>>>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Can you reproduce and attach the ceph.log from before you 
>>>>>>>>>> stop the
>>>>>>>>>> osd
>>>>>>>>>> until after you have started the osd and it has recovered?
>>>>>>>>>> -Sam
>>>>>>>>>>
>>>>>>>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> i still have recovery issues with cuttlefish. After the OSD 
>>>>>>>>>>> comes
>>>>>>>>>>> back
>>>>>>>>>>> it seem to hang for around 2-4 minutes and then recovery 
>>>>>>>>>>> seems to
>>>>>>>>>>> start
>>>>>>>>>>> (pgs in recovery_wait start to decrement). This is with ceph
>>>>>>>>>>> 0.61.7.
>>>>>>>>>>> I
>>>>>>>>>>> get a lot of slow request messages an hanging VMs.
>>>>>>>>>>>
>>>>>>>>>>> What i noticed today is that if i leave the OSD off as long as
>>>>>>>>>>> ceph
>>>>>>>>>>> starts to backfill - the recovery and "re" backfilling wents
>>>>>>>>>>> absolutely
>>>>>>>>>>> smooth without any issues and no slow request messages at all.
>>>>>>>>>>>
>>>>>>>>>>> Does anybody have an idea why?
>>>>>>>>>>>
>>>>>>>>>>> Greets,
>>>>>>>>>>> Stefan
>>>>>>>>>>> -- 
>>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>>> ceph-devel"
>>>>>>>>>>> in
>>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>>> More majordomo info at
>>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> -- 
>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>> ceph-devel"
>>>>>>>>>> in
>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>> More majordomo info at 
>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>>
>>>>>>>>
>>>>>>> -- 
>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>> ceph-devel" in
>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html
>>>>>>>
>>>>>>
>>>>
> -- 
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


-- 

Oliver Francke

filoo GmbH
Moltkestraße 25a
33330 Gütersloh
HRB4355 AG Gütersloh

Geschäftsführer: J.Rehpöhler | C.Kunz

Folgen Sie uns auf Twitter: http://twitter.com/filoogmbh

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-08 14:05                   ` Mike Dawson
  2013-08-08 15:43                     ` Oliver Francke
@ 2013-08-08 18:13                     ` Stefan Priebe
  2013-08-09 21:44                       ` Samuel Just
  1 sibling, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-08 18:13 UTC (permalink / raw)
  To: Mike Dawson
  Cc: Samuel Just, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

Hi Mike,

Am 08.08.2013 16:05, schrieb Mike Dawson:
> Stefan,
>
> I see the same behavior and I theorize it is linked to an issue detailed
> in another thread [0]. Do your VM guests ever hang while your cluster is
> HEALTH_OK like described in that other thread?
>
> [0] http://comments.gmane.org/gmane.comp.file-systems.ceph.user/2982

mhm no can't see that. All our VMs are working fine even under high load 
while ceph is OK.

> A few observations:
>
> - The VMs that hang do lots of writes (video surveillance).
> - I use rbd and qemu. The problem exists in both qemu 1.4.x and 1.5.2.
> - The problem exists with or without joshd's qemu async flush patch.
> - Windows VMs seem to be more vulnerable than Linux VMs.
> - If I restart the qemu-system-x86_64 process, the guest will come back
> to life.
> - A partial workaround seems to be console input (NoVNC or 'virsh
> screenshot'), but restarting qemu-system-x86_64 works better.
> - The issue of VMs hanging seems worse with RBD writeback cache enabled
> - I typically run virtio, but I believe I've seen it with e1000, too.
> - VM guests hang at different times, not all at once on a host (or
> across all hosts).
> - I co-mingle VM guests on servers that host ceph OSDs.
>
>
>
> Oliver,
>
> If your cluster has to recover/backfill, do your guest VMs hang with
> more frequency than under normal HEALTH_OK conditions, even if you
> prioritize client IO as Sam wrote below?
>
>
> Sam,
>
> Turning down all the settings you mentioned certainly does slow the
> recover/backfill process, but it doesn't prevent the VM guests backed by
> RBD volumes from hanging. In fact, I often try to prioritize
> recovery/backfill because my guests tend to hang until I get back to
> HEALTH_OK. Given this apparent bug, completing recovery/backfill quicker
> leads to less total outage, it seems.
>
>
> Josh,
>
> How can I help you investigate if RBD is the common source of both of
> these issues?
>
>
> Thanks,
> Mike Dawson
>
>
> On 8/2/2013 2:46 PM, Stefan Priebe wrote:
>> Hi,
>>
>>          osd recovery max active = 1
>>          osd max backfills = 1
>>          osd recovery op priority = 5
>>
>> still no difference...
>>
>> Stefan
>>
>> Am 02.08.2013 20:21, schrieb Samuel Just:
>>> Also, you have osd_recovery_op_priority at 50.  That is close to the
>>> priority of client IO.  You want it below 10 (defaults to 10), perhaps
>>> at 1.  You can also adjust down osd_recovery_max_active.
>>> -Sam
>>>
>>> On Fri, Aug 2, 2013 at 11:16 AM, Stefan Priebe <s.priebe@profihost.ag>
>>> wrote:
>>>> I already tried both values this makes no difference. The drives are
>>>> not the
>>>> bottleneck.
>>>>
>>>> Am 02.08.2013 19:35, schrieb Samuel Just:
>>>>
>>>>> You might try turning osd_max_backfills to 2 or 1.
>>>>> -Sam
>>>>>
>>>>> On Fri, Aug 2, 2013 at 12:44 AM, Stefan Priebe <s.priebe@profihost.ag>
>>>>> wrote:
>>>>>>
>>>>>> Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd
>>>>>> settings?
>>>>>>
>>>>>>> sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
>>>>>>
>>>>>>
>>>>>> Sure.
>>>>>>
>>>>>>
>>>>>>
>>>>>> { "name": "osd.0",
>>>>>>     "cluster": "ceph",
>>>>>>     "none": "0\/5",
>>>>>>     "lockdep": "0\/0",
>>>>>>     "context": "0\/0",
>>>>>>     "crush": "0\/0",
>>>>>>     "mds": "0\/0",
>>>>>>     "mds_balancer": "0\/0",
>>>>>>     "mds_locker": "0\/0",
>>>>>>     "mds_log": "0\/0",
>>>>>>     "mds_log_expire": "0\/0",
>>>>>>     "mds_migrator": "0\/0",
>>>>>>     "buffer": "0\/0",
>>>>>>     "timer": "0\/0",
>>>>>>     "filer": "0\/0",
>>>>>>     "striper": "0\/1",
>>>>>>     "objecter": "0\/0",
>>>>>>     "rados": "0\/0",
>>>>>>     "rbd": "0\/0",
>>>>>>     "journaler": "0\/0",
>>>>>>     "objectcacher": "0\/0",
>>>>>>     "client": "0\/0",
>>>>>>     "osd": "0\/0",
>>>>>>     "optracker": "0\/0",
>>>>>>     "objclass": "0\/0",
>>>>>>     "filestore": "0\/0",
>>>>>>     "journal": "0\/0",
>>>>>>     "ms": "0\/0",
>>>>>>     "mon": "0\/0",
>>>>>>     "monc": "0\/0",
>>>>>>     "paxos": "0\/0",
>>>>>>     "tp": "0\/0",
>>>>>>     "auth": "0\/0",
>>>>>>     "crypto": "1\/5",
>>>>>>     "finisher": "0\/0",
>>>>>>     "heartbeatmap": "0\/0",
>>>>>>     "perfcounter": "0\/0",
>>>>>>     "rgw": "0\/0",
>>>>>>     "hadoop": "0\/0",
>>>>>>     "javaclient": "1\/5",
>>>>>>     "asok": "0\/0",
>>>>>>     "throttle": "0\/0",
>>>>>>     "host": "cloud1-1268",
>>>>>>     "fsid": "00000000-0000-0000-0000-000000000000",
>>>>>>     "public_addr": "10.255.0.90:0\/0",
>>>>>>     "cluster_addr": "10.255.0.90:0\/0",
>>>>>>     "public_network": "10.255.0.1\/24",
>>>>>>     "cluster_network": "10.255.0.1\/24",
>>>>>>     "num_client": "1",
>>>>>>     "monmap": "",
>>>>>>     "mon_host": "",
>>>>>>     "lockdep": "false",
>>>>>>     "run_dir": "\/var\/run\/ceph",
>>>>>>     "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
>>>>>>     "daemonize": "true",
>>>>>>     "pid_file": "\/var\/run\/ceph\/osd.0.pid",
>>>>>>     "chdir": "\/",
>>>>>>     "max_open_files": "0",
>>>>>>     "fatal_signal_handlers": "true",
>>>>>>     "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
>>>>>>     "log_max_new": "1000",
>>>>>>     "log_max_recent": "10000",
>>>>>>     "log_to_stderr": "false",
>>>>>>     "err_to_stderr": "true",
>>>>>>     "log_to_syslog": "false",
>>>>>>     "err_to_syslog": "false",
>>>>>>     "log_flush_on_exit": "true",
>>>>>>     "log_stop_at_utilization": "0.97",
>>>>>>     "clog_to_monitors": "true",
>>>>>>     "clog_to_syslog": "false",
>>>>>>     "clog_to_syslog_level": "info",
>>>>>>     "clog_to_syslog_facility": "daemon",
>>>>>>     "mon_cluster_log_to_syslog": "false",
>>>>>>     "mon_cluster_log_to_syslog_level": "info",
>>>>>>     "mon_cluster_log_to_syslog_facility": "daemon",
>>>>>>     "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
>>>>>>     "key": "",
>>>>>>     "keyfile": "",
>>>>>>     "keyring": "\/etc\/ceph\/osd.0.keyring",
>>>>>>     "heartbeat_interval": "5",
>>>>>>     "heartbeat_file": "",
>>>>>>     "heartbeat_inject_failure": "0",
>>>>>>     "perf": "true",
>>>>>>     "ms_tcp_nodelay": "true",
>>>>>>     "ms_tcp_rcvbuf": "0",
>>>>>>     "ms_initial_backoff": "0.2",
>>>>>>     "ms_max_backoff": "15",
>>>>>>     "ms_nocrc": "false",
>>>>>>     "ms_die_on_bad_msg": "false",
>>>>>>     "ms_die_on_unhandled_msg": "false",
>>>>>>     "ms_dispatch_throttle_bytes": "104857600",
>>>>>>     "ms_bind_ipv6": "false",
>>>>>>     "ms_bind_port_min": "6800",
>>>>>>     "ms_bind_port_max": "7100",
>>>>>>     "ms_rwthread_stack_bytes": "1048576",
>>>>>>     "ms_tcp_read_timeout": "900",
>>>>>>     "ms_pq_max_tokens_per_priority": "4194304",
>>>>>>     "ms_pq_min_cost": "65536",
>>>>>>     "ms_inject_socket_failures": "0",
>>>>>>     "ms_inject_delay_type": "",
>>>>>>     "ms_inject_delay_max": "1",
>>>>>>     "ms_inject_delay_probability": "0",
>>>>>>     "ms_inject_internal_delays": "0",
>>>>>>     "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
>>>>>>     "mon_initial_members": "",
>>>>>>     "mon_sync_fs_threshold": "5",
>>>>>>     "mon_compact_on_start": "false",
>>>>>>     "mon_compact_on_bootstrap": "false",
>>>>>>     "mon_compact_on_trim": "true",
>>>>>>     "mon_tick_interval": "5",
>>>>>>     "mon_subscribe_interval": "300",
>>>>>>     "mon_osd_laggy_halflife": "3600",
>>>>>>     "mon_osd_laggy_weight": "0.3",
>>>>>>     "mon_osd_adjust_heartbeat_grace": "true",
>>>>>>     "mon_osd_adjust_down_out_interval": "true",
>>>>>>     "mon_osd_auto_mark_in": "false",
>>>>>>     "mon_osd_auto_mark_auto_out_in": "true",
>>>>>>     "mon_osd_auto_mark_new_in": "true",
>>>>>>     "mon_osd_down_out_interval": "300",
>>>>>>     "mon_osd_down_out_subtree_limit": "rack",
>>>>>>     "mon_osd_min_up_ratio": "0.3",
>>>>>>     "mon_osd_min_in_ratio": "0.3",
>>>>>>     "mon_stat_smooth_intervals": "2",
>>>>>>     "mon_lease": "5",
>>>>>>     "mon_lease_renew_interval": "3",
>>>>>>     "mon_lease_ack_timeout": "10",
>>>>>>     "mon_clock_drift_allowed": "0.05",
>>>>>>     "mon_clock_drift_warn_backoff": "5",
>>>>>>     "mon_timecheck_interval": "300",
>>>>>>     "mon_accept_timeout": "10",
>>>>>>     "mon_pg_create_interval": "30",
>>>>>>     "mon_pg_stuck_threshold": "300",
>>>>>>     "mon_osd_full_ratio": "0.95",
>>>>>>     "mon_osd_nearfull_ratio": "0.85",
>>>>>>     "mon_globalid_prealloc": "100",
>>>>>>     "mon_osd_report_timeout": "900",
>>>>>>     "mon_force_standby_active": "true",
>>>>>>     "mon_min_osdmap_epochs": "500",
>>>>>>     "mon_max_pgmap_epochs": "500",
>>>>>>     "mon_max_log_epochs": "500",
>>>>>>     "mon_max_osd": "10000",
>>>>>>     "mon_probe_timeout": "2",
>>>>>>     "mon_slurp_timeout": "10",
>>>>>>     "mon_slurp_bytes": "262144",
>>>>>>     "mon_client_bytes": "104857600",
>>>>>>     "mon_daemon_bytes": "419430400",
>>>>>>     "mon_max_log_entries_per_event": "4096",
>>>>>>     "mon_health_data_update_interval": "60",
>>>>>>     "mon_data_avail_crit": "5",
>>>>>>     "mon_data_avail_warn": "30",
>>>>>>     "mon_config_key_max_entry_size": "4096",
>>>>>>     "mon_sync_trim_timeout": "30",
>>>>>>     "mon_sync_heartbeat_timeout": "30",
>>>>>>     "mon_sync_heartbeat_interval": "5",
>>>>>>     "mon_sync_backoff_timeout": "30",
>>>>>>     "mon_sync_timeout": "30",
>>>>>>     "mon_sync_max_retries": "5",
>>>>>>     "mon_sync_max_payload_size": "1048576",
>>>>>>     "mon_sync_debug": "false",
>>>>>>     "mon_sync_debug_leader": "-1",
>>>>>>     "mon_sync_debug_provider": "-1",
>>>>>>     "mon_sync_debug_provider_fallback": "-1",
>>>>>>     "mon_debug_dump_transactions": "false",
>>>>>>     "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
>>>>>>     "mon_sync_leader_kill_at": "0",
>>>>>>     "mon_sync_provider_kill_at": "0",
>>>>>>     "mon_sync_requester_kill_at": "0",
>>>>>>     "mon_leveldb_write_buffer_size": "33554432",
>>>>>>     "mon_leveldb_cache_size": "268435456",
>>>>>>     "mon_leveldb_block_size": "65536",
>>>>>>     "mon_leveldb_bloom_size": "0",
>>>>>>     "mon_leveldb_max_open_files": "0",
>>>>>>     "mon_leveldb_compression": "false",
>>>>>>     "mon_leveldb_paranoid": "false",
>>>>>>     "mon_leveldb_log": "",
>>>>>>     "paxos_stash_full_interval": "25",
>>>>>>     "paxos_max_join_drift": "100",
>>>>>>     "paxos_propose_interval": "1",
>>>>>>     "paxos_min_wait": "0.05",
>>>>>>     "paxos_min": "500",
>>>>>>     "paxos_trim_min": "500",
>>>>>>     "paxos_trim_max": "1000",
>>>>>>     "paxos_trim_disabled_max_versions": "108000",
>>>>>>     "paxos_service_trim_min": "500",
>>>>>>     "paxos_service_trim_max": "1000",
>>>>>>     "clock_offset": "0",
>>>>>>     "auth_cluster_required": "none",
>>>>>>     "auth_service_required": "none",
>>>>>>     "auth_client_required": "none",
>>>>>>     "auth_supported": "none",
>>>>>>     "cephx_require_signatures": "false",
>>>>>>     "cephx_cluster_require_signatures": "false",
>>>>>>     "cephx_service_require_signatures": "false",
>>>>>>     "cephx_sign_messages": "true",
>>>>>>     "auth_mon_ticket_ttl": "43200",
>>>>>>     "auth_service_ticket_ttl": "3600",
>>>>>>     "auth_debug": "false",
>>>>>>     "mon_client_hunt_interval": "3",
>>>>>>     "mon_client_ping_interval": "10",
>>>>>>     "mon_client_max_log_entries_per_message": "1000",
>>>>>>     "mon_max_pool_pg_num": "65536",
>>>>>>     "mon_pool_quota_warn_threshold": "0",
>>>>>>     "mon_pool_quota_crit_threshold": "0",
>>>>>>     "client_cache_size": "16384",
>>>>>>     "client_cache_mid": "0.75",
>>>>>>     "client_use_random_mds": "false",
>>>>>>     "client_mount_timeout": "300",
>>>>>>     "client_tick_interval": "1",
>>>>>>     "client_trace": "",
>>>>>>     "client_readahead_min": "131072",
>>>>>>     "client_readahead_max_bytes": "0",
>>>>>>     "client_readahead_max_periods": "4",
>>>>>>     "client_snapdir": ".snap",
>>>>>>     "client_mountpoint": "\/",
>>>>>>     "client_notify_timeout": "10",
>>>>>>     "client_caps_release_delay": "5",
>>>>>>     "client_oc": "true",
>>>>>>     "client_oc_size": "209715200",
>>>>>>     "client_oc_max_dirty": "104857600",
>>>>>>     "client_oc_target_dirty": "8388608",
>>>>>>     "client_oc_max_dirty_age": "5",
>>>>>>     "client_oc_max_objects": "1000",
>>>>>>     "client_debug_force_sync_read": "false",
>>>>>>     "client_debug_inject_tick_delay": "0",
>>>>>>     "fuse_use_invalidate_cb": "false",
>>>>>>     "fuse_allow_other": "true",
>>>>>>     "fuse_default_permissions": "true",
>>>>>>     "fuse_big_writes": "true",
>>>>>>     "fuse_atomic_o_trunc": "true",
>>>>>>     "fuse_debug": "false",
>>>>>>     "objecter_tick_interval": "5",
>>>>>>     "objecter_timeout": "10",
>>>>>>     "objecter_inflight_op_bytes": "104857600",
>>>>>>     "objecter_inflight_ops": "1024",
>>>>>>     "journaler_allow_split_entries": "true",
>>>>>>     "journaler_write_head_interval": "15",
>>>>>>     "journaler_prefetch_periods": "10",
>>>>>>     "journaler_prezero_periods": "5",
>>>>>>     "journaler_batch_interval": "0.001",
>>>>>>     "journaler_batch_max": "0",
>>>>>>     "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
>>>>>>     "mds_max_file_size": "1099511627776",
>>>>>>     "mds_cache_size": "100000",
>>>>>>     "mds_cache_mid": "0.7",
>>>>>>     "mds_mem_max": "1048576",
>>>>>>     "mds_dir_commit_ratio": "0.5",
>>>>>>     "mds_dir_max_commit_size": "90",
>>>>>>     "mds_decay_halflife": "5",
>>>>>>     "mds_beacon_interval": "4",
>>>>>>     "mds_beacon_grace": "15",
>>>>>>     "mds_enforce_unique_name": "true",
>>>>>>     "mds_blacklist_interval": "1440",
>>>>>>     "mds_session_timeout": "60",
>>>>>>     "mds_session_autoclose": "300",
>>>>>>     "mds_reconnect_timeout": "45",
>>>>>>     "mds_tick_interval": "5",
>>>>>>     "mds_dirstat_min_interval": "1",
>>>>>>     "mds_scatter_nudge_interval": "5",
>>>>>>     "mds_client_prealloc_inos": "1000",
>>>>>>     "mds_early_reply": "true",
>>>>>>     "mds_use_tmap": "true",
>>>>>>     "mds_default_dir_hash": "2",
>>>>>>     "mds_log": "true",
>>>>>>     "mds_log_skip_corrupt_events": "false",
>>>>>>     "mds_log_max_events": "-1",
>>>>>>     "mds_log_segment_size": "0",
>>>>>>     "mds_log_max_segments": "30",
>>>>>>     "mds_log_max_expiring": "20",
>>>>>>     "mds_bal_sample_interval": "3",
>>>>>>     "mds_bal_replicate_threshold": "8000",
>>>>>>     "mds_bal_unreplicate_threshold": "0",
>>>>>>     "mds_bal_frag": "false",
>>>>>>     "mds_bal_split_size": "10000",
>>>>>>     "mds_bal_split_rd": "25000",
>>>>>>     "mds_bal_split_wr": "10000",
>>>>>>     "mds_bal_split_bits": "3",
>>>>>>     "mds_bal_merge_size": "50",
>>>>>>     "mds_bal_merge_rd": "1000",
>>>>>>     "mds_bal_merge_wr": "1000",
>>>>>>     "mds_bal_interval": "10",
>>>>>>     "mds_bal_fragment_interval": "5",
>>>>>>     "mds_bal_idle_threshold": "0",
>>>>>>     "mds_bal_max": "-1",
>>>>>>     "mds_bal_max_until": "-1",
>>>>>>     "mds_bal_mode": "0",
>>>>>>     "mds_bal_min_rebalance": "0.1",
>>>>>>     "mds_bal_min_start": "0.2",
>>>>>>     "mds_bal_need_min": "0.8",
>>>>>>     "mds_bal_need_max": "1.2",
>>>>>>     "mds_bal_midchunk": "0.3",
>>>>>>     "mds_bal_minchunk": "0.001",
>>>>>>     "mds_bal_target_removal_min": "5",
>>>>>>     "mds_bal_target_removal_max": "10",
>>>>>>     "mds_replay_interval": "1",
>>>>>>     "mds_shutdown_check": "0",
>>>>>>     "mds_thrash_exports": "0",
>>>>>>     "mds_thrash_fragments": "0",
>>>>>>     "mds_dump_cache_on_map": "false",
>>>>>>     "mds_dump_cache_after_rejoin": "false",
>>>>>>     "mds_verify_scatter": "false",
>>>>>>     "mds_debug_scatterstat": "false",
>>>>>>     "mds_debug_frag": "false",
>>>>>>     "mds_debug_auth_pins": "false",
>>>>>>     "mds_debug_subtrees": "false",
>>>>>>     "mds_kill_mdstable_at": "0",
>>>>>>     "mds_kill_export_at": "0",
>>>>>>     "mds_kill_import_at": "0",
>>>>>>     "mds_kill_link_at": "0",
>>>>>>     "mds_kill_rename_at": "0",
>>>>>>     "mds_kill_openc_at": "0",
>>>>>>     "mds_kill_journal_at": "0",
>>>>>>     "mds_kill_journal_expire_at": "0",
>>>>>>     "mds_kill_journal_replay_at": "0",
>>>>>>     "mds_inject_traceless_reply_probability": "0",
>>>>>>     "mds_wipe_sessions": "false",
>>>>>>     "mds_wipe_ino_prealloc": "false",
>>>>>>     "mds_skip_ino": "0",
>>>>>>     "max_mds": "1",
>>>>>>     "mds_standby_for_name": "",
>>>>>>     "mds_standby_for_rank": "-1",
>>>>>>     "mds_standby_replay": "false",
>>>>>>     "osd_auto_upgrade_tmap": "true",
>>>>>>     "osd_tmapput_sets_uses_tmap": "false",
>>>>>>     "osd_max_backfills": "5",
>>>>>>     "osd_backfill_full_ratio": "0.85",
>>>>>>     "osd_backfill_retry_interval": "10",
>>>>>>     "osd_uuid": "00000000-0000-0000-0000-000000000000",
>>>>>>     "osd_data": "\/ceph\/osd.0\/",
>>>>>>     "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
>>>>>>     "osd_journal_size": "5120",
>>>>>>     "osd_max_write_size": "90",
>>>>>>     "osd_max_pgls": "1024",
>>>>>>     "osd_client_message_size_cap": "524288000",
>>>>>>     "osd_client_message_cap": "100",
>>>>>>     "osd_pg_bits": "6",
>>>>>>     "osd_pgp_bits": "6",
>>>>>>     "osd_crush_chooseleaf_type": "1",
>>>>>>     "osd_min_rep": "1",
>>>>>>     "osd_max_rep": "10",
>>>>>>     "osd_pool_default_crush_rule": "0",
>>>>>>     "osd_pool_default_size": "2",
>>>>>>     "osd_pool_default_min_size": "0",
>>>>>>     "osd_pool_default_pg_num": "8",
>>>>>>     "osd_pool_default_pgp_num": "8",
>>>>>>     "osd_pool_default_flags": "0",
>>>>>>     "osd_map_dedup": "true",
>>>>>>     "osd_map_cache_size": "500",
>>>>>>     "osd_map_message_max": "100",
>>>>>>     "osd_map_share_max_epochs": "100",
>>>>>>     "osd_op_threads": "2",
>>>>>>     "osd_peering_wq_batch_size": "20",
>>>>>>     "osd_op_pq_max_tokens_per_priority": "4194304",
>>>>>>     "osd_op_pq_min_cost": "65536",
>>>>>>     "osd_disk_threads": "1",
>>>>>>     "osd_recovery_threads": "2",
>>>>>>     "osd_recover_clone_overlap": "true",
>>>>>>     "osd_backfill_scan_min": "64",
>>>>>>     "osd_backfill_scan_max": "512",
>>>>>>     "osd_op_thread_timeout": "15",
>>>>>>     "osd_recovery_thread_timeout": "30",
>>>>>>     "osd_snap_trim_thread_timeout": "3600",
>>>>>>     "osd_scrub_thread_timeout": "60",
>>>>>>     "osd_scrub_finalize_thread_timeout": "600",
>>>>>>     "osd_remove_thread_timeout": "3600",
>>>>>>     "osd_command_thread_timeout": "600",
>>>>>>     "osd_age": "0.8",
>>>>>>     "osd_age_time": "0",
>>>>>>     "osd_heartbeat_addr": ":\/0",
>>>>>>     "osd_heartbeat_interval": "6",
>>>>>>     "osd_heartbeat_grace": "20",
>>>>>>     "osd_mon_heartbeat_interval": "30",
>>>>>>     "osd_mon_report_interval_max": "120",
>>>>>>     "osd_mon_report_interval_min": "5",
>>>>>>     "osd_pg_stat_report_interval_max": "500",
>>>>>>     "osd_mon_ack_timeout": "30",
>>>>>>     "osd_min_down_reporters": "1",
>>>>>>     "osd_min_down_reports": "3",
>>>>>>     "osd_default_data_pool_replay_window": "45",
>>>>>>     "osd_preserve_trimmed_log": "false",
>>>>>>     "osd_auto_mark_unfound_lost": "false",
>>>>>>     "osd_recovery_delay_start": "0",
>>>>>>     "osd_recovery_max_active": "5",
>>>>>>     "osd_recovery_max_chunk": "8388608",
>>>>>>     "osd_recovery_forget_lost_objects": "false",
>>>>>>     "osd_max_scrubs": "1",
>>>>>>     "osd_scrub_load_threshold": "0.5",
>>>>>>     "osd_scrub_min_interval": "86400",
>>>>>>     "osd_scrub_max_interval": "604800",
>>>>>>     "osd_deep_scrub_interval": "604800",
>>>>>>     "osd_deep_scrub_stride": "524288",
>>>>>>     "osd_scan_list_ping_tp_interval": "100",
>>>>>>     "osd_auto_weight": "false",
>>>>>>     "osd_class_dir": "\/usr\/lib\/rados-classes",
>>>>>>     "osd_check_for_log_corruption": "false",
>>>>>>     "osd_use_stale_snap": "false",
>>>>>>     "osd_rollback_to_cluster_snap": "",
>>>>>>     "osd_default_notify_timeout": "30",
>>>>>>     "osd_kill_backfill_at": "0",
>>>>>>     "osd_pg_epoch_persisted_max_stale": "200",
>>>>>>     "osd_min_pg_log_entries": "500",
>>>>>>     "osd_max_pg_log_entries": "1500",
>>>>>>     "osd_op_complaint_time": "30",
>>>>>>     "osd_command_max_records": "256",
>>>>>>     "osd_op_log_threshold": "5",
>>>>>>     "osd_verify_sparse_read_holes": "false",
>>>>>>     "osd_debug_drop_ping_probability": "0",
>>>>>>     "osd_debug_drop_ping_duration": "0",
>>>>>>     "osd_debug_drop_pg_create_probability": "0",
>>>>>>     "osd_debug_drop_pg_create_duration": "1",
>>>>>>     "osd_debug_drop_op_probability": "0",
>>>>>>     "osd_debug_op_order": "false",
>>>>>>     "osd_debug_verify_snaps_on_info": "false",
>>>>>>     "osd_debug_verify_stray_on_activate": "false",
>>>>>>     "osd_debug_skip_full_check_in_backfill_reservation": "false",
>>>>>>     "osd_op_history_size": "20",
>>>>>>     "osd_op_history_duration": "600",
>>>>>>     "osd_target_transaction_size": "30",
>>>>>>     "osd_failsafe_full_ratio": "0.97",
>>>>>>     "osd_failsafe_nearfull_ratio": "0.9",
>>>>>>     "osd_leveldb_write_buffer_size": "0",
>>>>>>     "osd_leveldb_cache_size": "0",
>>>>>>     "osd_leveldb_block_size": "0",
>>>>>>     "osd_leveldb_bloom_size": "0",
>>>>>>     "osd_leveldb_max_open_files": "0",
>>>>>>     "osd_leveldb_compression": "true",
>>>>>>     "osd_leveldb_paranoid": "false",
>>>>>>     "osd_leveldb_log": "",
>>>>>>     "osd_client_op_priority": "63",
>>>>>>     "osd_recovery_op_priority": "50",
>>>>>>     "osd_mon_shutdown_timeout": "5",
>>>>>>     "filestore": "false",
>>>>>>     "filestore_index_retry_probability": "0",
>>>>>>     "filestore_debug_inject_read_err": "false",
>>>>>>     "filestore_debug_omap_check": "false",
>>>>>>     "filestore_xattr_use_omap": "false",
>>>>>>     "filestore_max_inline_xattr_size": "512",
>>>>>>     "filestore_max_inline_xattrs": "2",
>>>>>>     "filestore_max_sync_interval": "5",
>>>>>>     "filestore_min_sync_interval": "0.01",
>>>>>>     "filestore_btrfs_snap": "true",
>>>>>>     "filestore_btrfs_clone_range": "true",
>>>>>>     "filestore_fsync_flushes_journal_data": "false",
>>>>>>     "filestore_fiemap": "false",
>>>>>>     "filestore_flusher": "true",
>>>>>>     "filestore_flusher_max_fds": "512",
>>>>>>     "filestore_flush_min": "65536",
>>>>>>     "filestore_sync_flush": "false",
>>>>>>     "filestore_journal_parallel": "false",
>>>>>>     "filestore_journal_writeahead": "false",
>>>>>>     "filestore_journal_trailing": "false",
>>>>>>     "filestore_queue_max_ops": "500",
>>>>>>     "filestore_queue_max_bytes": "104857600",
>>>>>>     "filestore_queue_committing_max_ops": "5000",
>>>>>>     "filestore_queue_committing_max_bytes": "104857600",
>>>>>>     "filestore_op_threads": "2",
>>>>>>     "filestore_op_thread_timeout": "60",
>>>>>>     "filestore_op_thread_suicide_timeout": "180",
>>>>>>     "filestore_commit_timeout": "600",
>>>>>>     "filestore_fiemap_threshold": "4096",
>>>>>>     "filestore_merge_threshold": "10",
>>>>>>     "filestore_split_multiple": "2",
>>>>>>     "filestore_update_to": "1000",
>>>>>>     "filestore_blackhole": "false",
>>>>>>     "filestore_dump_file": "",
>>>>>>     "filestore_kill_at": "0",
>>>>>>     "filestore_inject_stall": "0",
>>>>>>     "filestore_fail_eio": "true",
>>>>>>     "filestore_replica_fadvise": "true",
>>>>>>     "filestore_debug_verify_split": "false",
>>>>>>     "journal_dio": "true",
>>>>>>     "journal_aio": "true",
>>>>>>     "journal_force_aio": "false",
>>>>>>     "journal_max_corrupt_search": "10485760",
>>>>>>     "journal_block_align": "true",
>>>>>>     "journal_write_header_frequency": "0",
>>>>>>     "journal_max_write_bytes": "10485760",
>>>>>>     "journal_max_write_entries": "100",
>>>>>>     "journal_queue_max_ops": "300",
>>>>>>     "journal_queue_max_bytes": "33554432",
>>>>>>     "journal_align_min_size": "65536",
>>>>>>     "journal_replay_from": "0",
>>>>>>     "journal_zero_on_create": "false",
>>>>>>     "journal_ignore_corruption": "false",
>>>>>>     "rbd_cache": "false",
>>>>>>     "rbd_cache_writethrough_until_flush": "false",
>>>>>>     "rbd_cache_size": "33554432",
>>>>>>     "rbd_cache_max_dirty": "25165824",
>>>>>>     "rbd_cache_target_dirty": "16777216",
>>>>>>     "rbd_cache_max_dirty_age": "1",
>>>>>>     "rbd_cache_block_writes_upfront": "false",
>>>>>>     "rbd_concurrent_management_ops": "10",
>>>>>>     "rbd_default_format": "1",
>>>>>>     "rbd_default_order": "22",
>>>>>>     "rbd_default_stripe_count": "1",
>>>>>>     "rbd_default_stripe_unit": "4194304",
>>>>>>     "rbd_default_features": "3",
>>>>>>     "nss_db_path": "",
>>>>>>     "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
>>>>>>     "rgw_enable_apis": "s3, swift, swift_auth, admin",
>>>>>>     "rgw_cache_enabled": "true",
>>>>>>     "rgw_cache_lru_size": "10000",
>>>>>>     "rgw_socket_path": "",
>>>>>>     "rgw_host": "",
>>>>>>     "rgw_port": "",
>>>>>>     "rgw_dns_name": "",
>>>>>>     "rgw_script_uri": "",
>>>>>>     "rgw_request_uri": "",
>>>>>>     "rgw_swift_url": "",
>>>>>>     "rgw_swift_url_prefix": "swift",
>>>>>>     "rgw_swift_auth_url": "",
>>>>>>     "rgw_swift_auth_entry": "auth",
>>>>>>     "rgw_keystone_url": "",
>>>>>>     "rgw_keystone_admin_token": "",
>>>>>>     "rgw_keystone_accepted_roles": "Member, admin",
>>>>>>     "rgw_keystone_token_cache_size": "10000",
>>>>>>     "rgw_keystone_revocation_interval": "900",
>>>>>>     "rgw_admin_entry": "admin",
>>>>>>     "rgw_enforce_swift_acls": "true",
>>>>>>     "rgw_swift_token_expiration": "86400",
>>>>>>     "rgw_print_continue": "true",
>>>>>>     "rgw_remote_addr_param": "REMOTE_ADDR",
>>>>>>     "rgw_op_thread_timeout": "600",
>>>>>>     "rgw_op_thread_suicide_timeout": "0",
>>>>>>     "rgw_thread_pool_size": "100",
>>>>>>     "rgw_num_control_oids": "8",
>>>>>>     "rgw_zone_root_pool": ".rgw.root",
>>>>>>     "rgw_log_nonexistent_bucket": "false",
>>>>>>     "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
>>>>>>     "rgw_log_object_name_utc": "false",
>>>>>>     "rgw_usage_max_shards": "32",
>>>>>>     "rgw_usage_max_user_shards": "1",
>>>>>>     "rgw_enable_ops_log": "false",
>>>>>>     "rgw_enable_usage_log": "false",
>>>>>>     "rgw_ops_log_rados": "true",
>>>>>>     "rgw_ops_log_socket_path": "",
>>>>>>     "rgw_ops_log_data_backlog": "5242880",
>>>>>>     "rgw_usage_log_flush_threshold": "1024",
>>>>>>     "rgw_usage_log_tick_interval": "30",
>>>>>>     "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
>>>>>>     "rgw_intent_log_object_name_utc": "false",
>>>>>>     "rgw_init_timeout": "300",
>>>>>>     "rgw_mime_types_file": "\/etc\/mime.types",
>>>>>>     "rgw_gc_max_objs": "32",
>>>>>>     "rgw_gc_obj_min_wait": "7200",
>>>>>>     "rgw_gc_processor_max_time": "3600",
>>>>>>     "rgw_gc_processor_period": "3600",
>>>>>>     "rgw_s3_success_create_obj_status": "0",
>>>>>>     "rgw_resolve_cname": "false",
>>>>>>     "rgw_obj_stripe_size": "4194304",
>>>>>>     "rgw_extended_http_attrs": "",
>>>>>>     "rgw_exit_timeout_secs": "120",
>>>>>>     "rgw_get_obj_window_size": "16777216",
>>>>>>     "rgw_get_obj_max_req_size": "4194304",
>>>>>>     "rgw_relaxed_s3_bucket_names": "false",
>>>>>>     "rgw_list_buckets_max_chunk": "1000",
>>>>>>     "mutex_perf_counter": "false",
>>>>>>     "internal_safe_to_start_threads": "true"}
>>>>>>
>>>>>>
>>>>>>
>>>>>> Stefan
>>>>>>
>>>>>>
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe
>>>>>>> <s.priebe@profihost.ag>
>>>>>>> wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>> Mike we already have the async patch running. Yes it helps but only
>>>>>>>> helps
>>>>>>>> it
>>>>>>>> does not solve. It just hides the issue ...
>>>>>>>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>>>>>>>
>>>>>>>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>>>>>>>
>>>>>>>>> - ceph osd set noout
>>>>>>>>>
>>>>>>>>> - Reboot one of the nodes hosting OSDs
>>>>>>>>>         - VMs mounted from RBD volumes work properly
>>>>>>>>>
>>>>>>>>> - I see the OSD's boot messages as they re-join the cluster
>>>>>>>>>
>>>>>>>>> - Start seeing active+recovery_wait, peering, and
>>>>>>>>> active+recovering
>>>>>>>>>         - VMs mounted from RBD volumes become unresponsive.
>>>>>>>>>
>>>>>>>>> - Recovery completes
>>>>>>>>>         - VMs mounted from RBD volumes regain responsiveness
>>>>>>>>>
>>>>>>>>> - ceph osd unset noout
>>>>>>>>>
>>>>>>>>> Would joshd's async patch for qemu help here, or is there
>>>>>>>>> something
>>>>>>>>> else
>>>>>>>>> going on?
>>>>>>>>>
>>>>>>>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>>
>>>>>>>>> Mike Dawson
>>>>>>>>> Co-Founder & Director of Cloud Architecture
>>>>>>>>> Cloudapt LLC
>>>>>>>>> 6330 East 75th Street, Suite 170
>>>>>>>>> Indianapolis, IN 46250
>>>>>>>>>
>>>>>>>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Can you reproduce and attach the ceph.log from before you stop
>>>>>>>>>> the
>>>>>>>>>> osd
>>>>>>>>>> until after you have started the osd and it has recovered?
>>>>>>>>>> -Sam
>>>>>>>>>>
>>>>>>>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> i still have recovery issues with cuttlefish. After the OSD
>>>>>>>>>>> comes
>>>>>>>>>>> back
>>>>>>>>>>> it seem to hang for around 2-4 minutes and then recovery
>>>>>>>>>>> seems to
>>>>>>>>>>> start
>>>>>>>>>>> (pgs in recovery_wait start to decrement). This is with ceph
>>>>>>>>>>> 0.61.7.
>>>>>>>>>>> I
>>>>>>>>>>> get a lot of slow request messages an hanging VMs.
>>>>>>>>>>>
>>>>>>>>>>> What i noticed today is that if i leave the OSD off as long as
>>>>>>>>>>> ceph
>>>>>>>>>>> starts to backfill - the recovery and "re" backfilling wents
>>>>>>>>>>> absolutely
>>>>>>>>>>> smooth without any issues and no slow request messages at all.
>>>>>>>>>>>
>>>>>>>>>>> Does anybody have an idea why?
>>>>>>>>>>>
>>>>>>>>>>> Greets,
>>>>>>>>>>> Stefan
>>>>>>>>>>> --
>>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>>> ceph-devel"
>>>>>>>>>>> in
>>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>>> More majordomo info at
>>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>> ceph-devel"
>>>>>>>>>> in
>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>> More majordomo info at
>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>>
>>>>>>>>
>>>>>>> --
>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>> ceph-devel" in
>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>
>>>>>>
>>>>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-08 18:13                     ` Stefan Priebe
@ 2013-08-09 21:44                       ` Samuel Just
  2013-08-10 19:08                         ` Stefan Priebe
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-09 21:44 UTC (permalink / raw)
  To: Stefan Priebe
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

I think Stefan's problem is probably distinct from Mike's.

Stefan: Can you reproduce the problem with

debug osd = 20
debug filestore = 20
debug ms = 1
debug optracker = 20

on a few osds (including the restarted osd), and upload those osd logs
along with the ceph.log from before killing the osd until after the
cluster becomes clean again?
-Sam

On Thu, Aug 8, 2013 at 11:13 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> Hi Mike,
>
> Am 08.08.2013 16:05, schrieb Mike Dawson:
>
>> Stefan,
>>
>> I see the same behavior and I theorize it is linked to an issue detailed
>> in another thread [0]. Do your VM guests ever hang while your cluster is
>> HEALTH_OK like described in that other thread?
>>
>> [0] http://comments.gmane.org/gmane.comp.file-systems.ceph.user/2982
>
>
> mhm no can't see that. All our VMs are working fine even under high load
> while ceph is OK.
>
>
>> A few observations:
>>
>> - The VMs that hang do lots of writes (video surveillance).
>> - I use rbd and qemu. The problem exists in both qemu 1.4.x and 1.5.2.
>> - The problem exists with or without joshd's qemu async flush patch.
>> - Windows VMs seem to be more vulnerable than Linux VMs.
>> - If I restart the qemu-system-x86_64 process, the guest will come back
>> to life.
>> - A partial workaround seems to be console input (NoVNC or 'virsh
>> screenshot'), but restarting qemu-system-x86_64 works better.
>> - The issue of VMs hanging seems worse with RBD writeback cache enabled
>> - I typically run virtio, but I believe I've seen it with e1000, too.
>> - VM guests hang at different times, not all at once on a host (or
>> across all hosts).
>> - I co-mingle VM guests on servers that host ceph OSDs.
>>
>>
>>
>> Oliver,
>>
>> If your cluster has to recover/backfill, do your guest VMs hang with
>> more frequency than under normal HEALTH_OK conditions, even if you
>> prioritize client IO as Sam wrote below?
>>
>>
>> Sam,
>>
>> Turning down all the settings you mentioned certainly does slow the
>> recover/backfill process, but it doesn't prevent the VM guests backed by
>> RBD volumes from hanging. In fact, I often try to prioritize
>> recovery/backfill because my guests tend to hang until I get back to
>> HEALTH_OK. Given this apparent bug, completing recovery/backfill quicker
>> leads to less total outage, it seems.
>>
>>
>> Josh,
>>
>> How can I help you investigate if RBD is the common source of both of
>> these issues?
>>
>>
>> Thanks,
>> Mike Dawson
>>
>>
>> On 8/2/2013 2:46 PM, Stefan Priebe wrote:
>>>
>>> Hi,
>>>
>>>          osd recovery max active = 1
>>>          osd max backfills = 1
>>>          osd recovery op priority = 5
>>>
>>> still no difference...
>>>
>>> Stefan
>>>
>>> Am 02.08.2013 20:21, schrieb Samuel Just:
>>>>
>>>> Also, you have osd_recovery_op_priority at 50.  That is close to the
>>>> priority of client IO.  You want it below 10 (defaults to 10), perhaps
>>>> at 1.  You can also adjust down osd_recovery_max_active.
>>>> -Sam
>>>>
>>>> On Fri, Aug 2, 2013 at 11:16 AM, Stefan Priebe <s.priebe@profihost.ag>
>>>> wrote:
>>>>>
>>>>> I already tried both values this makes no difference. The drives are
>>>>> not the
>>>>> bottleneck.
>>>>>
>>>>> Am 02.08.2013 19:35, schrieb Samuel Just:
>>>>>
>>>>>> You might try turning osd_max_backfills to 2 or 1.
>>>>>> -Sam
>>>>>>
>>>>>> On Fri, Aug 2, 2013 at 12:44 AM, Stefan Priebe <s.priebe@profihost.ag>
>>>>>> wrote:
>>>>>>>
>>>>>>>
>>>>>>> Am 01.08.2013 23:23, schrieb Samuel Just:> Can you dump your osd
>>>>>>> settings?
>>>>>>>
>>>>>>>> sudo ceph --admin-daemon ceph-osd.<osdid>.asok config show
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Sure.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> { "name": "osd.0",
>>>>>>>     "cluster": "ceph",
>>>>>>>     "none": "0\/5",
>>>>>>>     "lockdep": "0\/0",
>>>>>>>     "context": "0\/0",
>>>>>>>     "crush": "0\/0",
>>>>>>>     "mds": "0\/0",
>>>>>>>     "mds_balancer": "0\/0",
>>>>>>>     "mds_locker": "0\/0",
>>>>>>>     "mds_log": "0\/0",
>>>>>>>     "mds_log_expire": "0\/0",
>>>>>>>     "mds_migrator": "0\/0",
>>>>>>>     "buffer": "0\/0",
>>>>>>>     "timer": "0\/0",
>>>>>>>     "filer": "0\/0",
>>>>>>>     "striper": "0\/1",
>>>>>>>     "objecter": "0\/0",
>>>>>>>     "rados": "0\/0",
>>>>>>>     "rbd": "0\/0",
>>>>>>>     "journaler": "0\/0",
>>>>>>>     "objectcacher": "0\/0",
>>>>>>>     "client": "0\/0",
>>>>>>>     "osd": "0\/0",
>>>>>>>     "optracker": "0\/0",
>>>>>>>     "objclass": "0\/0",
>>>>>>>     "filestore": "0\/0",
>>>>>>>     "journal": "0\/0",
>>>>>>>     "ms": "0\/0",
>>>>>>>     "mon": "0\/0",
>>>>>>>     "monc": "0\/0",
>>>>>>>     "paxos": "0\/0",
>>>>>>>     "tp": "0\/0",
>>>>>>>     "auth": "0\/0",
>>>>>>>     "crypto": "1\/5",
>>>>>>>     "finisher": "0\/0",
>>>>>>>     "heartbeatmap": "0\/0",
>>>>>>>     "perfcounter": "0\/0",
>>>>>>>     "rgw": "0\/0",
>>>>>>>     "hadoop": "0\/0",
>>>>>>>     "javaclient": "1\/5",
>>>>>>>     "asok": "0\/0",
>>>>>>>     "throttle": "0\/0",
>>>>>>>     "host": "cloud1-1268",
>>>>>>>     "fsid": "00000000-0000-0000-0000-000000000000",
>>>>>>>     "public_addr": "10.255.0.90:0\/0",
>>>>>>>     "cluster_addr": "10.255.0.90:0\/0",
>>>>>>>     "public_network": "10.255.0.1\/24",
>>>>>>>     "cluster_network": "10.255.0.1\/24",
>>>>>>>     "num_client": "1",
>>>>>>>     "monmap": "",
>>>>>>>     "mon_host": "",
>>>>>>>     "lockdep": "false",
>>>>>>>     "run_dir": "\/var\/run\/ceph",
>>>>>>>     "admin_socket": "\/var\/run\/ceph\/ceph-osd.0.asok",
>>>>>>>     "daemonize": "true",
>>>>>>>     "pid_file": "\/var\/run\/ceph\/osd.0.pid",
>>>>>>>     "chdir": "\/",
>>>>>>>     "max_open_files": "0",
>>>>>>>     "fatal_signal_handlers": "true",
>>>>>>>     "log_file": "\/var\/log\/ceph\/ceph-osd.0.log",
>>>>>>>     "log_max_new": "1000",
>>>>>>>     "log_max_recent": "10000",
>>>>>>>     "log_to_stderr": "false",
>>>>>>>     "err_to_stderr": "true",
>>>>>>>     "log_to_syslog": "false",
>>>>>>>     "err_to_syslog": "false",
>>>>>>>     "log_flush_on_exit": "true",
>>>>>>>     "log_stop_at_utilization": "0.97",
>>>>>>>     "clog_to_monitors": "true",
>>>>>>>     "clog_to_syslog": "false",
>>>>>>>     "clog_to_syslog_level": "info",
>>>>>>>     "clog_to_syslog_facility": "daemon",
>>>>>>>     "mon_cluster_log_to_syslog": "false",
>>>>>>>     "mon_cluster_log_to_syslog_level": "info",
>>>>>>>     "mon_cluster_log_to_syslog_facility": "daemon",
>>>>>>>     "mon_cluster_log_file": "\/var\/log\/ceph\/ceph.log",
>>>>>>>     "key": "",
>>>>>>>     "keyfile": "",
>>>>>>>     "keyring": "\/etc\/ceph\/osd.0.keyring",
>>>>>>>     "heartbeat_interval": "5",
>>>>>>>     "heartbeat_file": "",
>>>>>>>     "heartbeat_inject_failure": "0",
>>>>>>>     "perf": "true",
>>>>>>>     "ms_tcp_nodelay": "true",
>>>>>>>     "ms_tcp_rcvbuf": "0",
>>>>>>>     "ms_initial_backoff": "0.2",
>>>>>>>     "ms_max_backoff": "15",
>>>>>>>     "ms_nocrc": "false",
>>>>>>>     "ms_die_on_bad_msg": "false",
>>>>>>>     "ms_die_on_unhandled_msg": "false",
>>>>>>>     "ms_dispatch_throttle_bytes": "104857600",
>>>>>>>     "ms_bind_ipv6": "false",
>>>>>>>     "ms_bind_port_min": "6800",
>>>>>>>     "ms_bind_port_max": "7100",
>>>>>>>     "ms_rwthread_stack_bytes": "1048576",
>>>>>>>     "ms_tcp_read_timeout": "900",
>>>>>>>     "ms_pq_max_tokens_per_priority": "4194304",
>>>>>>>     "ms_pq_min_cost": "65536",
>>>>>>>     "ms_inject_socket_failures": "0",
>>>>>>>     "ms_inject_delay_type": "",
>>>>>>>     "ms_inject_delay_max": "1",
>>>>>>>     "ms_inject_delay_probability": "0",
>>>>>>>     "ms_inject_internal_delays": "0",
>>>>>>>     "mon_data": "\/var\/lib\/ceph\/mon\/ceph-0",
>>>>>>>     "mon_initial_members": "",
>>>>>>>     "mon_sync_fs_threshold": "5",
>>>>>>>     "mon_compact_on_start": "false",
>>>>>>>     "mon_compact_on_bootstrap": "false",
>>>>>>>     "mon_compact_on_trim": "true",
>>>>>>>     "mon_tick_interval": "5",
>>>>>>>     "mon_subscribe_interval": "300",
>>>>>>>     "mon_osd_laggy_halflife": "3600",
>>>>>>>     "mon_osd_laggy_weight": "0.3",
>>>>>>>     "mon_osd_adjust_heartbeat_grace": "true",
>>>>>>>     "mon_osd_adjust_down_out_interval": "true",
>>>>>>>     "mon_osd_auto_mark_in": "false",
>>>>>>>     "mon_osd_auto_mark_auto_out_in": "true",
>>>>>>>     "mon_osd_auto_mark_new_in": "true",
>>>>>>>     "mon_osd_down_out_interval": "300",
>>>>>>>     "mon_osd_down_out_subtree_limit": "rack",
>>>>>>>     "mon_osd_min_up_ratio": "0.3",
>>>>>>>     "mon_osd_min_in_ratio": "0.3",
>>>>>>>     "mon_stat_smooth_intervals": "2",
>>>>>>>     "mon_lease": "5",
>>>>>>>     "mon_lease_renew_interval": "3",
>>>>>>>     "mon_lease_ack_timeout": "10",
>>>>>>>     "mon_clock_drift_allowed": "0.05",
>>>>>>>     "mon_clock_drift_warn_backoff": "5",
>>>>>>>     "mon_timecheck_interval": "300",
>>>>>>>     "mon_accept_timeout": "10",
>>>>>>>     "mon_pg_create_interval": "30",
>>>>>>>     "mon_pg_stuck_threshold": "300",
>>>>>>>     "mon_osd_full_ratio": "0.95",
>>>>>>>     "mon_osd_nearfull_ratio": "0.85",
>>>>>>>     "mon_globalid_prealloc": "100",
>>>>>>>     "mon_osd_report_timeout": "900",
>>>>>>>     "mon_force_standby_active": "true",
>>>>>>>     "mon_min_osdmap_epochs": "500",
>>>>>>>     "mon_max_pgmap_epochs": "500",
>>>>>>>     "mon_max_log_epochs": "500",
>>>>>>>     "mon_max_osd": "10000",
>>>>>>>     "mon_probe_timeout": "2",
>>>>>>>     "mon_slurp_timeout": "10",
>>>>>>>     "mon_slurp_bytes": "262144",
>>>>>>>     "mon_client_bytes": "104857600",
>>>>>>>     "mon_daemon_bytes": "419430400",
>>>>>>>     "mon_max_log_entries_per_event": "4096",
>>>>>>>     "mon_health_data_update_interval": "60",
>>>>>>>     "mon_data_avail_crit": "5",
>>>>>>>     "mon_data_avail_warn": "30",
>>>>>>>     "mon_config_key_max_entry_size": "4096",
>>>>>>>     "mon_sync_trim_timeout": "30",
>>>>>>>     "mon_sync_heartbeat_timeout": "30",
>>>>>>>     "mon_sync_heartbeat_interval": "5",
>>>>>>>     "mon_sync_backoff_timeout": "30",
>>>>>>>     "mon_sync_timeout": "30",
>>>>>>>     "mon_sync_max_retries": "5",
>>>>>>>     "mon_sync_max_payload_size": "1048576",
>>>>>>>     "mon_sync_debug": "false",
>>>>>>>     "mon_sync_debug_leader": "-1",
>>>>>>>     "mon_sync_debug_provider": "-1",
>>>>>>>     "mon_sync_debug_provider_fallback": "-1",
>>>>>>>     "mon_debug_dump_transactions": "false",
>>>>>>>     "mon_debug_dump_location": "\/var\/log\/ceph\/ceph-osd.0.tdump",
>>>>>>>     "mon_sync_leader_kill_at": "0",
>>>>>>>     "mon_sync_provider_kill_at": "0",
>>>>>>>     "mon_sync_requester_kill_at": "0",
>>>>>>>     "mon_leveldb_write_buffer_size": "33554432",
>>>>>>>     "mon_leveldb_cache_size": "268435456",
>>>>>>>     "mon_leveldb_block_size": "65536",
>>>>>>>     "mon_leveldb_bloom_size": "0",
>>>>>>>     "mon_leveldb_max_open_files": "0",
>>>>>>>     "mon_leveldb_compression": "false",
>>>>>>>     "mon_leveldb_paranoid": "false",
>>>>>>>     "mon_leveldb_log": "",
>>>>>>>     "paxos_stash_full_interval": "25",
>>>>>>>     "paxos_max_join_drift": "100",
>>>>>>>     "paxos_propose_interval": "1",
>>>>>>>     "paxos_min_wait": "0.05",
>>>>>>>     "paxos_min": "500",
>>>>>>>     "paxos_trim_min": "500",
>>>>>>>     "paxos_trim_max": "1000",
>>>>>>>     "paxos_trim_disabled_max_versions": "108000",
>>>>>>>     "paxos_service_trim_min": "500",
>>>>>>>     "paxos_service_trim_max": "1000",
>>>>>>>     "clock_offset": "0",
>>>>>>>     "auth_cluster_required": "none",
>>>>>>>     "auth_service_required": "none",
>>>>>>>     "auth_client_required": "none",
>>>>>>>     "auth_supported": "none",
>>>>>>>     "cephx_require_signatures": "false",
>>>>>>>     "cephx_cluster_require_signatures": "false",
>>>>>>>     "cephx_service_require_signatures": "false",
>>>>>>>     "cephx_sign_messages": "true",
>>>>>>>     "auth_mon_ticket_ttl": "43200",
>>>>>>>     "auth_service_ticket_ttl": "3600",
>>>>>>>     "auth_debug": "false",
>>>>>>>     "mon_client_hunt_interval": "3",
>>>>>>>     "mon_client_ping_interval": "10",
>>>>>>>     "mon_client_max_log_entries_per_message": "1000",
>>>>>>>     "mon_max_pool_pg_num": "65536",
>>>>>>>     "mon_pool_quota_warn_threshold": "0",
>>>>>>>     "mon_pool_quota_crit_threshold": "0",
>>>>>>>     "client_cache_size": "16384",
>>>>>>>     "client_cache_mid": "0.75",
>>>>>>>     "client_use_random_mds": "false",
>>>>>>>     "client_mount_timeout": "300",
>>>>>>>     "client_tick_interval": "1",
>>>>>>>     "client_trace": "",
>>>>>>>     "client_readahead_min": "131072",
>>>>>>>     "client_readahead_max_bytes": "0",
>>>>>>>     "client_readahead_max_periods": "4",
>>>>>>>     "client_snapdir": ".snap",
>>>>>>>     "client_mountpoint": "\/",
>>>>>>>     "client_notify_timeout": "10",
>>>>>>>     "client_caps_release_delay": "5",
>>>>>>>     "client_oc": "true",
>>>>>>>     "client_oc_size": "209715200",
>>>>>>>     "client_oc_max_dirty": "104857600",
>>>>>>>     "client_oc_target_dirty": "8388608",
>>>>>>>     "client_oc_max_dirty_age": "5",
>>>>>>>     "client_oc_max_objects": "1000",
>>>>>>>     "client_debug_force_sync_read": "false",
>>>>>>>     "client_debug_inject_tick_delay": "0",
>>>>>>>     "fuse_use_invalidate_cb": "false",
>>>>>>>     "fuse_allow_other": "true",
>>>>>>>     "fuse_default_permissions": "true",
>>>>>>>     "fuse_big_writes": "true",
>>>>>>>     "fuse_atomic_o_trunc": "true",
>>>>>>>     "fuse_debug": "false",
>>>>>>>     "objecter_tick_interval": "5",
>>>>>>>     "objecter_timeout": "10",
>>>>>>>     "objecter_inflight_op_bytes": "104857600",
>>>>>>>     "objecter_inflight_ops": "1024",
>>>>>>>     "journaler_allow_split_entries": "true",
>>>>>>>     "journaler_write_head_interval": "15",
>>>>>>>     "journaler_prefetch_periods": "10",
>>>>>>>     "journaler_prezero_periods": "5",
>>>>>>>     "journaler_batch_interval": "0.001",
>>>>>>>     "journaler_batch_max": "0",
>>>>>>>     "mds_data": "\/var\/lib\/ceph\/mds\/ceph-0",
>>>>>>>     "mds_max_file_size": "1099511627776",
>>>>>>>     "mds_cache_size": "100000",
>>>>>>>     "mds_cache_mid": "0.7",
>>>>>>>     "mds_mem_max": "1048576",
>>>>>>>     "mds_dir_commit_ratio": "0.5",
>>>>>>>     "mds_dir_max_commit_size": "90",
>>>>>>>     "mds_decay_halflife": "5",
>>>>>>>     "mds_beacon_interval": "4",
>>>>>>>     "mds_beacon_grace": "15",
>>>>>>>     "mds_enforce_unique_name": "true",
>>>>>>>     "mds_blacklist_interval": "1440",
>>>>>>>     "mds_session_timeout": "60",
>>>>>>>     "mds_session_autoclose": "300",
>>>>>>>     "mds_reconnect_timeout": "45",
>>>>>>>     "mds_tick_interval": "5",
>>>>>>>     "mds_dirstat_min_interval": "1",
>>>>>>>     "mds_scatter_nudge_interval": "5",
>>>>>>>     "mds_client_prealloc_inos": "1000",
>>>>>>>     "mds_early_reply": "true",
>>>>>>>     "mds_use_tmap": "true",
>>>>>>>     "mds_default_dir_hash": "2",
>>>>>>>     "mds_log": "true",
>>>>>>>     "mds_log_skip_corrupt_events": "false",
>>>>>>>     "mds_log_max_events": "-1",
>>>>>>>     "mds_log_segment_size": "0",
>>>>>>>     "mds_log_max_segments": "30",
>>>>>>>     "mds_log_max_expiring": "20",
>>>>>>>     "mds_bal_sample_interval": "3",
>>>>>>>     "mds_bal_replicate_threshold": "8000",
>>>>>>>     "mds_bal_unreplicate_threshold": "0",
>>>>>>>     "mds_bal_frag": "false",
>>>>>>>     "mds_bal_split_size": "10000",
>>>>>>>     "mds_bal_split_rd": "25000",
>>>>>>>     "mds_bal_split_wr": "10000",
>>>>>>>     "mds_bal_split_bits": "3",
>>>>>>>     "mds_bal_merge_size": "50",
>>>>>>>     "mds_bal_merge_rd": "1000",
>>>>>>>     "mds_bal_merge_wr": "1000",
>>>>>>>     "mds_bal_interval": "10",
>>>>>>>     "mds_bal_fragment_interval": "5",
>>>>>>>     "mds_bal_idle_threshold": "0",
>>>>>>>     "mds_bal_max": "-1",
>>>>>>>     "mds_bal_max_until": "-1",
>>>>>>>     "mds_bal_mode": "0",
>>>>>>>     "mds_bal_min_rebalance": "0.1",
>>>>>>>     "mds_bal_min_start": "0.2",
>>>>>>>     "mds_bal_need_min": "0.8",
>>>>>>>     "mds_bal_need_max": "1.2",
>>>>>>>     "mds_bal_midchunk": "0.3",
>>>>>>>     "mds_bal_minchunk": "0.001",
>>>>>>>     "mds_bal_target_removal_min": "5",
>>>>>>>     "mds_bal_target_removal_max": "10",
>>>>>>>     "mds_replay_interval": "1",
>>>>>>>     "mds_shutdown_check": "0",
>>>>>>>     "mds_thrash_exports": "0",
>>>>>>>     "mds_thrash_fragments": "0",
>>>>>>>     "mds_dump_cache_on_map": "false",
>>>>>>>     "mds_dump_cache_after_rejoin": "false",
>>>>>>>     "mds_verify_scatter": "false",
>>>>>>>     "mds_debug_scatterstat": "false",
>>>>>>>     "mds_debug_frag": "false",
>>>>>>>     "mds_debug_auth_pins": "false",
>>>>>>>     "mds_debug_subtrees": "false",
>>>>>>>     "mds_kill_mdstable_at": "0",
>>>>>>>     "mds_kill_export_at": "0",
>>>>>>>     "mds_kill_import_at": "0",
>>>>>>>     "mds_kill_link_at": "0",
>>>>>>>     "mds_kill_rename_at": "0",
>>>>>>>     "mds_kill_openc_at": "0",
>>>>>>>     "mds_kill_journal_at": "0",
>>>>>>>     "mds_kill_journal_expire_at": "0",
>>>>>>>     "mds_kill_journal_replay_at": "0",
>>>>>>>     "mds_inject_traceless_reply_probability": "0",
>>>>>>>     "mds_wipe_sessions": "false",
>>>>>>>     "mds_wipe_ino_prealloc": "false",
>>>>>>>     "mds_skip_ino": "0",
>>>>>>>     "max_mds": "1",
>>>>>>>     "mds_standby_for_name": "",
>>>>>>>     "mds_standby_for_rank": "-1",
>>>>>>>     "mds_standby_replay": "false",
>>>>>>>     "osd_auto_upgrade_tmap": "true",
>>>>>>>     "osd_tmapput_sets_uses_tmap": "false",
>>>>>>>     "osd_max_backfills": "5",
>>>>>>>     "osd_backfill_full_ratio": "0.85",
>>>>>>>     "osd_backfill_retry_interval": "10",
>>>>>>>     "osd_uuid": "00000000-0000-0000-0000-000000000000",
>>>>>>>     "osd_data": "\/ceph\/osd.0\/",
>>>>>>>     "osd_journal": "\/dev\/disk\/by-partlabel\/journalosd0",
>>>>>>>     "osd_journal_size": "5120",
>>>>>>>     "osd_max_write_size": "90",
>>>>>>>     "osd_max_pgls": "1024",
>>>>>>>     "osd_client_message_size_cap": "524288000",
>>>>>>>     "osd_client_message_cap": "100",
>>>>>>>     "osd_pg_bits": "6",
>>>>>>>     "osd_pgp_bits": "6",
>>>>>>>     "osd_crush_chooseleaf_type": "1",
>>>>>>>     "osd_min_rep": "1",
>>>>>>>     "osd_max_rep": "10",
>>>>>>>     "osd_pool_default_crush_rule": "0",
>>>>>>>     "osd_pool_default_size": "2",
>>>>>>>     "osd_pool_default_min_size": "0",
>>>>>>>     "osd_pool_default_pg_num": "8",
>>>>>>>     "osd_pool_default_pgp_num": "8",
>>>>>>>     "osd_pool_default_flags": "0",
>>>>>>>     "osd_map_dedup": "true",
>>>>>>>     "osd_map_cache_size": "500",
>>>>>>>     "osd_map_message_max": "100",
>>>>>>>     "osd_map_share_max_epochs": "100",
>>>>>>>     "osd_op_threads": "2",
>>>>>>>     "osd_peering_wq_batch_size": "20",
>>>>>>>     "osd_op_pq_max_tokens_per_priority": "4194304",
>>>>>>>     "osd_op_pq_min_cost": "65536",
>>>>>>>     "osd_disk_threads": "1",
>>>>>>>     "osd_recovery_threads": "2",
>>>>>>>     "osd_recover_clone_overlap": "true",
>>>>>>>     "osd_backfill_scan_min": "64",
>>>>>>>     "osd_backfill_scan_max": "512",
>>>>>>>     "osd_op_thread_timeout": "15",
>>>>>>>     "osd_recovery_thread_timeout": "30",
>>>>>>>     "osd_snap_trim_thread_timeout": "3600",
>>>>>>>     "osd_scrub_thread_timeout": "60",
>>>>>>>     "osd_scrub_finalize_thread_timeout": "600",
>>>>>>>     "osd_remove_thread_timeout": "3600",
>>>>>>>     "osd_command_thread_timeout": "600",
>>>>>>>     "osd_age": "0.8",
>>>>>>>     "osd_age_time": "0",
>>>>>>>     "osd_heartbeat_addr": ":\/0",
>>>>>>>     "osd_heartbeat_interval": "6",
>>>>>>>     "osd_heartbeat_grace": "20",
>>>>>>>     "osd_mon_heartbeat_interval": "30",
>>>>>>>     "osd_mon_report_interval_max": "120",
>>>>>>>     "osd_mon_report_interval_min": "5",
>>>>>>>     "osd_pg_stat_report_interval_max": "500",
>>>>>>>     "osd_mon_ack_timeout": "30",
>>>>>>>     "osd_min_down_reporters": "1",
>>>>>>>     "osd_min_down_reports": "3",
>>>>>>>     "osd_default_data_pool_replay_window": "45",
>>>>>>>     "osd_preserve_trimmed_log": "false",
>>>>>>>     "osd_auto_mark_unfound_lost": "false",
>>>>>>>     "osd_recovery_delay_start": "0",
>>>>>>>     "osd_recovery_max_active": "5",
>>>>>>>     "osd_recovery_max_chunk": "8388608",
>>>>>>>     "osd_recovery_forget_lost_objects": "false",
>>>>>>>     "osd_max_scrubs": "1",
>>>>>>>     "osd_scrub_load_threshold": "0.5",
>>>>>>>     "osd_scrub_min_interval": "86400",
>>>>>>>     "osd_scrub_max_interval": "604800",
>>>>>>>     "osd_deep_scrub_interval": "604800",
>>>>>>>     "osd_deep_scrub_stride": "524288",
>>>>>>>     "osd_scan_list_ping_tp_interval": "100",
>>>>>>>     "osd_auto_weight": "false",
>>>>>>>     "osd_class_dir": "\/usr\/lib\/rados-classes",
>>>>>>>     "osd_check_for_log_corruption": "false",
>>>>>>>     "osd_use_stale_snap": "false",
>>>>>>>     "osd_rollback_to_cluster_snap": "",
>>>>>>>     "osd_default_notify_timeout": "30",
>>>>>>>     "osd_kill_backfill_at": "0",
>>>>>>>     "osd_pg_epoch_persisted_max_stale": "200",
>>>>>>>     "osd_min_pg_log_entries": "500",
>>>>>>>     "osd_max_pg_log_entries": "1500",
>>>>>>>     "osd_op_complaint_time": "30",
>>>>>>>     "osd_command_max_records": "256",
>>>>>>>     "osd_op_log_threshold": "5",
>>>>>>>     "osd_verify_sparse_read_holes": "false",
>>>>>>>     "osd_debug_drop_ping_probability": "0",
>>>>>>>     "osd_debug_drop_ping_duration": "0",
>>>>>>>     "osd_debug_drop_pg_create_probability": "0",
>>>>>>>     "osd_debug_drop_pg_create_duration": "1",
>>>>>>>     "osd_debug_drop_op_probability": "0",
>>>>>>>     "osd_debug_op_order": "false",
>>>>>>>     "osd_debug_verify_snaps_on_info": "false",
>>>>>>>     "osd_debug_verify_stray_on_activate": "false",
>>>>>>>     "osd_debug_skip_full_check_in_backfill_reservation": "false",
>>>>>>>     "osd_op_history_size": "20",
>>>>>>>     "osd_op_history_duration": "600",
>>>>>>>     "osd_target_transaction_size": "30",
>>>>>>>     "osd_failsafe_full_ratio": "0.97",
>>>>>>>     "osd_failsafe_nearfull_ratio": "0.9",
>>>>>>>     "osd_leveldb_write_buffer_size": "0",
>>>>>>>     "osd_leveldb_cache_size": "0",
>>>>>>>     "osd_leveldb_block_size": "0",
>>>>>>>     "osd_leveldb_bloom_size": "0",
>>>>>>>     "osd_leveldb_max_open_files": "0",
>>>>>>>     "osd_leveldb_compression": "true",
>>>>>>>     "osd_leveldb_paranoid": "false",
>>>>>>>     "osd_leveldb_log": "",
>>>>>>>     "osd_client_op_priority": "63",
>>>>>>>     "osd_recovery_op_priority": "50",
>>>>>>>     "osd_mon_shutdown_timeout": "5",
>>>>>>>     "filestore": "false",
>>>>>>>     "filestore_index_retry_probability": "0",
>>>>>>>     "filestore_debug_inject_read_err": "false",
>>>>>>>     "filestore_debug_omap_check": "false",
>>>>>>>     "filestore_xattr_use_omap": "false",
>>>>>>>     "filestore_max_inline_xattr_size": "512",
>>>>>>>     "filestore_max_inline_xattrs": "2",
>>>>>>>     "filestore_max_sync_interval": "5",
>>>>>>>     "filestore_min_sync_interval": "0.01",
>>>>>>>     "filestore_btrfs_snap": "true",
>>>>>>>     "filestore_btrfs_clone_range": "true",
>>>>>>>     "filestore_fsync_flushes_journal_data": "false",
>>>>>>>     "filestore_fiemap": "false",
>>>>>>>     "filestore_flusher": "true",
>>>>>>>     "filestore_flusher_max_fds": "512",
>>>>>>>     "filestore_flush_min": "65536",
>>>>>>>     "filestore_sync_flush": "false",
>>>>>>>     "filestore_journal_parallel": "false",
>>>>>>>     "filestore_journal_writeahead": "false",
>>>>>>>     "filestore_journal_trailing": "false",
>>>>>>>     "filestore_queue_max_ops": "500",
>>>>>>>     "filestore_queue_max_bytes": "104857600",
>>>>>>>     "filestore_queue_committing_max_ops": "5000",
>>>>>>>     "filestore_queue_committing_max_bytes": "104857600",
>>>>>>>     "filestore_op_threads": "2",
>>>>>>>     "filestore_op_thread_timeout": "60",
>>>>>>>     "filestore_op_thread_suicide_timeout": "180",
>>>>>>>     "filestore_commit_timeout": "600",
>>>>>>>     "filestore_fiemap_threshold": "4096",
>>>>>>>     "filestore_merge_threshold": "10",
>>>>>>>     "filestore_split_multiple": "2",
>>>>>>>     "filestore_update_to": "1000",
>>>>>>>     "filestore_blackhole": "false",
>>>>>>>     "filestore_dump_file": "",
>>>>>>>     "filestore_kill_at": "0",
>>>>>>>     "filestore_inject_stall": "0",
>>>>>>>     "filestore_fail_eio": "true",
>>>>>>>     "filestore_replica_fadvise": "true",
>>>>>>>     "filestore_debug_verify_split": "false",
>>>>>>>     "journal_dio": "true",
>>>>>>>     "journal_aio": "true",
>>>>>>>     "journal_force_aio": "false",
>>>>>>>     "journal_max_corrupt_search": "10485760",
>>>>>>>     "journal_block_align": "true",
>>>>>>>     "journal_write_header_frequency": "0",
>>>>>>>     "journal_max_write_bytes": "10485760",
>>>>>>>     "journal_max_write_entries": "100",
>>>>>>>     "journal_queue_max_ops": "300",
>>>>>>>     "journal_queue_max_bytes": "33554432",
>>>>>>>     "journal_align_min_size": "65536",
>>>>>>>     "journal_replay_from": "0",
>>>>>>>     "journal_zero_on_create": "false",
>>>>>>>     "journal_ignore_corruption": "false",
>>>>>>>     "rbd_cache": "false",
>>>>>>>     "rbd_cache_writethrough_until_flush": "false",
>>>>>>>     "rbd_cache_size": "33554432",
>>>>>>>     "rbd_cache_max_dirty": "25165824",
>>>>>>>     "rbd_cache_target_dirty": "16777216",
>>>>>>>     "rbd_cache_max_dirty_age": "1",
>>>>>>>     "rbd_cache_block_writes_upfront": "false",
>>>>>>>     "rbd_concurrent_management_ops": "10",
>>>>>>>     "rbd_default_format": "1",
>>>>>>>     "rbd_default_order": "22",
>>>>>>>     "rbd_default_stripe_count": "1",
>>>>>>>     "rbd_default_stripe_unit": "4194304",
>>>>>>>     "rbd_default_features": "3",
>>>>>>>     "nss_db_path": "",
>>>>>>>     "rgw_data": "\/var\/lib\/ceph\/radosgw\/ceph-0",
>>>>>>>     "rgw_enable_apis": "s3, swift, swift_auth, admin",
>>>>>>>     "rgw_cache_enabled": "true",
>>>>>>>     "rgw_cache_lru_size": "10000",
>>>>>>>     "rgw_socket_path": "",
>>>>>>>     "rgw_host": "",
>>>>>>>     "rgw_port": "",
>>>>>>>     "rgw_dns_name": "",
>>>>>>>     "rgw_script_uri": "",
>>>>>>>     "rgw_request_uri": "",
>>>>>>>     "rgw_swift_url": "",
>>>>>>>     "rgw_swift_url_prefix": "swift",
>>>>>>>     "rgw_swift_auth_url": "",
>>>>>>>     "rgw_swift_auth_entry": "auth",
>>>>>>>     "rgw_keystone_url": "",
>>>>>>>     "rgw_keystone_admin_token": "",
>>>>>>>     "rgw_keystone_accepted_roles": "Member, admin",
>>>>>>>     "rgw_keystone_token_cache_size": "10000",
>>>>>>>     "rgw_keystone_revocation_interval": "900",
>>>>>>>     "rgw_admin_entry": "admin",
>>>>>>>     "rgw_enforce_swift_acls": "true",
>>>>>>>     "rgw_swift_token_expiration": "86400",
>>>>>>>     "rgw_print_continue": "true",
>>>>>>>     "rgw_remote_addr_param": "REMOTE_ADDR",
>>>>>>>     "rgw_op_thread_timeout": "600",
>>>>>>>     "rgw_op_thread_suicide_timeout": "0",
>>>>>>>     "rgw_thread_pool_size": "100",
>>>>>>>     "rgw_num_control_oids": "8",
>>>>>>>     "rgw_zone_root_pool": ".rgw.root",
>>>>>>>     "rgw_log_nonexistent_bucket": "false",
>>>>>>>     "rgw_log_object_name": "%Y-%m-%d-%H-%i-%n",
>>>>>>>     "rgw_log_object_name_utc": "false",
>>>>>>>     "rgw_usage_max_shards": "32",
>>>>>>>     "rgw_usage_max_user_shards": "1",
>>>>>>>     "rgw_enable_ops_log": "false",
>>>>>>>     "rgw_enable_usage_log": "false",
>>>>>>>     "rgw_ops_log_rados": "true",
>>>>>>>     "rgw_ops_log_socket_path": "",
>>>>>>>     "rgw_ops_log_data_backlog": "5242880",
>>>>>>>     "rgw_usage_log_flush_threshold": "1024",
>>>>>>>     "rgw_usage_log_tick_interval": "30",
>>>>>>>     "rgw_intent_log_object_name": "%Y-%m-%d-%i-%n",
>>>>>>>     "rgw_intent_log_object_name_utc": "false",
>>>>>>>     "rgw_init_timeout": "300",
>>>>>>>     "rgw_mime_types_file": "\/etc\/mime.types",
>>>>>>>     "rgw_gc_max_objs": "32",
>>>>>>>     "rgw_gc_obj_min_wait": "7200",
>>>>>>>     "rgw_gc_processor_max_time": "3600",
>>>>>>>     "rgw_gc_processor_period": "3600",
>>>>>>>     "rgw_s3_success_create_obj_status": "0",
>>>>>>>     "rgw_resolve_cname": "false",
>>>>>>>     "rgw_obj_stripe_size": "4194304",
>>>>>>>     "rgw_extended_http_attrs": "",
>>>>>>>     "rgw_exit_timeout_secs": "120",
>>>>>>>     "rgw_get_obj_window_size": "16777216",
>>>>>>>     "rgw_get_obj_max_req_size": "4194304",
>>>>>>>     "rgw_relaxed_s3_bucket_names": "false",
>>>>>>>     "rgw_list_buckets_max_chunk": "1000",
>>>>>>>     "mutex_perf_counter": "false",
>>>>>>>     "internal_safe_to_start_threads": "true"}
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Stefan
>>>>>>>
>>>>>>>
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Thu, Aug 1, 2013 at 12:07 PM, Stefan Priebe
>>>>>>>> <s.priebe@profihost.ag>
>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Mike we already have the async patch running. Yes it helps but only
>>>>>>>>> helps
>>>>>>>>> it
>>>>>>>>> does not solve. It just hides the issue ...
>>>>>>>>> Am 01.08.2013 20:54, schrieb Mike Dawson:
>>>>>>>>>
>>>>>>>>>> I am also seeing recovery issues with 0.61.7. Here's the process:
>>>>>>>>>>
>>>>>>>>>> - ceph osd set noout
>>>>>>>>>>
>>>>>>>>>> - Reboot one of the nodes hosting OSDs
>>>>>>>>>>         - VMs mounted from RBD volumes work properly
>>>>>>>>>>
>>>>>>>>>> - I see the OSD's boot messages as they re-join the cluster
>>>>>>>>>>
>>>>>>>>>> - Start seeing active+recovery_wait, peering, and
>>>>>>>>>> active+recovering
>>>>>>>>>>         - VMs mounted from RBD volumes become unresponsive.
>>>>>>>>>>
>>>>>>>>>> - Recovery completes
>>>>>>>>>>         - VMs mounted from RBD volumes regain responsiveness
>>>>>>>>>>
>>>>>>>>>> - ceph osd unset noout
>>>>>>>>>>
>>>>>>>>>> Would joshd's async patch for qemu help here, or is there
>>>>>>>>>> something
>>>>>>>>>> else
>>>>>>>>>> going on?
>>>>>>>>>>
>>>>>>>>>> Output of ceph -w at: http://pastebin.com/raw.php?i=JLcZYFzY
>>>>>>>>>>
>>>>>>>>>> Thanks,
>>>>>>>>>>
>>>>>>>>>> Mike Dawson
>>>>>>>>>> Co-Founder & Director of Cloud Architecture
>>>>>>>>>> Cloudapt LLC
>>>>>>>>>> 6330 East 75th Street, Suite 170
>>>>>>>>>> Indianapolis, IN 46250
>>>>>>>>>>
>>>>>>>>>> On 8/1/2013 2:34 PM, Samuel Just wrote:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Can you reproduce and attach the ceph.log from before you stop
>>>>>>>>>>> the
>>>>>>>>>>> osd
>>>>>>>>>>> until after you have started the osd and it has recovered?
>>>>>>>>>>> -Sam
>>>>>>>>>>>
>>>>>>>>>>> On Thu, Aug 1, 2013 at 1:22 AM, Stefan Priebe - Profihost AG
>>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Hi,
>>>>>>>>>>>>
>>>>>>>>>>>> i still have recovery issues with cuttlefish. After the OSD
>>>>>>>>>>>> comes
>>>>>>>>>>>> back
>>>>>>>>>>>> it seem to hang for around 2-4 minutes and then recovery
>>>>>>>>>>>> seems to
>>>>>>>>>>>> start
>>>>>>>>>>>> (pgs in recovery_wait start to decrement). This is with ceph
>>>>>>>>>>>> 0.61.7.
>>>>>>>>>>>> I
>>>>>>>>>>>> get a lot of slow request messages an hanging VMs.
>>>>>>>>>>>>
>>>>>>>>>>>> What i noticed today is that if i leave the OSD off as long as
>>>>>>>>>>>> ceph
>>>>>>>>>>>> starts to backfill - the recovery and "re" backfilling wents
>>>>>>>>>>>> absolutely
>>>>>>>>>>>> smooth without any issues and no slow request messages at all.
>>>>>>>>>>>>
>>>>>>>>>>>> Does anybody have an idea why?
>>>>>>>>>>>>
>>>>>>>>>>>> Greets,
>>>>>>>>>>>> Stefan
>>>>>>>>>>>> --
>>>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>>>> ceph-devel"
>>>>>>>>>>>> in
>>>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>>>> More majordomo info at
>>>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>>> ceph-devel"
>>>>>>>>>>> in
>>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>>> More majordomo info at
>>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>>>
>>>>>>>>>
>>>>>>>> --
>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>> ceph-devel" in
>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>>
>>>>>>>
>>>>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-09 21:44                       ` Samuel Just
@ 2013-08-10 19:08                         ` Stefan Priebe
  2013-08-11  3:50                           ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-10 19:08 UTC (permalink / raw)
  To: Samuel Just
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

Hi Samual,

Am 09.08.2013 23:44, schrieb Samuel Just:
> I think Stefan's problem is probably distinct from Mike's.
>
> Stefan: Can you reproduce the problem with
>
> debug osd = 20
> debug filestore = 20
> debug ms = 1
> debug optracker = 20
>
> on a few osds (including the restarted osd), and upload those osd logs
> along with the ceph.log from before killing the osd until after the
> cluster becomes clean again?

done - you'll find the logs at cephdrop folder: 
slow_requests_recovering_cuttlefish

osd.52 was the one recovering

Thanks!

Greets,
Stefan

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-10 19:08                         ` Stefan Priebe
@ 2013-08-11  3:50                           ` Samuel Just
  2013-08-13  4:39                             ` Stefan Priebe - Profihost AG
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-11  3:50 UTC (permalink / raw)
  To: Stefan Priebe
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

Great!  I'll take a look on Monday.
-Sam

On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> Hi Samual,
>
> Am 09.08.2013 23:44, schrieb Samuel Just:
>
>> I think Stefan's problem is probably distinct from Mike's.
>>
>> Stefan: Can you reproduce the problem with
>>
>> debug osd = 20
>> debug filestore = 20
>> debug ms = 1
>> debug optracker = 20
>>
>> on a few osds (including the restarted osd), and upload those osd logs
>> along with the ceph.log from before killing the osd until after the
>> cluster becomes clean again?
>
>
> done - you'll find the logs at cephdrop folder:
> slow_requests_recovering_cuttlefish
>
> osd.52 was the one recovering
>
> Thanks!
>
> Greets,
> Stefan

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-11  3:50                           ` Samuel Just
@ 2013-08-13  4:39                             ` Stefan Priebe - Profihost AG
  2013-08-13  5:34                               ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe - Profihost AG @ 2013-08-13  4:39 UTC (permalink / raw)
  To: Samuel Just
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

Did you take a look?

Stefan

Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:

> Great!  I'll take a look on Monday.
> -Sam
> 
> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> Hi Samual,
>> 
>> Am 09.08.2013 23:44, schrieb Samuel Just:
>> 
>>> I think Stefan's problem is probably distinct from Mike's.
>>> 
>>> Stefan: Can you reproduce the problem with
>>> 
>>> debug osd = 20
>>> debug filestore = 20
>>> debug ms = 1
>>> debug optracker = 20
>>> 
>>> on a few osds (including the restarted osd), and upload those osd logs
>>> along with the ceph.log from before killing the osd until after the
>>> cluster becomes clean again?
>> 
>> 
>> done - you'll find the logs at cephdrop folder:
>> slow_requests_recovering_cuttlefish
>> 
>> osd.52 was the one recovering
>> 
>> Thanks!
>> 
>> Greets,
>> Stefan
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-13  4:39                             ` Stefan Priebe - Profihost AG
@ 2013-08-13  5:34                               ` Samuel Just
  2013-08-13 20:43                                 ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-13  5:34 UTC (permalink / raw)
  To: Stefan Priebe - Profihost AG
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

I got swamped today.  I should be able to look tomorrow.  Sorry!
-Sam

On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
<s.priebe@profihost.ag> wrote:
> Did you take a look?
>
> Stefan
>
> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>
>> Great!  I'll take a look on Monday.
>> -Sam
>>
>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>>> Hi Samual,
>>>
>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>
>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>
>>>> Stefan: Can you reproduce the problem with
>>>>
>>>> debug osd = 20
>>>> debug filestore = 20
>>>> debug ms = 1
>>>> debug optracker = 20
>>>>
>>>> on a few osds (including the restarted osd), and upload those osd logs
>>>> along with the ceph.log from before killing the osd until after the
>>>> cluster becomes clean again?
>>>
>>>
>>> done - you'll find the logs at cephdrop folder:
>>> slow_requests_recovering_cuttlefish
>>>
>>> osd.52 was the one recovering
>>>
>>> Thanks!
>>>
>>> Greets,
>>> Stefan
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-13  5:34                               ` Samuel Just
@ 2013-08-13 20:43                                 ` Samuel Just
  2013-08-13 21:03                                   ` Stefan Priebe - Profihost AG
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-13 20:43 UTC (permalink / raw)
  To: Stefan Priebe - Profihost AG
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

I just backported a couple of patches from next to fix a bug where we
weren't respecting the osd_recovery_max_active config in some cases
(1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
current cuttlefish branch or wait for a 61.8 release.
-Sam

On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com> wrote:
> I got swamped today.  I should be able to look tomorrow.  Sorry!
> -Sam
>
> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
> <s.priebe@profihost.ag> wrote:
>> Did you take a look?
>>
>> Stefan
>>
>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>
>>> Great!  I'll take a look on Monday.
>>> -Sam
>>>
>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>>>> Hi Samual,
>>>>
>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>
>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>
>>>>> Stefan: Can you reproduce the problem with
>>>>>
>>>>> debug osd = 20
>>>>> debug filestore = 20
>>>>> debug ms = 1
>>>>> debug optracker = 20
>>>>>
>>>>> on a few osds (including the restarted osd), and upload those osd logs
>>>>> along with the ceph.log from before killing the osd until after the
>>>>> cluster becomes clean again?
>>>>
>>>>
>>>> done - you'll find the logs at cephdrop folder:
>>>> slow_requests_recovering_cuttlefish
>>>>
>>>> osd.52 was the one recovering
>>>>
>>>> Thanks!
>>>>
>>>> Greets,
>>>> Stefan
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-13 20:43                                 ` Samuel Just
@ 2013-08-13 21:03                                   ` Stefan Priebe - Profihost AG
  2013-08-13 23:11                                     ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe - Profihost AG @ 2013-08-13 21:03 UTC (permalink / raw)
  To: Samuel Just
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi


Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:

> I just backported a couple of patches from next to fix a bug where we
> weren't respecting the osd_recovery_max_active config in some cases
> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
> current cuttlefish branch or wait for a 61.8 release.

Thanks! Are you sure that this is the issue? I don't believe that but i'll give it a try. I already tested a branch from sage where he fixed a race regarding max active some weeks ago. So active recovering was max 1 but the issue didn't went away.

Stefan

> -Sam
> 
> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com> wrote:
>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>> -Sam
>> 
>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>> <s.priebe@profihost.ag> wrote:
>>> Did you take a look?
>>> 
>>> Stefan
>>> 
>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>> 
>>>> Great!  I'll take a look on Monday.
>>>> -Sam
>>>> 
>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>>>>> Hi Samual,
>>>>> 
>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>> 
>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>> 
>>>>>> Stefan: Can you reproduce the problem with
>>>>>> 
>>>>>> debug osd = 20
>>>>>> debug filestore = 20
>>>>>> debug ms = 1
>>>>>> debug optracker = 20
>>>>>> 
>>>>>> on a few osds (including the restarted osd), and upload those osd logs
>>>>>> along with the ceph.log from before killing the osd until after the
>>>>>> cluster becomes clean again?
>>>>> 
>>>>> 
>>>>> done - you'll find the logs at cephdrop folder:
>>>>> slow_requests_recovering_cuttlefish
>>>>> 
>>>>> osd.52 was the one recovering
>>>>> 
>>>>> Thanks!
>>>>> 
>>>>> Greets,
>>>>> Stefan
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-13 21:03                                   ` Stefan Priebe - Profihost AG
@ 2013-08-13 23:11                                     ` Samuel Just
  2013-08-14  7:04                                       ` Stefan Priebe - Profihost AG
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-13 23:11 UTC (permalink / raw)
  To: Stefan Priebe - Profihost AG
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

I'm not sure, but your logs did show that you had >16 recovery ops in
flight, so it's worth a try.  If it doesn't help, you should collect
the same set of logs I'll look again.  Also, there are a few other
patches between 61.7 and current cuttlefish which may help.
-Sam

On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
<s.priebe@profihost.ag> wrote:
>
> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>
>> I just backported a couple of patches from next to fix a bug where we
>> weren't respecting the osd_recovery_max_active config in some cases
>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>> current cuttlefish branch or wait for a 61.8 release.
>
> Thanks! Are you sure that this is the issue? I don't believe that but i'll give it a try. I already tested a branch from sage where he fixed a race regarding max active some weeks ago. So active recovering was max 1 but the issue didn't went away.
>
> Stefan
>
>> -Sam
>>
>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com> wrote:
>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>> -Sam
>>>
>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>> <s.priebe@profihost.ag> wrote:
>>>> Did you take a look?
>>>>
>>>> Stefan
>>>>
>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>
>>>>> Great!  I'll take a look on Monday.
>>>>> -Sam
>>>>>
>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>>>>>> Hi Samual,
>>>>>>
>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>
>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>
>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>
>>>>>>> debug osd = 20
>>>>>>> debug filestore = 20
>>>>>>> debug ms = 1
>>>>>>> debug optracker = 20
>>>>>>>
>>>>>>> on a few osds (including the restarted osd), and upload those osd logs
>>>>>>> along with the ceph.log from before killing the osd until after the
>>>>>>> cluster becomes clean again?
>>>>>>
>>>>>>
>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>> slow_requests_recovering_cuttlefish
>>>>>>
>>>>>> osd.52 was the one recovering
>>>>>>
>>>>>> Thanks!
>>>>>>
>>>>>> Greets,
>>>>>> Stefan
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>>> the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-13 23:11                                     ` Samuel Just
@ 2013-08-14  7:04                                       ` Stefan Priebe - Profihost AG
  2013-08-21 15:28                                         ` Mike Dawson
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe - Profihost AG @ 2013-08-14  7:04 UTC (permalink / raw)
  To: Samuel Just
  Cc: Mike Dawson, josh.durgin, Oliver Francke, ceph-devel, Stefan Hajnoczi

the same problem still occours. Will need to check when i've time to
gather logs again.

Am 14.08.2013 01:11, schrieb Samuel Just:
> I'm not sure, but your logs did show that you had >16 recovery ops in
> flight, so it's worth a try.  If it doesn't help, you should collect
> the same set of logs I'll look again.  Also, there are a few other
> patches between 61.7 and current cuttlefish which may help.
> -Sam
> 
> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
> <s.priebe@profihost.ag> wrote:
>>
>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>
>>> I just backported a couple of patches from next to fix a bug where we
>>> weren't respecting the osd_recovery_max_active config in some cases
>>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>>> current cuttlefish branch or wait for a 61.8 release.
>>
>> Thanks! Are you sure that this is the issue? I don't believe that but i'll give it a try. I already tested a branch from sage where he fixed a race regarding max active some weeks ago. So active recovering was max 1 but the issue didn't went away.
>>
>> Stefan
>>
>>> -Sam
>>>
>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com> wrote:
>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>> -Sam
>>>>
>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>> <s.priebe@profihost.ag> wrote:
>>>>> Did you take a look?
>>>>>
>>>>> Stefan
>>>>>
>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>
>>>>>> Great!  I'll take a look on Monday.
>>>>>> -Sam
>>>>>>
>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>>>>>>> Hi Samual,
>>>>>>>
>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>
>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>
>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>
>>>>>>>> debug osd = 20
>>>>>>>> debug filestore = 20
>>>>>>>> debug ms = 1
>>>>>>>> debug optracker = 20
>>>>>>>>
>>>>>>>> on a few osds (including the restarted osd), and upload those osd logs
>>>>>>>> along with the ceph.log from before killing the osd until after the
>>>>>>>> cluster becomes clean again?
>>>>>>>
>>>>>>>
>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>
>>>>>>> osd.52 was the one recovering
>>>>>>>
>>>>>>> Thanks!
>>>>>>>
>>>>>>> Greets,
>>>>>>> Stefan
>>>>>> --
>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-14  7:04                                       ` Stefan Priebe - Profihost AG
@ 2013-08-21 15:28                                         ` Mike Dawson
  2013-08-21 15:32                                           ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Mike Dawson @ 2013-08-21 15:28 UTC (permalink / raw)
  To: Stefan Priebe - Profihost AG; +Cc: Samuel Just, josh.durgin, ceph-devel

Sam/Josh,

We upgraded from 0.61.7 to 0.67.1 during a maintenance window this 
morning, hoping it would improve this situation, but there was no 
appreciable change.

One node in our cluster fsck'ed after a reboot and got a bit behind. Our 
instances backed by RBD volumes were OK at that point, but once the node 
booted fully and the OSDs started, all Windows instances with rbd 
volumes experienced very choppy performance and were unable to ingest 
video surveillance traffic and commit it to disk. Once the cluster got 
back to HEALTH_OK, they resumed normal operation.

I tried for a time with conservative recovery settings (osd max 
backfills = 1, osd recovery op priority = 1, and osd recovery max active 
= 1). No improvement for the guests. So I went to more aggressive 
settings to get things moving faster. That decreased the duration of the 
outage.

During the entire period of recovery/backfill, the network looked 
fine...no where close to saturation. iowait on all drives look fine as well.

Any ideas?

Thanks,
Mike Dawson


On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
> the same problem still occours. Will need to check when i've time to
> gather logs again.
>
> Am 14.08.2013 01:11, schrieb Samuel Just:
>> I'm not sure, but your logs did show that you had >16 recovery ops in
>> flight, so it's worth a try.  If it doesn't help, you should collect
>> the same set of logs I'll look again.  Also, there are a few other
>> patches between 61.7 and current cuttlefish which may help.
>> -Sam
>>
>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>> <s.priebe@profihost.ag> wrote:
>>>
>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>
>>>> I just backported a couple of patches from next to fix a bug where we
>>>> weren't respecting the osd_recovery_max_active config in some cases
>>>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>>>> current cuttlefish branch or wait for a 61.8 release.
>>>
>>> Thanks! Are you sure that this is the issue? I don't believe that but i'll give it a try. I already tested a branch from sage where he fixed a race regarding max active some weeks ago. So active recovering was max 1 but the issue didn't went away.
>>>
>>> Stefan
>>>
>>>> -Sam
>>>>
>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com> wrote:
>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>> -Sam
>>>>>
>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>> <s.priebe@profihost.ag> wrote:
>>>>>> Did you take a look?
>>>>>>
>>>>>> Stefan
>>>>>>
>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>
>>>>>>> Great!  I'll take a look on Monday.
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>>>>>>>> Hi Samual,
>>>>>>>>
>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>
>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>
>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>
>>>>>>>>> debug osd = 20
>>>>>>>>> debug filestore = 20
>>>>>>>>> debug ms = 1
>>>>>>>>> debug optracker = 20
>>>>>>>>>
>>>>>>>>> on a few osds (including the restarted osd), and upload those osd logs
>>>>>>>>> along with the ceph.log from before killing the osd until after the
>>>>>>>>> cluster becomes clean again?
>>>>>>>>
>>>>>>>>
>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>
>>>>>>>> osd.52 was the one recovering
>>>>>>>>
>>>>>>>> Thanks!
>>>>>>>>
>>>>>>>> Greets,
>>>>>>>> Stefan
>>>>>>> --
>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-21 15:28                                         ` Mike Dawson
@ 2013-08-21 15:32                                           ` Samuel Just
  2013-08-21 16:25                                             ` Yann ROBIN
  2013-08-21 18:21                                             ` Stefan Priebe
  0 siblings, 2 replies; 40+ messages in thread
From: Samuel Just @ 2013-08-21 15:32 UTC (permalink / raw)
  To: Mike Dawson; +Cc: Stefan Priebe - Profihost AG, josh.durgin, ceph-devel

Have you tried setting osd_recovery_clone_overlap to false?  That
seemed to help with Stefan's issue.
-Sam

On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com> wrote:
> Sam/Josh,
>
> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this morning,
> hoping it would improve this situation, but there was no appreciable change.
>
> One node in our cluster fsck'ed after a reboot and got a bit behind. Our
> instances backed by RBD volumes were OK at that point, but once the node
> booted fully and the OSDs started, all Windows instances with rbd volumes
> experienced very choppy performance and were unable to ingest video
> surveillance traffic and commit it to disk. Once the cluster got back to
> HEALTH_OK, they resumed normal operation.
>
> I tried for a time with conservative recovery settings (osd max backfills =
> 1, osd recovery op priority = 1, and osd recovery max active = 1). No
> improvement for the guests. So I went to more aggressive settings to get
> things moving faster. That decreased the duration of the outage.
>
> During the entire period of recovery/backfill, the network looked fine...no
> where close to saturation. iowait on all drives look fine as well.
>
> Any ideas?
>
> Thanks,
> Mike Dawson
>
>
>
> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>
>> the same problem still occours. Will need to check when i've time to
>> gather logs again.
>>
>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>
>>> I'm not sure, but your logs did show that you had >16 recovery ops in
>>> flight, so it's worth a try.  If it doesn't help, you should collect
>>> the same set of logs I'll look again.  Also, there are a few other
>>> patches between 61.7 and current cuttlefish which may help.
>>> -Sam
>>>
>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>> <s.priebe@profihost.ag> wrote:
>>>>
>>>>
>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>
>>>>> I just backported a couple of patches from next to fix a bug where we
>>>>> weren't respecting the osd_recovery_max_active config in some cases
>>>>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>>>>> current cuttlefish branch or wait for a 61.8 release.
>>>>
>>>>
>>>> Thanks! Are you sure that this is the issue? I don't believe that but
>>>> i'll give it a try. I already tested a branch from sage where he fixed a
>>>> race regarding max active some weeks ago. So active recovering was max 1 but
>>>> the issue didn't went away.
>>>>
>>>> Stefan
>>>>
>>>>> -Sam
>>>>>
>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com>
>>>>> wrote:
>>>>>>
>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>> -Sam
>>>>>>
>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>
>>>>>>> Did you take a look?
>>>>>>>
>>>>>>> Stefan
>>>>>>>
>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>
>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>
>>>>>>>>> Hi Samual,
>>>>>>>>>
>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>
>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>
>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>
>>>>>>>>>> debug osd = 20
>>>>>>>>>> debug filestore = 20
>>>>>>>>>> debug ms = 1
>>>>>>>>>> debug optracker = 20
>>>>>>>>>>
>>>>>>>>>> on a few osds (including the restarted osd), and upload those osd
>>>>>>>>>> logs
>>>>>>>>>> along with the ceph.log from before killing the osd until after
>>>>>>>>>> the
>>>>>>>>>> cluster becomes clean again?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>
>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>
>>>>>>>>> Thanks!
>>>>>>>>>
>>>>>>>>> Greets,
>>>>>>>>> Stefan
>>>>>>>>
>>>>>>>> --
>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>> ceph-devel" in
>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>> in
>>>>> the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* RE: still recovery issues with cuttlefish
  2013-08-21 15:32                                           ` Samuel Just
@ 2013-08-21 16:25                                             ` Yann ROBIN
  2013-08-21 17:12                                               ` Samuel Just
  2013-08-21 18:21                                             ` Stefan Priebe
  1 sibling, 1 reply; 40+ messages in thread
From: Yann ROBIN @ 2013-08-21 16:25 UTC (permalink / raw)
  To: Samuel Just, Mike Dawson
  Cc: Stefan Priebe - Profihost AG, josh.durgin, ceph-devel

It's osd recover clone overlap (see http://tracker.ceph.com/issues/5401)

-----Original Message-----
From: ceph-devel-owner@vger.kernel.org [mailto:ceph-devel-owner@vger.kernel.org] On Behalf Of Samuel Just
Sent: mercredi 21 août 2013 17:33
To: Mike Dawson
Cc: Stefan Priebe - Profihost AG; josh.durgin@inktank.com; ceph-devel@vger.kernel.org
Subject: Re: still recovery issues with cuttlefish

Have you tried setting osd_recovery_clone_overlap to false?  That seemed to help with Stefan's issue.
-Sam

On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com> wrote:
> Sam/Josh,
>
> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this 
> morning, hoping it would improve this situation, but there was no appreciable change.
>
> One node in our cluster fsck'ed after a reboot and got a bit behind. 
> Our instances backed by RBD volumes were OK at that point, but once 
> the node booted fully and the OSDs started, all Windows instances with 
> rbd volumes experienced very choppy performance and were unable to 
> ingest video surveillance traffic and commit it to disk. Once the 
> cluster got back to HEALTH_OK, they resumed normal operation.
>
> I tried for a time with conservative recovery settings (osd max 
> backfills = 1, osd recovery op priority = 1, and osd recovery max 
> active = 1). No improvement for the guests. So I went to more 
> aggressive settings to get things moving faster. That decreased the duration of the outage.
>
> During the entire period of recovery/backfill, the network looked 
> fine...no where close to saturation. iowait on all drives look fine as well.
>
> Any ideas?
>
> Thanks,
> Mike Dawson
>
>
>
> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>
>> the same problem still occours. Will need to check when i've time to 
>> gather logs again.
>>
>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>
>>> I'm not sure, but your logs did show that you had >16 recovery ops 
>>> in flight, so it's worth a try.  If it doesn't help, you should 
>>> collect the same set of logs I'll look again.  Also, there are a few 
>>> other patches between 61.7 and current cuttlefish which may help.
>>> -Sam
>>>
>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG 
>>> <s.priebe@profihost.ag> wrote:
>>>>
>>>>
>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>
>>>>> I just backported a couple of patches from next to fix a bug where 
>>>>> we weren't respecting the osd_recovery_max_active config in some 
>>>>> cases (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either 
>>>>> try the current cuttlefish branch or wait for a 61.8 release.
>>>>
>>>>
>>>> Thanks! Are you sure that this is the issue? I don't believe that 
>>>> but i'll give it a try. I already tested a branch from sage where 
>>>> he fixed a race regarding max active some weeks ago. So active 
>>>> recovering was max 1 but the issue didn't went away.
>>>>
>>>> Stefan
>>>>
>>>>> -Sam
>>>>>
>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just 
>>>>> <sam.just@inktank.com>
>>>>> wrote:
>>>>>>
>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>> -Sam
>>>>>>
>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG 
>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>
>>>>>>> Did you take a look?
>>>>>>>
>>>>>>> Stefan
>>>>>>>
>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>
>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe 
>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>
>>>>>>>>> Hi Samual,
>>>>>>>>>
>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>
>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>
>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>
>>>>>>>>>> debug osd = 20
>>>>>>>>>> debug filestore = 20
>>>>>>>>>> debug ms = 1
>>>>>>>>>> debug optracker = 20
>>>>>>>>>>
>>>>>>>>>> on a few osds (including the restarted osd), and upload those 
>>>>>>>>>> osd logs along with the ceph.log from before killing the osd 
>>>>>>>>>> until after the cluster becomes clean again?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>
>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>
>>>>>>>>> Thanks!
>>>>>>>>>
>>>>>>>>> Greets,
>>>>>>>>> Stefan
>>>>>>>>
>>>>>>>> --
>>>>>>>> To unsubscribe from this list: send the line "unsubscribe 
>>>>>>>> ceph-devel" in the body of a message to 
>>>>>>>> majordomo@vger.kernel.org More majordomo info at  
>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>> in
>>>>> the body of a message to majordomo@vger.kernel.org More majordomo 
>>>>> info at  http://vger.kernel.org/majordomo-info.html
>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe 
>>> ceph-devel" in the body of a message to majordomo@vger.kernel.org 
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@vger.kernel.org More majordomo info at  http://vger.kernel.org/majordomo-info.html


--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-21 16:25                                             ` Yann ROBIN
@ 2013-08-21 17:12                                               ` Samuel Just
  2013-08-21 17:55                                                 ` Mike Dawson
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-21 17:12 UTC (permalink / raw)
  To: Yann ROBIN
  Cc: Mike Dawson, Stefan Priebe - Profihost AG, josh.durgin, ceph-devel

Ah, thanks for the correction.
-Sam

On Wed, Aug 21, 2013 at 9:25 AM, Yann ROBIN <yann.robin@youscribe.com> wrote:
> It's osd recover clone overlap (see http://tracker.ceph.com/issues/5401)
>
> -----Original Message-----
> From: ceph-devel-owner@vger.kernel.org [mailto:ceph-devel-owner@vger.kernel.org] On Behalf Of Samuel Just
> Sent: mercredi 21 août 2013 17:33
> To: Mike Dawson
> Cc: Stefan Priebe - Profihost AG; josh.durgin@inktank.com; ceph-devel@vger.kernel.org
> Subject: Re: still recovery issues with cuttlefish
>
> Have you tried setting osd_recovery_clone_overlap to false?  That seemed to help with Stefan's issue.
> -Sam
>
> On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com> wrote:
>> Sam/Josh,
>>
>> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this
>> morning, hoping it would improve this situation, but there was no appreciable change.
>>
>> One node in our cluster fsck'ed after a reboot and got a bit behind.
>> Our instances backed by RBD volumes were OK at that point, but once
>> the node booted fully and the OSDs started, all Windows instances with
>> rbd volumes experienced very choppy performance and were unable to
>> ingest video surveillance traffic and commit it to disk. Once the
>> cluster got back to HEALTH_OK, they resumed normal operation.
>>
>> I tried for a time with conservative recovery settings (osd max
>> backfills = 1, osd recovery op priority = 1, and osd recovery max
>> active = 1). No improvement for the guests. So I went to more
>> aggressive settings to get things moving faster. That decreased the duration of the outage.
>>
>> During the entire period of recovery/backfill, the network looked
>> fine...no where close to saturation. iowait on all drives look fine as well.
>>
>> Any ideas?
>>
>> Thanks,
>> Mike Dawson
>>
>>
>>
>> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>>
>>> the same problem still occours. Will need to check when i've time to
>>> gather logs again.
>>>
>>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>>
>>>> I'm not sure, but your logs did show that you had >16 recovery ops
>>>> in flight, so it's worth a try.  If it doesn't help, you should
>>>> collect the same set of logs I'll look again.  Also, there are a few
>>>> other patches between 61.7 and current cuttlefish which may help.
>>>> -Sam
>>>>
>>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>>> <s.priebe@profihost.ag> wrote:
>>>>>
>>>>>
>>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>
>>>>>> I just backported a couple of patches from next to fix a bug where
>>>>>> we weren't respecting the osd_recovery_max_active config in some
>>>>>> cases (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either
>>>>>> try the current cuttlefish branch or wait for a 61.8 release.
>>>>>
>>>>>
>>>>> Thanks! Are you sure that this is the issue? I don't believe that
>>>>> but i'll give it a try. I already tested a branch from sage where
>>>>> he fixed a race regarding max active some weeks ago. So active
>>>>> recovering was max 1 but the issue didn't went away.
>>>>>
>>>>> Stefan
>>>>>
>>>>>> -Sam
>>>>>>
>>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just
>>>>>> <sam.just@inktank.com>
>>>>>> wrote:
>>>>>>>
>>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>
>>>>>>>> Did you take a look?
>>>>>>>>
>>>>>>>> Stefan
>>>>>>>>
>>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>>
>>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>>> -Sam
>>>>>>>>>
>>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>
>>>>>>>>>> Hi Samual,
>>>>>>>>>>
>>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>>
>>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>>
>>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>>
>>>>>>>>>>> debug osd = 20
>>>>>>>>>>> debug filestore = 20
>>>>>>>>>>> debug ms = 1
>>>>>>>>>>> debug optracker = 20
>>>>>>>>>>>
>>>>>>>>>>> on a few osds (including the restarted osd), and upload those
>>>>>>>>>>> osd logs along with the ceph.log from before killing the osd
>>>>>>>>>>> until after the cluster becomes clean again?
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>>
>>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>>
>>>>>>>>>> Thanks!
>>>>>>>>>>
>>>>>>>>>> Greets,
>>>>>>>>>> Stefan
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>> ceph-devel" in the body of a message to
>>>>>>>>> majordomo@vger.kernel.org More majordomo info at
>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>>> --
>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>> in
>>>>>> the body of a message to majordomo@vger.kernel.org More majordomo
>>>>>> info at  http://vger.kernel.org/majordomo-info.html
>>>>
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe
>>>> ceph-devel" in the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>
>>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@vger.kernel.org More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-21 17:12                                               ` Samuel Just
@ 2013-08-21 17:55                                                 ` Mike Dawson
  2013-08-21 18:05                                                   ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Mike Dawson @ 2013-08-21 17:55 UTC (permalink / raw)
  To: Samuel Just
  Cc: Yann ROBIN, Stefan Priebe - Profihost AG, josh.durgin, ceph-devel

Sam,

Tried it. Injected with 'ceph tell osd.* injectargs -- 
--no_osd_recover_clone_overlap', then stopped one OSD for ~1 minute. 
Upon restart, all my Windows VMs have issues until HEALTH_OK.

The recovery was taking an abnormally long time, so I reverted away from 
--no_osd_recover_clone_overlap after about 10mins, to get back to HEALTH_OK.

Interestingly, a Raring guest running a different video surveillance 
package proceeded without any issue whatsoever.

Here is an image of the traffic to some of these Windows guests:

http://www.gammacode.com/upload/rbd-hang-with-clone-overlap.jpg

Ceph is outside of HEALTH_OK between ~12:55 and 13:10. Most of these 
instances rebooted due to an app error caused by the i/o hang shortly 
after 13:10.

These Windows instances are booted as COW clones from a Glance image 
using Cinder. They also have a second RBD volume for bulk storage. I'm 
using qemu 1.5.2.

Thanks,
Mike


On 8/21/2013 1:12 PM, Samuel Just wrote:
> Ah, thanks for the correction.
> -Sam
>
> On Wed, Aug 21, 2013 at 9:25 AM, Yann ROBIN <yann.robin@youscribe.com> wrote:
>> It's osd recover clone overlap (see http://tracker.ceph.com/issues/5401)
>>
>> -----Original Message-----
>> From: ceph-devel-owner@vger.kernel.org [mailto:ceph-devel-owner@vger.kernel.org] On Behalf Of Samuel Just
>> Sent: mercredi 21 août 2013 17:33
>> To: Mike Dawson
>> Cc: Stefan Priebe - Profihost AG; josh.durgin@inktank.com; ceph-devel@vger.kernel.org
>> Subject: Re: still recovery issues with cuttlefish
>>
>> Have you tried setting osd_recovery_clone_overlap to false?  That seemed to help with Stefan's issue.
>> -Sam
>>
>> On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com> wrote:
>>> Sam/Josh,
>>>
>>> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this
>>> morning, hoping it would improve this situation, but there was no appreciable change.
>>>
>>> One node in our cluster fsck'ed after a reboot and got a bit behind.
>>> Our instances backed by RBD volumes were OK at that point, but once
>>> the node booted fully and the OSDs started, all Windows instances with
>>> rbd volumes experienced very choppy performance and were unable to
>>> ingest video surveillance traffic and commit it to disk. Once the
>>> cluster got back to HEALTH_OK, they resumed normal operation.
>>>
>>> I tried for a time with conservative recovery settings (osd max
>>> backfills = 1, osd recovery op priority = 1, and osd recovery max
>>> active = 1). No improvement for the guests. So I went to more
>>> aggressive settings to get things moving faster. That decreased the duration of the outage.
>>>
>>> During the entire period of recovery/backfill, the network looked
>>> fine...no where close to saturation. iowait on all drives look fine as well.
>>>
>>> Any ideas?
>>>
>>> Thanks,
>>> Mike Dawson
>>>
>>>
>>>
>>> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>>>
>>>> the same problem still occours. Will need to check when i've time to
>>>> gather logs again.
>>>>
>>>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>>>
>>>>> I'm not sure, but your logs did show that you had >16 recovery ops
>>>>> in flight, so it's worth a try.  If it doesn't help, you should
>>>>> collect the same set of logs I'll look again.  Also, there are a few
>>>>> other patches between 61.7 and current cuttlefish which may help.
>>>>> -Sam
>>>>>
>>>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>
>>>>>>
>>>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>
>>>>>>> I just backported a couple of patches from next to fix a bug where
>>>>>>> we weren't respecting the osd_recovery_max_active config in some
>>>>>>> cases (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either
>>>>>>> try the current cuttlefish branch or wait for a 61.8 release.
>>>>>>
>>>>>>
>>>>>> Thanks! Are you sure that this is the issue? I don't believe that
>>>>>> but i'll give it a try. I already tested a branch from sage where
>>>>>> he fixed a race regarding max active some weeks ago. So active
>>>>>> recovering was max 1 but the issue didn't went away.
>>>>>>
>>>>>> Stefan
>>>>>>
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just
>>>>>>> <sam.just@inktank.com>
>>>>>>> wrote:
>>>>>>>>
>>>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>
>>>>>>>>> Did you take a look?
>>>>>>>>>
>>>>>>>>> Stefan
>>>>>>>>>
>>>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>>>
>>>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>>>> -Sam
>>>>>>>>>>
>>>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>
>>>>>>>>>>> Hi Samual,
>>>>>>>>>>>
>>>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>>>
>>>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>>>
>>>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>>>
>>>>>>>>>>>> debug osd = 20
>>>>>>>>>>>> debug filestore = 20
>>>>>>>>>>>> debug ms = 1
>>>>>>>>>>>> debug optracker = 20
>>>>>>>>>>>>
>>>>>>>>>>>> on a few osds (including the restarted osd), and upload those
>>>>>>>>>>>> osd logs along with the ceph.log from before killing the osd
>>>>>>>>>>>> until after the cluster becomes clean again?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>>>
>>>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>>>
>>>>>>>>>>> Thanks!
>>>>>>>>>>>
>>>>>>>>>>> Greets,
>>>>>>>>>>> Stefan
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>> ceph-devel" in the body of a message to
>>>>>>>>>> majordomo@vger.kernel.org More majordomo info at
>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>
>>>>>>> --
>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>>> in
>>>>>>> the body of a message to majordomo@vger.kernel.org More majordomo
>>>>>>> info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>> ceph-devel" in the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@vger.kernel.org More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-21 17:55                                                 ` Mike Dawson
@ 2013-08-21 18:05                                                   ` Samuel Just
  0 siblings, 0 replies; 40+ messages in thread
From: Samuel Just @ 2013-08-21 18:05 UTC (permalink / raw)
  To: Mike Dawson
  Cc: Yann ROBIN, Stefan Priebe - Profihost AG, josh.durgin, ceph-devel

If the raring guest was fine, I suspect that the issue is not on the OSDs.
-Sam

On Wed, Aug 21, 2013 at 10:55 AM, Mike Dawson <mike.dawson@cloudapt.com> wrote:
> Sam,
>
> Tried it. Injected with 'ceph tell osd.* injectargs --
> --no_osd_recover_clone_overlap', then stopped one OSD for ~1 minute. Upon
> restart, all my Windows VMs have issues until HEALTH_OK.
>
> The recovery was taking an abnormally long time, so I reverted away from
> --no_osd_recover_clone_overlap after about 10mins, to get back to HEALTH_OK.
>
> Interestingly, a Raring guest running a different video surveillance package
> proceeded without any issue whatsoever.
>
> Here is an image of the traffic to some of these Windows guests:
>
> http://www.gammacode.com/upload/rbd-hang-with-clone-overlap.jpg
>
> Ceph is outside of HEALTH_OK between ~12:55 and 13:10. Most of these
> instances rebooted due to an app error caused by the i/o hang shortly after
> 13:10.
>
> These Windows instances are booted as COW clones from a Glance image using
> Cinder. They also have a second RBD volume for bulk storage. I'm using qemu
> 1.5.2.
>
> Thanks,
> Mike
>
>
>
> On 8/21/2013 1:12 PM, Samuel Just wrote:
>>
>> Ah, thanks for the correction.
>> -Sam
>>
>> On Wed, Aug 21, 2013 at 9:25 AM, Yann ROBIN <yann.robin@youscribe.com>
>> wrote:
>>>
>>> It's osd recover clone overlap (see http://tracker.ceph.com/issues/5401)
>>>
>>> -----Original Message-----
>>> From: ceph-devel-owner@vger.kernel.org
>>> [mailto:ceph-devel-owner@vger.kernel.org] On Behalf Of Samuel Just
>>> Sent: mercredi 21 août 2013 17:33
>>> To: Mike Dawson
>>> Cc: Stefan Priebe - Profihost AG; josh.durgin@inktank.com;
>>> ceph-devel@vger.kernel.org
>>> Subject: Re: still recovery issues with cuttlefish
>>>
>>> Have you tried setting osd_recovery_clone_overlap to false?  That seemed
>>> to help with Stefan's issue.
>>> -Sam
>>>
>>> On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com>
>>> wrote:
>>>>
>>>> Sam/Josh,
>>>>
>>>> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this
>>>> morning, hoping it would improve this situation, but there was no
>>>> appreciable change.
>>>>
>>>> One node in our cluster fsck'ed after a reboot and got a bit behind.
>>>> Our instances backed by RBD volumes were OK at that point, but once
>>>> the node booted fully and the OSDs started, all Windows instances with
>>>> rbd volumes experienced very choppy performance and were unable to
>>>> ingest video surveillance traffic and commit it to disk. Once the
>>>> cluster got back to HEALTH_OK, they resumed normal operation.
>>>>
>>>> I tried for a time with conservative recovery settings (osd max
>>>> backfills = 1, osd recovery op priority = 1, and osd recovery max
>>>> active = 1). No improvement for the guests. So I went to more
>>>> aggressive settings to get things moving faster. That decreased the
>>>> duration of the outage.
>>>>
>>>> During the entire period of recovery/backfill, the network looked
>>>> fine...no where close to saturation. iowait on all drives look fine as
>>>> well.
>>>>
>>>> Any ideas?
>>>>
>>>> Thanks,
>>>> Mike Dawson
>>>>
>>>>
>>>>
>>>> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>>>>
>>>>>
>>>>> the same problem still occours. Will need to check when i've time to
>>>>> gather logs again.
>>>>>
>>>>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>>>>
>>>>>>
>>>>>> I'm not sure, but your logs did show that you had >16 recovery ops
>>>>>> in flight, so it's worth a try.  If it doesn't help, you should
>>>>>> collect the same set of logs I'll look again.  Also, there are a few
>>>>>> other patches between 61.7 and current cuttlefish which may help.
>>>>>> -Sam
>>>>>>
>>>>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>
>>>>>>>> I just backported a couple of patches from next to fix a bug where
>>>>>>>> we weren't respecting the osd_recovery_max_active config in some
>>>>>>>> cases (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either
>>>>>>>> try the current cuttlefish branch or wait for a 61.8 release.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Thanks! Are you sure that this is the issue? I don't believe that
>>>>>>> but i'll give it a try. I already tested a branch from sage where
>>>>>>> he fixed a race regarding max active some weeks ago. So active
>>>>>>> recovering was max 1 but the issue didn't went away.
>>>>>>>
>>>>>>> Stefan
>>>>>>>
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just
>>>>>>>> <sam.just@inktank.com>
>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>>>>> -Sam
>>>>>>>>>
>>>>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Did you take a look?
>>>>>>>>>>
>>>>>>>>>> Stefan
>>>>>>>>>>
>>>>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>>>>
>>>>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>>>>> -Sam
>>>>>>>>>>>
>>>>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Hi Samual,
>>>>>>>>>>>>
>>>>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>>>>
>>>>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>>>>
>>>>>>>>>>>>> debug osd = 20
>>>>>>>>>>>>> debug filestore = 20
>>>>>>>>>>>>> debug ms = 1
>>>>>>>>>>>>> debug optracker = 20
>>>>>>>>>>>>>
>>>>>>>>>>>>> on a few osds (including the restarted osd), and upload those
>>>>>>>>>>>>> osd logs along with the ceph.log from before killing the osd
>>>>>>>>>>>>> until after the cluster becomes clean again?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>>>>
>>>>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>
>>>>>>>>>>>> Greets,
>>>>>>>>>>>> Stefan
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>>> ceph-devel" in the body of a message to
>>>>>>>>>>> majordomo@vger.kernel.org More majordomo info at
>>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>> ceph-devel"
>>>>>>>> in
>>>>>>>> the body of a message to majordomo@vger.kernel.org More majordomo
>>>>>>>> info at  http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>>>
>>>>>> --
>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>> ceph-devel" in the body of a message to majordomo@vger.kernel.org
>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org More majordomo info at
>>> http://vger.kernel.org/majordomo-info.html
>>>
>>>
>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-21 15:32                                           ` Samuel Just
  2013-08-21 16:25                                             ` Yann ROBIN
@ 2013-08-21 18:21                                             ` Stefan Priebe
  2013-08-21 19:13                                               ` Samuel Just
  1 sibling, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-21 18:21 UTC (permalink / raw)
  To: Samuel Just; +Cc: Mike Dawson, josh.durgin, ceph-devel, Sage Weil

Am 21.08.2013 17:32, schrieb Samuel Just:
> Have you tried setting osd_recovery_clone_overlap to false?  That
> seemed to help with Stefan's issue.

This might sound a bug harsh but maybe due to my limited english skills ;-)

I still think that Cephs recovery system is broken by design. If an OSD 
comes back (was offline) all write requests regarding PGs where this one 
is primary are targeted immediatly to this OSD. If this one is not 
up2date for an PG it tries to recover that one immediatly which costs 
4MB / block. If you have a lot of small write all over your OSDs and PGs 
you're sucked as your OSD has to recover ALL it's PGs immediatly or at 
least lots of them WHICH can't work. This is totally crazy.

I think the right way would be:
1.) if an OSD goes down the replicas got primaries

or

2.) an OSD which does not have an up2date PG should redirect to the OSD 
holding the secondary or third replica.

Both results in being able to have a really smooth and slow recovery 
without any stress even under heavy 4K workloads like rbd backed VMs.

Thanks for reading!

Greets Stefan


> -Sam
>
> On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com> wrote:
>> Sam/Josh,
>>
>> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this morning,
>> hoping it would improve this situation, but there was no appreciable change.
>>
>> One node in our cluster fsck'ed after a reboot and got a bit behind. Our
>> instances backed by RBD volumes were OK at that point, but once the node
>> booted fully and the OSDs started, all Windows instances with rbd volumes
>> experienced very choppy performance and were unable to ingest video
>> surveillance traffic and commit it to disk. Once the cluster got back to
>> HEALTH_OK, they resumed normal operation.
>>
>> I tried for a time with conservative recovery settings (osd max backfills =
>> 1, osd recovery op priority = 1, and osd recovery max active = 1). No
>> improvement for the guests. So I went to more aggressive settings to get
>> things moving faster. That decreased the duration of the outage.
>>
>> During the entire period of recovery/backfill, the network looked fine...no
>> where close to saturation. iowait on all drives look fine as well.
>>
>> Any ideas?
>>
>> Thanks,
>> Mike Dawson
>>
>>
>>
>> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>>
>>> the same problem still occours. Will need to check when i've time to
>>> gather logs again.
>>>
>>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>>
>>>> I'm not sure, but your logs did show that you had >16 recovery ops in
>>>> flight, so it's worth a try.  If it doesn't help, you should collect
>>>> the same set of logs I'll look again.  Also, there are a few other
>>>> patches between 61.7 and current cuttlefish which may help.
>>>> -Sam
>>>>
>>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>>> <s.priebe@profihost.ag> wrote:
>>>>>
>>>>>
>>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>
>>>>>> I just backported a couple of patches from next to fix a bug where we
>>>>>> weren't respecting the osd_recovery_max_active config in some cases
>>>>>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>>>>>> current cuttlefish branch or wait for a 61.8 release.
>>>>>
>>>>>
>>>>> Thanks! Are you sure that this is the issue? I don't believe that but
>>>>> i'll give it a try. I already tested a branch from sage where he fixed a
>>>>> race regarding max active some weeks ago. So active recovering was max 1 but
>>>>> the issue didn't went away.
>>>>>
>>>>> Stefan
>>>>>
>>>>>> -Sam
>>>>>>
>>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com>
>>>>>> wrote:
>>>>>>>
>>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>
>>>>>>>> Did you take a look?
>>>>>>>>
>>>>>>>> Stefan
>>>>>>>>
>>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>>
>>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>>> -Sam
>>>>>>>>>
>>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>
>>>>>>>>>> Hi Samual,
>>>>>>>>>>
>>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>>
>>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>>
>>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>>
>>>>>>>>>>> debug osd = 20
>>>>>>>>>>> debug filestore = 20
>>>>>>>>>>> debug ms = 1
>>>>>>>>>>> debug optracker = 20
>>>>>>>>>>>
>>>>>>>>>>> on a few osds (including the restarted osd), and upload those osd
>>>>>>>>>>> logs
>>>>>>>>>>> along with the ceph.log from before killing the osd until after
>>>>>>>>>>> the
>>>>>>>>>>> cluster becomes clean again?
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>>
>>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>>
>>>>>>>>>> Thanks!
>>>>>>>>>>
>>>>>>>>>> Greets,
>>>>>>>>>> Stefan
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>> ceph-devel" in
>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>>> --
>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>> in
>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>
>>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-21 18:21                                             ` Stefan Priebe
@ 2013-08-21 19:13                                               ` Samuel Just
  2013-08-21 19:37                                                 ` Stefan Priebe
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-21 19:13 UTC (permalink / raw)
  To: Stefan Priebe; +Cc: Mike Dawson, josh.durgin, ceph-devel, Sage Weil

As long as the request is for an object which is up to date on the
primary, the request will be served without waiting for recovery.  A
request only waits on recovery if the particular object being read or
written must be recovered.  Your issue was that recovering the
particular object being requested was unreasonably slow due to
silliness in the recovery code which you disabled by disabling
osd_recover_clone_overlap.

In cases where the primary osd is significantly behind, we do make one
of the other osds primary during recovery in order to expedite
requests (pgs in this state are shown as remapped).
-Sam

On Wed, Aug 21, 2013 at 11:21 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> Am 21.08.2013 17:32, schrieb Samuel Just:
>
>> Have you tried setting osd_recovery_clone_overlap to false?  That
>> seemed to help with Stefan's issue.
>
>
> This might sound a bug harsh but maybe due to my limited english skills ;-)
>
> I still think that Cephs recovery system is broken by design. If an OSD
> comes back (was offline) all write requests regarding PGs where this one is
> primary are targeted immediatly to this OSD. If this one is not up2date for
> an PG it tries to recover that one immediatly which costs 4MB / block. If
> you have a lot of small write all over your OSDs and PGs you're sucked as
> your OSD has to recover ALL it's PGs immediatly or at least lots of them
> WHICH can't work. This is totally crazy.
>
> I think the right way would be:
> 1.) if an OSD goes down the replicas got primaries
>
> or
>
> 2.) an OSD which does not have an up2date PG should redirect to the OSD
> holding the secondary or third replica.
>
> Both results in being able to have a really smooth and slow recovery without
> any stress even under heavy 4K workloads like rbd backed VMs.
>
> Thanks for reading!
>
> Greets Stefan
>
>
>
>> -Sam
>>
>> On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com>
>> wrote:
>>>
>>> Sam/Josh,
>>>
>>> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this
>>> morning,
>>> hoping it would improve this situation, but there was no appreciable
>>> change.
>>>
>>> One node in our cluster fsck'ed after a reboot and got a bit behind. Our
>>> instances backed by RBD volumes were OK at that point, but once the node
>>> booted fully and the OSDs started, all Windows instances with rbd volumes
>>> experienced very choppy performance and were unable to ingest video
>>> surveillance traffic and commit it to disk. Once the cluster got back to
>>> HEALTH_OK, they resumed normal operation.
>>>
>>> I tried for a time with conservative recovery settings (osd max backfills
>>> =
>>> 1, osd recovery op priority = 1, and osd recovery max active = 1). No
>>> improvement for the guests. So I went to more aggressive settings to get
>>> things moving faster. That decreased the duration of the outage.
>>>
>>> During the entire period of recovery/backfill, the network looked
>>> fine...no
>>> where close to saturation. iowait on all drives look fine as well.
>>>
>>> Any ideas?
>>>
>>> Thanks,
>>> Mike Dawson
>>>
>>>
>>>
>>> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>>>
>>>>
>>>> the same problem still occours. Will need to check when i've time to
>>>> gather logs again.
>>>>
>>>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>>>
>>>>>
>>>>> I'm not sure, but your logs did show that you had >16 recovery ops in
>>>>> flight, so it's worth a try.  If it doesn't help, you should collect
>>>>> the same set of logs I'll look again.  Also, there are a few other
>>>>> patches between 61.7 and current cuttlefish which may help.
>>>>> -Sam
>>>>>
>>>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>
>>>>>>> I just backported a couple of patches from next to fix a bug where we
>>>>>>> weren't respecting the osd_recovery_max_active config in some cases
>>>>>>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>>>>>>> current cuttlefish branch or wait for a 61.8 release.
>>>>>>
>>>>>>
>>>>>>
>>>>>> Thanks! Are you sure that this is the issue? I don't believe that but
>>>>>> i'll give it a try. I already tested a branch from sage where he fixed
>>>>>> a
>>>>>> race regarding max active some weeks ago. So active recovering was max
>>>>>> 1 but
>>>>>> the issue didn't went away.
>>>>>>
>>>>>> Stefan
>>>>>>
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com>
>>>>>>> wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Did you take a look?
>>>>>>>>>
>>>>>>>>> Stefan
>>>>>>>>>
>>>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>>>
>>>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>>>> -Sam
>>>>>>>>>>
>>>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Hi Samual,
>>>>>>>>>>>
>>>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>>>
>>>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>>>
>>>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>>>
>>>>>>>>>>>> debug osd = 20
>>>>>>>>>>>> debug filestore = 20
>>>>>>>>>>>> debug ms = 1
>>>>>>>>>>>> debug optracker = 20
>>>>>>>>>>>>
>>>>>>>>>>>> on a few osds (including the restarted osd), and upload those
>>>>>>>>>>>> osd
>>>>>>>>>>>> logs
>>>>>>>>>>>> along with the ceph.log from before killing the osd until after
>>>>>>>>>>>> the
>>>>>>>>>>>> cluster becomes clean again?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>>>
>>>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>>>
>>>>>>>>>>> Thanks!
>>>>>>>>>>>
>>>>>>>>>>> Greets,
>>>>>>>>>>> Stefan
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>> ceph-devel" in
>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>>> in
>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>>>
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>> in
>>>>> the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-21 19:13                                               ` Samuel Just
@ 2013-08-21 19:37                                                 ` Stefan Priebe
  2013-08-22  3:34                                                   ` Samuel Just
  0 siblings, 1 reply; 40+ messages in thread
From: Stefan Priebe @ 2013-08-21 19:37 UTC (permalink / raw)
  To: Samuel Just; +Cc: Mike Dawson, josh.durgin, ceph-devel, Sage Weil

Hi Sam,
Am 21.08.2013 21:13, schrieb Samuel Just:
> As long as the request is for an object which is up to date on the
> primary, the request will be served without waiting for recovery.

Sure but remember if you have VM random 4K workload a lot of objects go 
out of date pretty soon.

 > A request only waits on recovery if the particular object being read or
> written must be recovered.

Yes but on 4k load this can be a lot.

> Your issue was that recovering the
> particular object being requested was unreasonably slow due to
> silliness in the recovery code which you disabled by disabling
> osd_recover_clone_overlap.

Yes and no. It's better now but far away from being good or perfect. My 
VMs do not crash anymore but i still have a bunch of slow requests (just 
around 10 messages) and still a VERY high I/O load on the disks during 
recovery.

> In cases where the primary osd is significantly behind, we do make one
> of the other osds primary during recovery in order to expedite
> requests (pgs in this state are shown as remapped).

oh never seen that but at least in my case even 60s are a very long 
timeframe and the OSD is very stressed during recovery. Is it possible 
for me to set this value?

Stefan

> -Sam
>
> On Wed, Aug 21, 2013 at 11:21 AM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> Am 21.08.2013 17:32, schrieb Samuel Just:
>>
>>> Have you tried setting osd_recovery_clone_overlap to false?  That
>>> seemed to help with Stefan's issue.
>>
>>
>> This might sound a bug harsh but maybe due to my limited english skills ;-)
>>
>> I still think that Cephs recovery system is broken by design. If an OSD
>> comes back (was offline) all write requests regarding PGs where this one is
>> primary are targeted immediatly to this OSD. If this one is not up2date for
>> an PG it tries to recover that one immediatly which costs 4MB / block. If
>> you have a lot of small write all over your OSDs and PGs you're sucked as
>> your OSD has to recover ALL it's PGs immediatly or at least lots of them
>> WHICH can't work. This is totally crazy.
>>
>> I think the right way would be:
>> 1.) if an OSD goes down the replicas got primaries
>>
>> or
>>
>> 2.) an OSD which does not have an up2date PG should redirect to the OSD
>> holding the secondary or third replica.
>>
>> Both results in being able to have a really smooth and slow recovery without
>> any stress even under heavy 4K workloads like rbd backed VMs.
>>
>> Thanks for reading!
>>
>> Greets Stefan
>>
>>
>>
>>> -Sam
>>>
>>> On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com>
>>> wrote:
>>>>
>>>> Sam/Josh,
>>>>
>>>> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this
>>>> morning,
>>>> hoping it would improve this situation, but there was no appreciable
>>>> change.
>>>>
>>>> One node in our cluster fsck'ed after a reboot and got a bit behind. Our
>>>> instances backed by RBD volumes were OK at that point, but once the node
>>>> booted fully and the OSDs started, all Windows instances with rbd volumes
>>>> experienced very choppy performance and were unable to ingest video
>>>> surveillance traffic and commit it to disk. Once the cluster got back to
>>>> HEALTH_OK, they resumed normal operation.
>>>>
>>>> I tried for a time with conservative recovery settings (osd max backfills
>>>> =
>>>> 1, osd recovery op priority = 1, and osd recovery max active = 1). No
>>>> improvement for the guests. So I went to more aggressive settings to get
>>>> things moving faster. That decreased the duration of the outage.
>>>>
>>>> During the entire period of recovery/backfill, the network looked
>>>> fine...no
>>>> where close to saturation. iowait on all drives look fine as well.
>>>>
>>>> Any ideas?
>>>>
>>>> Thanks,
>>>> Mike Dawson
>>>>
>>>>
>>>>
>>>> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>>>>
>>>>>
>>>>> the same problem still occours. Will need to check when i've time to
>>>>> gather logs again.
>>>>>
>>>>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>>>>
>>>>>>
>>>>>> I'm not sure, but your logs did show that you had >16 recovery ops in
>>>>>> flight, so it's worth a try.  If it doesn't help, you should collect
>>>>>> the same set of logs I'll look again.  Also, there are a few other
>>>>>> patches between 61.7 and current cuttlefish which may help.
>>>>>> -Sam
>>>>>>
>>>>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>
>>>>>>>> I just backported a couple of patches from next to fix a bug where we
>>>>>>>> weren't respecting the osd_recovery_max_active config in some cases
>>>>>>>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>>>>>>>> current cuttlefish branch or wait for a 61.8 release.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Thanks! Are you sure that this is the issue? I don't believe that but
>>>>>>> i'll give it a try. I already tested a branch from sage where he fixed
>>>>>>> a
>>>>>>> race regarding max active some weeks ago. So active recovering was max
>>>>>>> 1 but
>>>>>>> the issue didn't went away.
>>>>>>>
>>>>>>> Stefan
>>>>>>>
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just <sam.just@inktank.com>
>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>>>>> -Sam
>>>>>>>>>
>>>>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Did you take a look?
>>>>>>>>>>
>>>>>>>>>> Stefan
>>>>>>>>>>
>>>>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>>>>
>>>>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>>>>> -Sam
>>>>>>>>>>>
>>>>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Hi Samual,
>>>>>>>>>>>>
>>>>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>>>>
>>>>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>>>>
>>>>>>>>>>>>> debug osd = 20
>>>>>>>>>>>>> debug filestore = 20
>>>>>>>>>>>>> debug ms = 1
>>>>>>>>>>>>> debug optracker = 20
>>>>>>>>>>>>>
>>>>>>>>>>>>> on a few osds (including the restarted osd), and upload those
>>>>>>>>>>>>> osd
>>>>>>>>>>>>> logs
>>>>>>>>>>>>> along with the ceph.log from before killing the osd until after
>>>>>>>>>>>>> the
>>>>>>>>>>>>> cluster becomes clean again?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>>>>
>>>>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>
>>>>>>>>>>>> Greets,
>>>>>>>>>>>> Stefan
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>>> ceph-devel" in
>>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>>>> in
>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>>>
>>>>>> --
>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>> in
>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>
>>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-21 19:37                                                 ` Stefan Priebe
@ 2013-08-22  3:34                                                   ` Samuel Just
  2013-08-22  7:41                                                     ` Stefan Priebe - Profihost AG
  0 siblings, 1 reply; 40+ messages in thread
From: Samuel Just @ 2013-08-22  3:34 UTC (permalink / raw)
  To: Stefan Priebe; +Cc: Mike Dawson, josh.durgin, ceph-devel, Sage Weil

It's not really possible at this time to control that limit because
changing the primary is actually fairly expensive and doing it
unnecessarily would probably make the situation much worse (it's
mostly necessary for backfilling, which is expensive anyway).  It
seems like forwarding IO on an object which needs to be recovered to a
replica with the object would be the next step.  Certainly something
to consider for the future.
-Sam

On Wed, Aug 21, 2013 at 12:37 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> Hi Sam,
> Am 21.08.2013 21:13, schrieb Samuel Just:
>
>> As long as the request is for an object which is up to date on the
>> primary, the request will be served without waiting for recovery.
>
>
> Sure but remember if you have VM random 4K workload a lot of objects go out
> of date pretty soon.
>
>
>> A request only waits on recovery if the particular object being read or
>>
>> written must be recovered.
>
>
> Yes but on 4k load this can be a lot.
>
>
>> Your issue was that recovering the
>> particular object being requested was unreasonably slow due to
>> silliness in the recovery code which you disabled by disabling
>> osd_recover_clone_overlap.
>
>
> Yes and no. It's better now but far away from being good or perfect. My VMs
> do not crash anymore but i still have a bunch of slow requests (just around
> 10 messages) and still a VERY high I/O load on the disks during recovery.
>
>
>> In cases where the primary osd is significantly behind, we do make one
>> of the other osds primary during recovery in order to expedite
>> requests (pgs in this state are shown as remapped).
>
>
> oh never seen that but at least in my case even 60s are a very long
> timeframe and the OSD is very stressed during recovery. Is it possible for
> me to set this value?
>
>
> Stefan
>
>> -Sam
>>
>> On Wed, Aug 21, 2013 at 11:21 AM, Stefan Priebe <s.priebe@profihost.ag>
>> wrote:
>>>
>>> Am 21.08.2013 17:32, schrieb Samuel Just:
>>>
>>>> Have you tried setting osd_recovery_clone_overlap to false?  That
>>>> seemed to help with Stefan's issue.
>>>
>>>
>>>
>>> This might sound a bug harsh but maybe due to my limited english skills
>>> ;-)
>>>
>>> I still think that Cephs recovery system is broken by design. If an OSD
>>> comes back (was offline) all write requests regarding PGs where this one
>>> is
>>> primary are targeted immediatly to this OSD. If this one is not up2date
>>> for
>>> an PG it tries to recover that one immediatly which costs 4MB / block. If
>>> you have a lot of small write all over your OSDs and PGs you're sucked as
>>> your OSD has to recover ALL it's PGs immediatly or at least lots of them
>>> WHICH can't work. This is totally crazy.
>>>
>>> I think the right way would be:
>>> 1.) if an OSD goes down the replicas got primaries
>>>
>>> or
>>>
>>> 2.) an OSD which does not have an up2date PG should redirect to the OSD
>>> holding the secondary or third replica.
>>>
>>> Both results in being able to have a really smooth and slow recovery
>>> without
>>> any stress even under heavy 4K workloads like rbd backed VMs.
>>>
>>> Thanks for reading!
>>>
>>> Greets Stefan
>>>
>>>
>>>
>>>> -Sam
>>>>
>>>> On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com>
>>>> wrote:
>>>>>
>>>>>
>>>>> Sam/Josh,
>>>>>
>>>>> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this
>>>>> morning,
>>>>> hoping it would improve this situation, but there was no appreciable
>>>>> change.
>>>>>
>>>>> One node in our cluster fsck'ed after a reboot and got a bit behind.
>>>>> Our
>>>>> instances backed by RBD volumes were OK at that point, but once the
>>>>> node
>>>>> booted fully and the OSDs started, all Windows instances with rbd
>>>>> volumes
>>>>> experienced very choppy performance and were unable to ingest video
>>>>> surveillance traffic and commit it to disk. Once the cluster got back
>>>>> to
>>>>> HEALTH_OK, they resumed normal operation.
>>>>>
>>>>> I tried for a time with conservative recovery settings (osd max
>>>>> backfills
>>>>> =
>>>>> 1, osd recovery op priority = 1, and osd recovery max active = 1). No
>>>>> improvement for the guests. So I went to more aggressive settings to
>>>>> get
>>>>> things moving faster. That decreased the duration of the outage.
>>>>>
>>>>> During the entire period of recovery/backfill, the network looked
>>>>> fine...no
>>>>> where close to saturation. iowait on all drives look fine as well.
>>>>>
>>>>> Any ideas?
>>>>>
>>>>> Thanks,
>>>>> Mike Dawson
>>>>>
>>>>>
>>>>>
>>>>> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>> the same problem still occours. Will need to check when i've time to
>>>>>> gather logs again.
>>>>>>
>>>>>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> I'm not sure, but your logs did show that you had >16 recovery ops in
>>>>>>> flight, so it's worth a try.  If it doesn't help, you should collect
>>>>>>> the same set of logs I'll look again.  Also, there are a few other
>>>>>>> patches between 61.7 and current cuttlefish which may help.
>>>>>>> -Sam
>>>>>>>
>>>>>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>>
>>>>>>>>> I just backported a couple of patches from next to fix a bug where
>>>>>>>>> we
>>>>>>>>> weren't respecting the osd_recovery_max_active config in some cases
>>>>>>>>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>>>>>>>>> current cuttlefish branch or wait for a 61.8 release.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks! Are you sure that this is the issue? I don't believe that
>>>>>>>> but
>>>>>>>> i'll give it a try. I already tested a branch from sage where he
>>>>>>>> fixed
>>>>>>>> a
>>>>>>>> race regarding max active some weeks ago. So active recovering was
>>>>>>>> max
>>>>>>>> 1 but
>>>>>>>> the issue didn't went away.
>>>>>>>>
>>>>>>>> Stefan
>>>>>>>>
>>>>>>>>> -Sam
>>>>>>>>>
>>>>>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just
>>>>>>>>> <sam.just@inktank.com>
>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>>>>>> -Sam
>>>>>>>>>>
>>>>>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Did you take a look?
>>>>>>>>>>>
>>>>>>>>>>> Stefan
>>>>>>>>>>>
>>>>>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just
>>>>>>>>>>> <sam.just@inktank.com>:
>>>>>>>>>>>
>>>>>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>>>>>> -Sam
>>>>>>>>>>>>
>>>>>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> Hi Samual,
>>>>>>>>>>>>>
>>>>>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> debug osd = 20
>>>>>>>>>>>>>> debug filestore = 20
>>>>>>>>>>>>>> debug ms = 1
>>>>>>>>>>>>>> debug optracker = 20
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> on a few osds (including the restarted osd), and upload those
>>>>>>>>>>>>>> osd
>>>>>>>>>>>>>> logs
>>>>>>>>>>>>>> along with the ceph.log from before killing the osd until
>>>>>>>>>>>>>> after
>>>>>>>>>>>>>> the
>>>>>>>>>>>>>> cluster becomes clean again?
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>>>>>
>>>>>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>
>>>>>>>>>>>>> Greets,
>>>>>>>>>>>>> Stefan
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>>>> ceph-devel" in
>>>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>>>> More majordomo info at
>>>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>> ceph-devel"
>>>>>>>>> in
>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>>> in
>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>
>>>>>
>>>> --
>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>> the body of a message to majordomo@vger.kernel.org
>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>
>>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>

^ permalink raw reply	[flat|nested] 40+ messages in thread

* Re: still recovery issues with cuttlefish
  2013-08-22  3:34                                                   ` Samuel Just
@ 2013-08-22  7:41                                                     ` Stefan Priebe - Profihost AG
  0 siblings, 0 replies; 40+ messages in thread
From: Stefan Priebe - Profihost AG @ 2013-08-22  7:41 UTC (permalink / raw)
  To: Samuel Just; +Cc: Mike Dawson, josh.durgin, ceph-devel, Sage Weil

Am 22.08.2013 05:34, schrieb Samuel Just:
> It's not really possible at this time to control that limit because
> changing the primary is actually fairly expensive and doing it
> unnecessarily would probably make the situation much worse

I'm sorry but remapping or backfilling is far less expensive on all of
my machines than recovering.

While backfilling i've around 8-10% I/O waits while under recovery i
have 40%-50%


 (it's
> mostly necessary for backfilling, which is expensive anyway).  It
> seems like forwarding IO on an object which needs to be recovered to a
> replica with the object would be the next step.  Certainly something
> to consider for the future.

Yes this would be the solution.

Stefan

> -Sam
> 
> On Wed, Aug 21, 2013 at 12:37 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> Hi Sam,
>> Am 21.08.2013 21:13, schrieb Samuel Just:
>>
>>> As long as the request is for an object which is up to date on the
>>> primary, the request will be served without waiting for recovery.
>>
>>
>> Sure but remember if you have VM random 4K workload a lot of objects go out
>> of date pretty soon.
>>
>>
>>> A request only waits on recovery if the particular object being read or
>>>
>>> written must be recovered.
>>
>>
>> Yes but on 4k load this can be a lot.
>>
>>
>>> Your issue was that recovering the
>>> particular object being requested was unreasonably slow due to
>>> silliness in the recovery code which you disabled by disabling
>>> osd_recover_clone_overlap.
>>
>>
>> Yes and no. It's better now but far away from being good or perfect. My VMs
>> do not crash anymore but i still have a bunch of slow requests (just around
>> 10 messages) and still a VERY high I/O load on the disks during recovery.
>>
>>
>>> In cases where the primary osd is significantly behind, we do make one
>>> of the other osds primary during recovery in order to expedite
>>> requests (pgs in this state are shown as remapped).
>>
>>
>> oh never seen that but at least in my case even 60s are a very long
>> timeframe and the OSD is very stressed during recovery. Is it possible for
>> me to set this value?
>>
>>
>> Stefan
>>
>>> -Sam
>>>
>>> On Wed, Aug 21, 2013 at 11:21 AM, Stefan Priebe <s.priebe@profihost.ag>
>>> wrote:
>>>>
>>>> Am 21.08.2013 17:32, schrieb Samuel Just:
>>>>
>>>>> Have you tried setting osd_recovery_clone_overlap to false?  That
>>>>> seemed to help with Stefan's issue.
>>>>
>>>>
>>>>
>>>> This might sound a bug harsh but maybe due to my limited english skills
>>>> ;-)
>>>>
>>>> I still think that Cephs recovery system is broken by design. If an OSD
>>>> comes back (was offline) all write requests regarding PGs where this one
>>>> is
>>>> primary are targeted immediatly to this OSD. If this one is not up2date
>>>> for
>>>> an PG it tries to recover that one immediatly which costs 4MB / block. If
>>>> you have a lot of small write all over your OSDs and PGs you're sucked as
>>>> your OSD has to recover ALL it's PGs immediatly or at least lots of them
>>>> WHICH can't work. This is totally crazy.
>>>>
>>>> I think the right way would be:
>>>> 1.) if an OSD goes down the replicas got primaries
>>>>
>>>> or
>>>>
>>>> 2.) an OSD which does not have an up2date PG should redirect to the OSD
>>>> holding the secondary or third replica.
>>>>
>>>> Both results in being able to have a really smooth and slow recovery
>>>> without
>>>> any stress even under heavy 4K workloads like rbd backed VMs.
>>>>
>>>> Thanks for reading!
>>>>
>>>> Greets Stefan
>>>>
>>>>
>>>>
>>>>> -Sam
>>>>>
>>>>> On Wed, Aug 21, 2013 at 8:28 AM, Mike Dawson <mike.dawson@cloudapt.com>
>>>>> wrote:
>>>>>>
>>>>>>
>>>>>> Sam/Josh,
>>>>>>
>>>>>> We upgraded from 0.61.7 to 0.67.1 during a maintenance window this
>>>>>> morning,
>>>>>> hoping it would improve this situation, but there was no appreciable
>>>>>> change.
>>>>>>
>>>>>> One node in our cluster fsck'ed after a reboot and got a bit behind.
>>>>>> Our
>>>>>> instances backed by RBD volumes were OK at that point, but once the
>>>>>> node
>>>>>> booted fully and the OSDs started, all Windows instances with rbd
>>>>>> volumes
>>>>>> experienced very choppy performance and were unable to ingest video
>>>>>> surveillance traffic and commit it to disk. Once the cluster got back
>>>>>> to
>>>>>> HEALTH_OK, they resumed normal operation.
>>>>>>
>>>>>> I tried for a time with conservative recovery settings (osd max
>>>>>> backfills
>>>>>> =
>>>>>> 1, osd recovery op priority = 1, and osd recovery max active = 1). No
>>>>>> improvement for the guests. So I went to more aggressive settings to
>>>>>> get
>>>>>> things moving faster. That decreased the duration of the outage.
>>>>>>
>>>>>> During the entire period of recovery/backfill, the network looked
>>>>>> fine...no
>>>>>> where close to saturation. iowait on all drives look fine as well.
>>>>>>
>>>>>> Any ideas?
>>>>>>
>>>>>> Thanks,
>>>>>> Mike Dawson
>>>>>>
>>>>>>
>>>>>>
>>>>>> On 8/14/2013 3:04 AM, Stefan Priebe - Profihost AG wrote:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> the same problem still occours. Will need to check when i've time to
>>>>>>> gather logs again.
>>>>>>>
>>>>>>> Am 14.08.2013 01:11, schrieb Samuel Just:
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> I'm not sure, but your logs did show that you had >16 recovery ops in
>>>>>>>> flight, so it's worth a try.  If it doesn't help, you should collect
>>>>>>>> the same set of logs I'll look again.  Also, there are a few other
>>>>>>>> patches between 61.7 and current cuttlefish which may help.
>>>>>>>> -Sam
>>>>>>>>
>>>>>>>> On Tue, Aug 13, 2013 at 2:03 PM, Stefan Priebe - Profihost AG
>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Am 13.08.2013 um 22:43 schrieb Samuel Just <sam.just@inktank.com>:
>>>>>>>>>
>>>>>>>>>> I just backported a couple of patches from next to fix a bug where
>>>>>>>>>> we
>>>>>>>>>> weren't respecting the osd_recovery_max_active config in some cases
>>>>>>>>>> (1ea6b56170fc9e223e7c30635db02fa2ad8f4b4e).  You can either try the
>>>>>>>>>> current cuttlefish branch or wait for a 61.8 release.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Thanks! Are you sure that this is the issue? I don't believe that
>>>>>>>>> but
>>>>>>>>> i'll give it a try. I already tested a branch from sage where he
>>>>>>>>> fixed
>>>>>>>>> a
>>>>>>>>> race regarding max active some weeks ago. So active recovering was
>>>>>>>>> max
>>>>>>>>> 1 but
>>>>>>>>> the issue didn't went away.
>>>>>>>>>
>>>>>>>>> Stefan
>>>>>>>>>
>>>>>>>>>> -Sam
>>>>>>>>>>
>>>>>>>>>> On Mon, Aug 12, 2013 at 10:34 PM, Samuel Just
>>>>>>>>>> <sam.just@inktank.com>
>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> I got swamped today.  I should be able to look tomorrow.  Sorry!
>>>>>>>>>>> -Sam
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Aug 12, 2013 at 9:39 PM, Stefan Priebe - Profihost AG
>>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Did you take a look?
>>>>>>>>>>>>
>>>>>>>>>>>> Stefan
>>>>>>>>>>>>
>>>>>>>>>>>> Am 11.08.2013 um 05:50 schrieb Samuel Just
>>>>>>>>>>>> <sam.just@inktank.com>:
>>>>>>>>>>>>
>>>>>>>>>>>>> Great!  I'll take a look on Monday.
>>>>>>>>>>>>> -Sam
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Sat, Aug 10, 2013 at 12:08 PM, Stefan Priebe
>>>>>>>>>>>>> <s.priebe@profihost.ag> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hi Samual,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Am 09.08.2013 23:44, schrieb Samuel Just:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I think Stefan's problem is probably distinct from Mike's.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Stefan: Can you reproduce the problem with
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> debug osd = 20
>>>>>>>>>>>>>>> debug filestore = 20
>>>>>>>>>>>>>>> debug ms = 1
>>>>>>>>>>>>>>> debug optracker = 20
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> on a few osds (including the restarted osd), and upload those
>>>>>>>>>>>>>>> osd
>>>>>>>>>>>>>>> logs
>>>>>>>>>>>>>>> along with the ceph.log from before killing the osd until
>>>>>>>>>>>>>>> after
>>>>>>>>>>>>>>> the
>>>>>>>>>>>>>>> cluster becomes clean again?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> done - you'll find the logs at cephdrop folder:
>>>>>>>>>>>>>> slow_requests_recovering_cuttlefish
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> osd.52 was the one recovering
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks!
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Greets,
>>>>>>>>>>>>>> Stefan
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> --
>>>>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>>>>> ceph-devel" in
>>>>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>>>>> More majordomo info at
>>>>>>>>>>>>> http://vger.kernel.org/majordomo-info.html
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> To unsubscribe from this list: send the line "unsubscribe
>>>>>>>>>> ceph-devel"
>>>>>>>>>> in
>>>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
>>>>>>>> in
>>>>>>>> the body of a message to majordomo@vger.kernel.org
>>>>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>>>>
>>>>>>
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>>>> the body of a message to majordomo@vger.kernel.org
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>>>
>>>>
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 

^ permalink raw reply	[flat|nested] 40+ messages in thread

end of thread, other threads:[~2013-08-22  7:41 UTC | newest]

Thread overview: 40+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2013-08-01  8:22 still recovery issues with cuttlefish Stefan Priebe - Profihost AG
2013-08-01 14:50 ` Andrey Korolyov
2013-08-01 18:38   ` Samuel Just
2013-08-02 17:56     ` Andrey Korolyov
2013-08-01 18:34 ` Samuel Just
2013-08-01 18:34   ` Stefan Priebe
2013-08-01 18:36     ` Samuel Just
2013-08-01 18:36       ` Samuel Just
2013-08-01 18:46       ` Stefan Priebe
2013-08-01 18:54   ` Mike Dawson
2013-08-01 19:07     ` Stefan Priebe
2013-08-01 21:23       ` Samuel Just
2013-08-02  7:44         ` Stefan Priebe
2013-08-02 17:35           ` Samuel Just
2013-08-02 18:16             ` Stefan Priebe
2013-08-02 18:21               ` Samuel Just
2013-08-02 18:46                 ` Stefan Priebe
2013-08-08 14:05                   ` Mike Dawson
2013-08-08 15:43                     ` Oliver Francke
2013-08-08 18:13                     ` Stefan Priebe
2013-08-09 21:44                       ` Samuel Just
2013-08-10 19:08                         ` Stefan Priebe
2013-08-11  3:50                           ` Samuel Just
2013-08-13  4:39                             ` Stefan Priebe - Profihost AG
2013-08-13  5:34                               ` Samuel Just
2013-08-13 20:43                                 ` Samuel Just
2013-08-13 21:03                                   ` Stefan Priebe - Profihost AG
2013-08-13 23:11                                     ` Samuel Just
2013-08-14  7:04                                       ` Stefan Priebe - Profihost AG
2013-08-21 15:28                                         ` Mike Dawson
2013-08-21 15:32                                           ` Samuel Just
2013-08-21 16:25                                             ` Yann ROBIN
2013-08-21 17:12                                               ` Samuel Just
2013-08-21 17:55                                                 ` Mike Dawson
2013-08-21 18:05                                                   ` Samuel Just
2013-08-21 18:21                                             ` Stefan Priebe
2013-08-21 19:13                                               ` Samuel Just
2013-08-21 19:37                                                 ` Stefan Priebe
2013-08-22  3:34                                                   ` Samuel Just
2013-08-22  7:41                                                     ` Stefan Priebe - Profihost AG

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.