All of lore.kernel.org
 help / color / mirror / Atom feed
* ceph mon_data_size_warn limits for large cluster
@ 2019-02-06  9:28 M Ranga Swami Reddy
       [not found] ` <CANA9Uk437gWgTABHq66dw8v2KN6LqGF5DVc+ecaqoV24ABt++w-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  0 siblings, 1 reply; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-06  9:28 UTC (permalink / raw)
  To: ceph-users, ceph-devel

Hello -  Are the any limits for mon_data_size for cluster with 2PB
(with 2000+ OSDs)?

Currently it set as 15G. What is logic behind this? Can we increase
when we get the mon_data_size_warn messages?

I am getting the mon_data_size_warn message even though there a ample
of free space on the disk (around 300G free disk)

Earlier thread on the same discusion:
https://www.spinics.net/lists/ceph-users/msg42456.html

Thanks
Swami

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found] ` <CANA9Uk437gWgTABHq66dw8v2KN6LqGF5DVc+ecaqoV24ABt++w-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-06 12:26   ` Sage Weil
       [not found]     ` <alpine.DEB.2.11.1902061225330.4436-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
  0 siblings, 1 reply; 16+ messages in thread
From: Sage Weil @ 2019-02-06 12:26 UTC (permalink / raw)
  To: M Ranga Swami Reddy; +Cc: ceph-users, ceph-devel

Hi Swami

The limit is somewhat arbitrary, based on cluster sizes we had seen when 
we picked it.  In your case it should be perfectly safe to increase it.

sage


On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:

> Hello -  Are the any limits for mon_data_size for cluster with 2PB
> (with 2000+ OSDs)?
> 
> Currently it set as 15G. What is logic behind this? Can we increase
> when we get the mon_data_size_warn messages?
> 
> I am getting the mon_data_size_warn message even though there a ample
> of free space on the disk (around 300G free disk)
> 
> Earlier thread on the same discusion:
> https://www.spinics.net/lists/ceph-users/msg42456.html
> 
> Thanks
> Swami
> 
> 
> 

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]     ` <alpine.DEB.2.11.1902061225330.4436-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
@ 2019-02-06 12:54       ` Dan van der Ster
       [not found]         ` <CABZ+qqmBshV3DvSrP2UnvKLo1pSEDtVy_xfgP2YNd8e1ctW=xw-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  2019-02-07 11:18       ` M Ranga Swami Reddy
  1 sibling, 1 reply; 16+ messages in thread
From: Dan van der Ster @ 2019-02-06 12:54 UTC (permalink / raw)
  To: M Ranga Swami Reddy; +Cc: bokeoa-Re5JQEeQqe8AvxtiuMwx3w, ceph-users, ceph-devel

Hi,

With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.

During backfilling scenarios, the mons keep old maps and grow quite
quickly. So if you have balancing, pg splitting, etc. ongoing for
awhile, the mon stores will eventually trigger that 15GB alarm.
But the intended behavior is that once the PGs are all active+clean,
the old maps should be trimmed and the disk space freed.

However, several people have noted that (at least in luminous
releases) the old maps are not trimmed until after HEALTH_OK *and* all
mons are restarted. This ticket seems related:
http://tracker.ceph.com/issues/37875

(Over here we're restarting mons every ~2-3 weeks, resulting in the
mon stores dropping from >15GB to ~700MB each time).

-- Dan


On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
>
> Hi Swami
>
> The limit is somewhat arbitrary, based on cluster sizes we had seen when
> we picked it.  In your case it should be perfectly safe to increase it.
>
> sage
>
>
> On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
>
> > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > (with 2000+ OSDs)?
> >
> > Currently it set as 15G. What is logic behind this? Can we increase
> > when we get the mon_data_size_warn messages?
> >
> > I am getting the mon_data_size_warn message even though there a ample
> > of free space on the disk (around 300G free disk)
> >
> > Earlier thread on the same discusion:
> > https://www.spinics.net/lists/ceph-users/msg42456.html
> >
> > Thanks
> > Swami
> >
> >
> >
> _______________________________________________
> ceph-users mailing list
> ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]         ` <CABZ+qqmBshV3DvSrP2UnvKLo1pSEDtVy_xfgP2YNd8e1ctW=xw-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-07 11:17           ` M Ranga Swami Reddy
       [not found]             ` <CANA9Uk5byzNw8ewi3YHKtnxZnWprMYcmtvB9N3Qq3DaqPtMOwg-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  0 siblings, 1 reply; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-07 11:17 UTC (permalink / raw)
  To: Dan van der Ster; +Cc: bokeoa-Re5JQEeQqe8AvxtiuMwx3w, ceph-users, ceph-devel

Hi Dan,
>During backfilling scenarios, the mons keep old maps and grow quite
>quickly. So if you have balancing, pg splitting, etc. ongoing for
>awhile, the mon stores will eventually trigger that 15GB alarm.
>But the intended behavior is that once the PGs are all active+clean,
>the old maps should be trimmed and the disk space freed.

old maps not trimmed after cluster reached to "all+clean" state for all PGs.
Is there (known) bug here?
As the size of dB showing > 15G, do I need to run the compact commands
to do the trimming?

Thanks
Swami

On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
>
> Hi,
>
> With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
>
> During backfilling scenarios, the mons keep old maps and grow quite
> quickly. So if you have balancing, pg splitting, etc. ongoing for
> awhile, the mon stores will eventually trigger that 15GB alarm.
> But the intended behavior is that once the PGs are all active+clean,
> the old maps should be trimmed and the disk space freed.
>
> However, several people have noted that (at least in luminous
> releases) the old maps are not trimmed until after HEALTH_OK *and* all
> mons are restarted. This ticket seems related:
> http://tracker.ceph.com/issues/37875
>
> (Over here we're restarting mons every ~2-3 weeks, resulting in the
> mon stores dropping from >15GB to ~700MB each time).
>
> -- Dan
>
>
> On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> >
> > Hi Swami
> >
> > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > we picked it.  In your case it should be perfectly safe to increase it.
> >
> > sage
> >
> >
> > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> >
> > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > (with 2000+ OSDs)?
> > >
> > > Currently it set as 15G. What is logic behind this? Can we increase
> > > when we get the mon_data_size_warn messages?
> > >
> > > I am getting the mon_data_size_warn message even though there a ample
> > > of free space on the disk (around 300G free disk)
> > >
> > > Earlier thread on the same discusion:
> > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > >
> > > Thanks
> > > Swami
> > >
> > >
> > >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]     ` <alpine.DEB.2.11.1902061225330.4436-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
  2019-02-06 12:54       ` Dan van der Ster
@ 2019-02-07 11:18       ` M Ranga Swami Reddy
  1 sibling, 0 replies; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-07 11:18 UTC (permalink / raw)
  To: Sage Weil; +Cc: ceph-users, ceph-devel

Hi Sage

Sure, we will increase the mon_data_size to 30G to avoid this type of
warning. And currently we are using 500G disk here. I guss, which
should good enough here.

Thanks
Swami

On Wed, Feb 6, 2019 at 5:56 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
>
> Hi Swami
>
> The limit is somewhat arbitrary, based on cluster sizes we had seen when
> we picked it.  In your case it should be perfectly safe to increase it.
>
> sage
>
>
> On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
>
> > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > (with 2000+ OSDs)?
> >
> > Currently it set as 15G. What is logic behind this? Can we increase
> > when we get the mon_data_size_warn messages?
> >
> > I am getting the mon_data_size_warn message even though there a ample
> > of free space on the disk (around 300G free disk)
> >
> > Earlier thread on the same discusion:
> > https://www.spinics.net/lists/ceph-users/msg42456.html
> >
> > Thanks
> > Swami
> >
> >
> >

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]             ` <CANA9Uk5byzNw8ewi3YHKtnxZnWprMYcmtvB9N3Qq3DaqPtMOwg-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-07 12:59               ` Dan van der Ster
       [not found]                 ` <CABZ+qqn=NtW5oEb-qiqwTku2N4A04PvYBPLGspUnr29wCXqwjg-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  0 siblings, 1 reply; 16+ messages in thread
From: Dan van der Ster @ 2019-02-07 12:59 UTC (permalink / raw)
  To: M Ranga Swami Reddy; +Cc: Bryan Stillwell, ceph-users, ceph-devel

On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
<swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
>
> Hi Dan,
> >During backfilling scenarios, the mons keep old maps and grow quite
> >quickly. So if you have balancing, pg splitting, etc. ongoing for
> >awhile, the mon stores will eventually trigger that 15GB alarm.
> >But the intended behavior is that once the PGs are all active+clean,
> >the old maps should be trimmed and the disk space freed.
>
> old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> Is there (known) bug here?
> As the size of dB showing > 15G, do I need to run the compact commands
> to do the trimming?

Compaction isn't necessary -- you should only need to restart all
peon's then the leader. A few minutes later the db's should start
trimming.

-- dan


>
> Thanks
> Swami
>
> On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> >
> > Hi,
> >
> > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> >
> > During backfilling scenarios, the mons keep old maps and grow quite
> > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > awhile, the mon stores will eventually trigger that 15GB alarm.
> > But the intended behavior is that once the PGs are all active+clean,
> > the old maps should be trimmed and the disk space freed.
> >
> > However, several people have noted that (at least in luminous
> > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > mons are restarted. This ticket seems related:
> > http://tracker.ceph.com/issues/37875
> >
> > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > mon stores dropping from >15GB to ~700MB each time).
> >
> > -- Dan
> >
> >
> > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > >
> > > Hi Swami
> > >
> > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > we picked it.  In your case it should be perfectly safe to increase it.
> > >
> > > sage
> > >
> > >
> > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > >
> > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > (with 2000+ OSDs)?
> > > >
> > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > when we get the mon_data_size_warn messages?
> > > >
> > > > I am getting the mon_data_size_warn message even though there a ample
> > > > of free space on the disk (around 300G free disk)
> > > >
> > > > Earlier thread on the same discusion:
> > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > >
> > > > Thanks
> > > > Swami
> > > >
> > > >
> > > >
> > > _______________________________________________
> > > ceph-users mailing list
> > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                 ` <CABZ+qqn=NtW5oEb-qiqwTku2N4A04PvYBPLGspUnr29wCXqwjg-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-07 15:12                   ` M Ranga Swami Reddy
       [not found]                     ` <CANA9Uk4t05bvPV42=QgczSaewVU5-amxjc7NvFKxDcDWiRUAaw-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  2019-02-14 13:31                   ` Sage Weil
  1 sibling, 1 reply; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-07 15:12 UTC (permalink / raw)
  To: Dan van der Ster; +Cc: Bryan Stillwell, ceph-users, ceph-devel

>Compaction isn't necessary -- you should only need to restart all
>peon's then the leader. A few minutes later the db's should start
>trimming.

As we on production cluster, which may not be safe to restart the
ceph-mon, instead prefer to do the compact on non-leader mons.
Is this ok?

Thanks
Swami

On Thu, Feb 7, 2019 at 6:30 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
>
> On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> >
> > Hi Dan,
> > >During backfilling scenarios, the mons keep old maps and grow quite
> > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > >But the intended behavior is that once the PGs are all active+clean,
> > >the old maps should be trimmed and the disk space freed.
> >
> > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > Is there (known) bug here?
> > As the size of dB showing > 15G, do I need to run the compact commands
> > to do the trimming?
>
> Compaction isn't necessary -- you should only need to restart all
> peon's then the leader. A few minutes later the db's should start
> trimming.
>
> -- dan
>
>
> >
> > Thanks
> > Swami
> >
> > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > >
> > > Hi,
> > >
> > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > >
> > > During backfilling scenarios, the mons keep old maps and grow quite
> > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > But the intended behavior is that once the PGs are all active+clean,
> > > the old maps should be trimmed and the disk space freed.
> > >
> > > However, several people have noted that (at least in luminous
> > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > mons are restarted. This ticket seems related:
> > > http://tracker.ceph.com/issues/37875
> > >
> > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > mon stores dropping from >15GB to ~700MB each time).
> > >
> > > -- Dan
> > >
> > >
> > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > >
> > > > Hi Swami
> > > >
> > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > >
> > > > sage
> > > >
> > > >
> > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > >
> > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > (with 2000+ OSDs)?
> > > > >
> > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > when we get the mon_data_size_warn messages?
> > > > >
> > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > of free space on the disk (around 300G free disk)
> > > > >
> > > > > Earlier thread on the same discusion:
> > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > >
> > > > > Thanks
> > > > > Swami
> > > > >
> > > > >
> > > > >
> > > > _______________________________________________
> > > > ceph-users mailing list
> > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                     ` <CANA9Uk4t05bvPV42=QgczSaewVU5-amxjc7NvFKxDcDWiRUAaw-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-07 15:13                       ` Dan van der Ster
       [not found]                         ` <CABZ+qqn6vEk51S=0zxqxzLj0UgDe8P-bOvwzpM4=0TnAsP8uhQ-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  0 siblings, 1 reply; 16+ messages in thread
From: Dan van der Ster @ 2019-02-07 15:13 UTC (permalink / raw)
  To: M Ranga Swami Reddy; +Cc: Bryan Stillwell, ceph-users, ceph-devel

On Thu, Feb 7, 2019 at 4:12 PM M Ranga Swami Reddy <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
>
> >Compaction isn't necessary -- you should only need to restart all
> >peon's then the leader. A few minutes later the db's should start
> >trimming.
>
> As we on production cluster, which may not be safe to restart the
> ceph-mon, instead prefer to do the compact on non-leader mons.
> Is this ok?
>

Compaction doesn't solve this particular problem, because the maps
have not yet been deleted by the ceph-mon process.

-- dan


> Thanks
> Swami
>
> On Thu, Feb 7, 2019 at 6:30 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> >
> > On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > >
> > > Hi Dan,
> > > >During backfilling scenarios, the mons keep old maps and grow quite
> > > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > > >But the intended behavior is that once the PGs are all active+clean,
> > > >the old maps should be trimmed and the disk space freed.
> > >
> > > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > > Is there (known) bug here?
> > > As the size of dB showing > 15G, do I need to run the compact commands
> > > to do the trimming?
> >
> > Compaction isn't necessary -- you should only need to restart all
> > peon's then the leader. A few minutes later the db's should start
> > trimming.
> >
> > -- dan
> >
> >
> > >
> > > Thanks
> > > Swami
> > >
> > > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > > >
> > > > Hi,
> > > >
> > > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > > >
> > > > During backfilling scenarios, the mons keep old maps and grow quite
> > > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > But the intended behavior is that once the PGs are all active+clean,
> > > > the old maps should be trimmed and the disk space freed.
> > > >
> > > > However, several people have noted that (at least in luminous
> > > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > > mons are restarted. This ticket seems related:
> > > > http://tracker.ceph.com/issues/37875
> > > >
> > > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > > mon stores dropping from >15GB to ~700MB each time).
> > > >
> > > > -- Dan
> > > >
> > > >
> > > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > > >
> > > > > Hi Swami
> > > > >
> > > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > > >
> > > > > sage
> > > > >
> > > > >
> > > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > > >
> > > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > > (with 2000+ OSDs)?
> > > > > >
> > > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > > when we get the mon_data_size_warn messages?
> > > > > >
> > > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > > of free space on the disk (around 300G free disk)
> > > > > >
> > > > > > Earlier thread on the same discusion:
> > > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > > >
> > > > > > Thanks
> > > > > > Swami
> > > > > >
> > > > > >
> > > > > >
> > > > > _______________________________________________
> > > > > ceph-users mailing list
> > > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                         ` <CABZ+qqn6vEk51S=0zxqxzLj0UgDe8P-bOvwzpM4=0TnAsP8uhQ-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-07 15:37                           ` M Ranga Swami Reddy
  0 siblings, 0 replies; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-07 15:37 UTC (permalink / raw)
  To: Dan van der Ster; +Cc: Bryan Stillwell, ceph-users, ceph-devel

Alternatively, will increase the mon_data_size to 30G (from 15G)..

Thanks
Swami

On Thu, Feb 7, 2019 at 8:44 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
>
> On Thu, Feb 7, 2019 at 4:12 PM M Ranga Swami Reddy <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> >
> > >Compaction isn't necessary -- you should only need to restart all
> > >peon's then the leader. A few minutes later the db's should start
> > >trimming.
> >
> > As we on production cluster, which may not be safe to restart the
> > ceph-mon, instead prefer to do the compact on non-leader mons.
> > Is this ok?
> >
>
> Compaction doesn't solve this particular problem, because the maps
> have not yet been deleted by the ceph-mon process.
>
> -- dan
>
>
> > Thanks
> > Swami
> >
> > On Thu, Feb 7, 2019 at 6:30 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > >
> > > On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> > > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > > >
> > > > Hi Dan,
> > > > >During backfilling scenarios, the mons keep old maps and grow quite
> > > > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > >But the intended behavior is that once the PGs are all active+clean,
> > > > >the old maps should be trimmed and the disk space freed.
> > > >
> > > > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > > > Is there (known) bug here?
> > > > As the size of dB showing > 15G, do I need to run the compact commands
> > > > to do the trimming?
> > >
> > > Compaction isn't necessary -- you should only need to restart all
> > > peon's then the leader. A few minutes later the db's should start
> > > trimming.
> > >
> > > -- dan
> > >
> > >
> > > >
> > > > Thanks
> > > > Swami
> > > >
> > > > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > > > >
> > > > > Hi,
> > > > >
> > > > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > > > >
> > > > > During backfilling scenarios, the mons keep old maps and grow quite
> > > > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > > But the intended behavior is that once the PGs are all active+clean,
> > > > > the old maps should be trimmed and the disk space freed.
> > > > >
> > > > > However, several people have noted that (at least in luminous
> > > > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > > > mons are restarted. This ticket seems related:
> > > > > http://tracker.ceph.com/issues/37875
> > > > >
> > > > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > > > mon stores dropping from >15GB to ~700MB each time).
> > > > >
> > > > > -- Dan
> > > > >
> > > > >
> > > > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > > > >
> > > > > > Hi Swami
> > > > > >
> > > > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > > > >
> > > > > > sage
> > > > > >
> > > > > >
> > > > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > > > >
> > > > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > > > (with 2000+ OSDs)?
> > > > > > >
> > > > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > > > when we get the mon_data_size_warn messages?
> > > > > > >
> > > > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > > > of free space on the disk (around 300G free disk)
> > > > > > >
> > > > > > > Earlier thread on the same discusion:
> > > > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > > > >
> > > > > > > Thanks
> > > > > > > Swami
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > _______________________________________________
> > > > > > ceph-users mailing list
> > > > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                 ` <CABZ+qqn=NtW5oEb-qiqwTku2N4A04PvYBPLGspUnr29wCXqwjg-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  2019-02-07 15:12                   ` M Ranga Swami Reddy
@ 2019-02-14 13:31                   ` Sage Weil
       [not found]                     ` <alpine.DEB.2.11.1902141327510.14565-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
  1 sibling, 1 reply; 16+ messages in thread
From: Sage Weil @ 2019-02-14 13:31 UTC (permalink / raw)
  To: Dan van der Ster; +Cc: Bryan Stillwell, ceph-users, ceph-devel

On Thu, 7 Feb 2019, Dan van der Ster wrote:
> On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> >
> > Hi Dan,
> > >During backfilling scenarios, the mons keep old maps and grow quite
> > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > >But the intended behavior is that once the PGs are all active+clean,
> > >the old maps should be trimmed and the disk space freed.
> >
> > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > Is there (known) bug here?
> > As the size of dB showing > 15G, do I need to run the compact commands
> > to do the trimming?
> 
> Compaction isn't necessary -- you should only need to restart all
> peon's then the leader. A few minutes later the db's should start
> trimming.

The next time someone sees this behavior, can you please

- enable debug_mon = 20 on all mons (*before* restarting)
   ceph tell mon.* injectargs '--debug-mon 20'
- wait for 10 minutes or so to generate some logs
- add 'debug mon = 20' to ceph.conf (on mons only)
- restart the monitors
- wait for them to start trimming
- remove 'debug mon = 20' from ceph.conf (on mons only)
- tar up the log files, ceph-post-file them, and share them with ticket 
http://tracker.ceph.com/issues/38322

Thanks!
sage



 
> -- dan
> 
> 
> >
> > Thanks
> > Swami
> >
> > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > >
> > > Hi,
> > >
> > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > >
> > > During backfilling scenarios, the mons keep old maps and grow quite
> > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > But the intended behavior is that once the PGs are all active+clean,
> > > the old maps should be trimmed and the disk space freed.
> > >
> > > However, several people have noted that (at least in luminous
> > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > mons are restarted. This ticket seems related:
> > > http://tracker.ceph.com/issues/37875
> > >
> > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > mon stores dropping from >15GB to ~700MB each time).
> > >
> > > -- Dan
> > >
> > >
> > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > >
> > > > Hi Swami
> > > >
> > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > >
> > > > sage
> > > >
> > > >
> > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > >
> > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > (with 2000+ OSDs)?
> > > > >
> > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > when we get the mon_data_size_warn messages?
> > > > >
> > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > of free space on the disk (around 300G free disk)
> > > > >
> > > > > Earlier thread on the same discusion:
> > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > >
> > > > > Thanks
> > > > > Swami
> > > > >
> > > > >
> > > > >
> > > > _______________________________________________
> > > > ceph-users mailing list
> > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
> 

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                     ` <alpine.DEB.2.11.1902141327510.14565-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
@ 2019-02-14 14:11                       ` M Ranga Swami Reddy
  2019-02-15  9:43                       ` M Ranga Swami Reddy
  2019-02-18 11:27                       ` Dan van der Ster
  2 siblings, 0 replies; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-14 14:11 UTC (permalink / raw)
  To: Sage Weil; +Cc: Bryan Stillwell, ceph-users, ceph-devel

Sure, will this. For now I have creased the size to 30G (from 15G).

On Thu, Feb 14, 2019 at 7:39 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
>
> On Thu, 7 Feb 2019, Dan van der Ster wrote:
> > On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > >
> > > Hi Dan,
> > > >During backfilling scenarios, the mons keep old maps and grow quite
> > > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > > >But the intended behavior is that once the PGs are all active+clean,
> > > >the old maps should be trimmed and the disk space freed.
> > >
> > > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > > Is there (known) bug here?
> > > As the size of dB showing > 15G, do I need to run the compact commands
> > > to do the trimming?
> >
> > Compaction isn't necessary -- you should only need to restart all
> > peon's then the leader. A few minutes later the db's should start
> > trimming.
>
> The next time someone sees this behavior, can you please
>
> - enable debug_mon = 20 on all mons (*before* restarting)
>    ceph tell mon.* injectargs '--debug-mon 20'
> - wait for 10 minutes or so to generate some logs
> - add 'debug mon = 20' to ceph.conf (on mons only)
> - restart the monitors
> - wait for them to start trimming
> - remove 'debug mon = 20' from ceph.conf (on mons only)
> - tar up the log files, ceph-post-file them, and share them with ticket
> http://tracker.ceph.com/issues/38322
>
> Thanks!
> sage
>
>
>
>
> > -- dan
> >
> >
> > >
> > > Thanks
> > > Swami
> > >
> > > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > > >
> > > > Hi,
> > > >
> > > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > > >
> > > > During backfilling scenarios, the mons keep old maps and grow quite
> > > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > But the intended behavior is that once the PGs are all active+clean,
> > > > the old maps should be trimmed and the disk space freed.
> > > >
> > > > However, several people have noted that (at least in luminous
> > > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > > mons are restarted. This ticket seems related:
> > > > http://tracker.ceph.com/issues/37875
> > > >
> > > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > > mon stores dropping from >15GB to ~700MB each time).
> > > >
> > > > -- Dan
> > > >
> > > >
> > > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > > >
> > > > > Hi Swami
> > > > >
> > > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > > >
> > > > > sage
> > > > >
> > > > >
> > > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > > >
> > > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > > (with 2000+ OSDs)?
> > > > > >
> > > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > > when we get the mon_data_size_warn messages?
> > > > > >
> > > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > > of free space on the disk (around 300G free disk)
> > > > > >
> > > > > > Earlier thread on the same discusion:
> > > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > > >
> > > > > > Thanks
> > > > > > Swami
> > > > > >
> > > > > >
> > > > > >
> > > > > _______________________________________________
> > > > > ceph-users mailing list
> > > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
> >

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                     ` <alpine.DEB.2.11.1902141327510.14565-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
  2019-02-14 14:11                       ` M Ranga Swami Reddy
@ 2019-02-15  9:43                       ` M Ranga Swami Reddy
       [not found]                         ` <CANA9Uk4BHvaZk9dQ8v7+a780U+c5usYqcQJB_M6fuNR0nRnzTA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  2019-02-18 11:27                       ` Dan van der Ster
  2 siblings, 1 reply; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-15  9:43 UTC (permalink / raw)
  To: Sage Weil; +Cc: Bryan Stillwell, ceph-users, ceph-devel

today I again hit the warn with 30G also...

On Thu, Feb 14, 2019 at 7:39 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
>
> On Thu, 7 Feb 2019, Dan van der Ster wrote:
> > On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > >
> > > Hi Dan,
> > > >During backfilling scenarios, the mons keep old maps and grow quite
> > > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > > >But the intended behavior is that once the PGs are all active+clean,
> > > >the old maps should be trimmed and the disk space freed.
> > >
> > > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > > Is there (known) bug here?
> > > As the size of dB showing > 15G, do I need to run the compact commands
> > > to do the trimming?
> >
> > Compaction isn't necessary -- you should only need to restart all
> > peon's then the leader. A few minutes later the db's should start
> > trimming.
>
> The next time someone sees this behavior, can you please
>
> - enable debug_mon = 20 on all mons (*before* restarting)
>    ceph tell mon.* injectargs '--debug-mon 20'
> - wait for 10 minutes or so to generate some logs
> - add 'debug mon = 20' to ceph.conf (on mons only)
> - restart the monitors
> - wait for them to start trimming
> - remove 'debug mon = 20' from ceph.conf (on mons only)
> - tar up the log files, ceph-post-file them, and share them with ticket
> http://tracker.ceph.com/issues/38322
>
> Thanks!
> sage
>
>
>
>
> > -- dan
> >
> >
> > >
> > > Thanks
> > > Swami
> > >
> > > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > > >
> > > > Hi,
> > > >
> > > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > > >
> > > > During backfilling scenarios, the mons keep old maps and grow quite
> > > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > But the intended behavior is that once the PGs are all active+clean,
> > > > the old maps should be trimmed and the disk space freed.
> > > >
> > > > However, several people have noted that (at least in luminous
> > > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > > mons are restarted. This ticket seems related:
> > > > http://tracker.ceph.com/issues/37875
> > > >
> > > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > > mon stores dropping from >15GB to ~700MB each time).
> > > >
> > > > -- Dan
> > > >
> > > >
> > > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > > >
> > > > > Hi Swami
> > > > >
> > > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > > >
> > > > > sage
> > > > >
> > > > >
> > > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > > >
> > > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > > (with 2000+ OSDs)?
> > > > > >
> > > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > > when we get the mon_data_size_warn messages?
> > > > > >
> > > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > > of free space on the disk (around 300G free disk)
> > > > > >
> > > > > > Earlier thread on the same discusion:
> > > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > > >
> > > > > > Thanks
> > > > > > Swami
> > > > > >
> > > > > >
> > > > > >
> > > > > _______________________________________________
> > > > > ceph-users mailing list
> > > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
> >

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                         ` <CANA9Uk4BHvaZk9dQ8v7+a780U+c5usYqcQJB_M6fuNR0nRnzTA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-18 11:11                           ` M Ranga Swami Reddy
       [not found]                             ` <CANA9Uk4yVo65S5gsfzMOxzE0ya80Brqkmhxb3tFVW49a1c_MFA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  0 siblings, 1 reply; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-18 11:11 UTC (permalink / raw)
  To: Sage Weil; +Cc: Bryan Stillwell, ceph-users, ceph-devel

Hi Sage - If the mon data increases, is this impacts the ceph cluster
performance (ie on ceph osd bench, etc)?

On Fri, Feb 15, 2019 at 3:13 PM M Ranga Swami Reddy
<swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
>
> today I again hit the warn with 30G also...
>
> On Thu, Feb 14, 2019 at 7:39 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> >
> > On Thu, 7 Feb 2019, Dan van der Ster wrote:
> > > On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> > > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > > >
> > > > Hi Dan,
> > > > >During backfilling scenarios, the mons keep old maps and grow quite
> > > > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > >But the intended behavior is that once the PGs are all active+clean,
> > > > >the old maps should be trimmed and the disk space freed.
> > > >
> > > > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > > > Is there (known) bug here?
> > > > As the size of dB showing > 15G, do I need to run the compact commands
> > > > to do the trimming?
> > >
> > > Compaction isn't necessary -- you should only need to restart all
> > > peon's then the leader. A few minutes later the db's should start
> > > trimming.
> >
> > The next time someone sees this behavior, can you please
> >
> > - enable debug_mon = 20 on all mons (*before* restarting)
> >    ceph tell mon.* injectargs '--debug-mon 20'
> > - wait for 10 minutes or so to generate some logs
> > - add 'debug mon = 20' to ceph.conf (on mons only)
> > - restart the monitors
> > - wait for them to start trimming
> > - remove 'debug mon = 20' from ceph.conf (on mons only)
> > - tar up the log files, ceph-post-file them, and share them with ticket
> > http://tracker.ceph.com/issues/38322
> >
> > Thanks!
> > sage
> >
> >
> >
> >
> > > -- dan
> > >
> > >
> > > >
> > > > Thanks
> > > > Swami
> > > >
> > > > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > > > >
> > > > > Hi,
> > > > >
> > > > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > > > >
> > > > > During backfilling scenarios, the mons keep old maps and grow quite
> > > > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > > But the intended behavior is that once the PGs are all active+clean,
> > > > > the old maps should be trimmed and the disk space freed.
> > > > >
> > > > > However, several people have noted that (at least in luminous
> > > > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > > > mons are restarted. This ticket seems related:
> > > > > http://tracker.ceph.com/issues/37875
> > > > >
> > > > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > > > mon stores dropping from >15GB to ~700MB each time).
> > > > >
> > > > > -- Dan
> > > > >
> > > > >
> > > > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > > > >
> > > > > > Hi Swami
> > > > > >
> > > > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > > > >
> > > > > > sage
> > > > > >
> > > > > >
> > > > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > > > >
> > > > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > > > (with 2000+ OSDs)?
> > > > > > >
> > > > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > > > when we get the mon_data_size_warn messages?
> > > > > > >
> > > > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > > > of free space on the disk (around 300G free disk)
> > > > > > >
> > > > > > > Earlier thread on the same discusion:
> > > > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > > > >
> > > > > > > Thanks
> > > > > > > Swami
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > _______________________________________________
> > > > > > ceph-users mailing list
> > > > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > >
> > >
> > >

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                     ` <alpine.DEB.2.11.1902141327510.14565-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
  2019-02-14 14:11                       ` M Ranga Swami Reddy
  2019-02-15  9:43                       ` M Ranga Swami Reddy
@ 2019-02-18 11:27                       ` Dan van der Ster
  2 siblings, 0 replies; 16+ messages in thread
From: Dan van der Ster @ 2019-02-18 11:27 UTC (permalink / raw)
  To: Sage Weil; +Cc: Bryan Stillwell, ceph-users, ceph-devel

On Thu, Feb 14, 2019 at 2:31 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
>
> On Thu, 7 Feb 2019, Dan van der Ster wrote:
> > On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > >
> > > Hi Dan,
> > > >During backfilling scenarios, the mons keep old maps and grow quite
> > > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > > >But the intended behavior is that once the PGs are all active+clean,
> > > >the old maps should be trimmed and the disk space freed.
> > >
> > > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > > Is there (known) bug here?
> > > As the size of dB showing > 15G, do I need to run the compact commands
> > > to do the trimming?
> >
> > Compaction isn't necessary -- you should only need to restart all
> > peon's then the leader. A few minutes later the db's should start
> > trimming.
>
> The next time someone sees this behavior, can you please
>
> - enable debug_mon = 20 on all mons (*before* restarting)
>    ceph tell mon.* injectargs '--debug-mon 20'
> - wait for 10 minutes or so to generate some logs
> - add 'debug mon = 20' to ceph.conf (on mons only)
> - restart the monitors
> - wait for them to start trimming
> - remove 'debug mon = 20' from ceph.conf (on mons only)
> - tar up the log files, ceph-post-file them, and share them with ticket
> http://tracker.ceph.com/issues/38322
>

Not sure if you noticed, but we sent some logs Friday.

-- dan

> Thanks!
> sage
>
>
>
>
> > -- dan
> >
> >
> > >
> > > Thanks
> > > Swami
> > >
> > > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > > >
> > > > Hi,
> > > >
> > > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > > >
> > > > During backfilling scenarios, the mons keep old maps and grow quite
> > > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > But the intended behavior is that once the PGs are all active+clean,
> > > > the old maps should be trimmed and the disk space freed.
> > > >
> > > > However, several people have noted that (at least in luminous
> > > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > > mons are restarted. This ticket seems related:
> > > > http://tracker.ceph.com/issues/37875
> > > >
> > > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > > mon stores dropping from >15GB to ~700MB each time).
> > > >
> > > > -- Dan
> > > >
> > > >
> > > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > > >
> > > > > Hi Swami
> > > > >
> > > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > > >
> > > > > sage
> > > > >
> > > > >
> > > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > > >
> > > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > > (with 2000+ OSDs)?
> > > > > >
> > > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > > when we get the mon_data_size_warn messages?
> > > > > >
> > > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > > of free space on the disk (around 300G free disk)
> > > > > >
> > > > > > Earlier thread on the same discusion:
> > > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > > >
> > > > > > Thanks
> > > > > > Swami
> > > > > >
> > > > > >
> > > > > >
> > > > > _______________________________________________
> > > > > ceph-users mailing list
> > > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
> >

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                             ` <CANA9Uk4yVo65S5gsfzMOxzE0ya80Brqkmhxb3tFVW49a1c_MFA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-18 11:28                               ` Dan van der Ster
       [not found]                                 ` <CABZ+qqktu6RXFSA9-9tHcUGYy2CkuGPJU9s=uvrParrzujoiEA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
  0 siblings, 1 reply; 16+ messages in thread
From: Dan van der Ster @ 2019-02-18 11:28 UTC (permalink / raw)
  To: M Ranga Swami Reddy; +Cc: Bryan Stillwell, ceph-users, ceph-devel

Not really.

You should just restart your mons though -- if done one at a time it
has zero impact on your clients.

-- dan


On Mon, Feb 18, 2019 at 12:11 PM M Ranga Swami Reddy
<swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
>
> Hi Sage - If the mon data increases, is this impacts the ceph cluster
> performance (ie on ceph osd bench, etc)?
>
> On Fri, Feb 15, 2019 at 3:13 PM M Ranga Swami Reddy
> <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> >
> > today I again hit the warn with 30G also...
> >
> > On Thu, Feb 14, 2019 at 7:39 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > >
> > > On Thu, 7 Feb 2019, Dan van der Ster wrote:
> > > > On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> > > > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > > > >
> > > > > Hi Dan,
> > > > > >During backfilling scenarios, the mons keep old maps and grow quite
> > > > > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > > >But the intended behavior is that once the PGs are all active+clean,
> > > > > >the old maps should be trimmed and the disk space freed.
> > > > >
> > > > > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > > > > Is there (known) bug here?
> > > > > As the size of dB showing > 15G, do I need to run the compact commands
> > > > > to do the trimming?
> > > >
> > > > Compaction isn't necessary -- you should only need to restart all
> > > > peon's then the leader. A few minutes later the db's should start
> > > > trimming.
> > >
> > > The next time someone sees this behavior, can you please
> > >
> > > - enable debug_mon = 20 on all mons (*before* restarting)
> > >    ceph tell mon.* injectargs '--debug-mon 20'
> > > - wait for 10 minutes or so to generate some logs
> > > - add 'debug mon = 20' to ceph.conf (on mons only)
> > > - restart the monitors
> > > - wait for them to start trimming
> > > - remove 'debug mon = 20' from ceph.conf (on mons only)
> > > - tar up the log files, ceph-post-file them, and share them with ticket
> > > http://tracker.ceph.com/issues/38322
> > >
> > > Thanks!
> > > sage
> > >
> > >
> > >
> > >
> > > > -- dan
> > > >
> > > >
> > > > >
> > > > > Thanks
> > > > > Swami
> > > > >
> > > > > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > > > > >
> > > > > > Hi,
> > > > > >
> > > > > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > > > > >
> > > > > > During backfilling scenarios, the mons keep old maps and grow quite
> > > > > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > > > But the intended behavior is that once the PGs are all active+clean,
> > > > > > the old maps should be trimmed and the disk space freed.
> > > > > >
> > > > > > However, several people have noted that (at least in luminous
> > > > > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > > > > mons are restarted. This ticket seems related:
> > > > > > http://tracker.ceph.com/issues/37875
> > > > > >
> > > > > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > > > > mon stores dropping from >15GB to ~700MB each time).
> > > > > >
> > > > > > -- Dan
> > > > > >
> > > > > >
> > > > > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > > > > >
> > > > > > > Hi Swami
> > > > > > >
> > > > > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > > > > >
> > > > > > > sage
> > > > > > >
> > > > > > >
> > > > > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > > > > >
> > > > > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > > > > (with 2000+ OSDs)?
> > > > > > > >
> > > > > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > > > > when we get the mon_data_size_warn messages?
> > > > > > > >
> > > > > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > > > > of free space on the disk (around 300G free disk)
> > > > > > > >
> > > > > > > > Earlier thread on the same discusion:
> > > > > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > > > > >
> > > > > > > > Thanks
> > > > > > > > Swami
> > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > _______________________________________________
> > > > > > > ceph-users mailing list
> > > > > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > >
> > > >
> > > >

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: ceph mon_data_size_warn limits for large cluster
       [not found]                                 ` <CABZ+qqktu6RXFSA9-9tHcUGYy2CkuGPJU9s=uvrParrzujoiEA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
@ 2019-02-18 11:30                                   ` M Ranga Swami Reddy
  0 siblings, 0 replies; 16+ messages in thread
From: M Ranga Swami Reddy @ 2019-02-18 11:30 UTC (permalink / raw)
  To: Dan van der Ster; +Cc: Bryan Stillwell, ceph-users, ceph-devel

OK, sure will restart the ceph-mon (starting from non leader first,
and then last leader node).


On Mon, Feb 18, 2019 at 4:59 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
>
> Not really.
>
> You should just restart your mons though -- if done one at a time it
> has zero impact on your clients.
>
> -- dan
>
>
> On Mon, Feb 18, 2019 at 12:11 PM M Ranga Swami Reddy
> <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> >
> > Hi Sage - If the mon data increases, is this impacts the ceph cluster
> > performance (ie on ceph osd bench, etc)?
> >
> > On Fri, Feb 15, 2019 at 3:13 PM M Ranga Swami Reddy
> > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > >
> > > today I again hit the warn with 30G also...
> > >
> > > On Thu, Feb 14, 2019 at 7:39 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > >
> > > > On Thu, 7 Feb 2019, Dan van der Ster wrote:
> > > > > On Thu, Feb 7, 2019 at 12:17 PM M Ranga Swami Reddy
> > > > > <swamireddy-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> wrote:
> > > > > >
> > > > > > Hi Dan,
> > > > > > >During backfilling scenarios, the mons keep old maps and grow quite
> > > > > > >quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > > > >awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > > > >But the intended behavior is that once the PGs are all active+clean,
> > > > > > >the old maps should be trimmed and the disk space freed.
> > > > > >
> > > > > > old maps not trimmed after cluster reached to "all+clean" state for all PGs.
> > > > > > Is there (known) bug here?
> > > > > > As the size of dB showing > 15G, do I need to run the compact commands
> > > > > > to do the trimming?
> > > > >
> > > > > Compaction isn't necessary -- you should only need to restart all
> > > > > peon's then the leader. A few minutes later the db's should start
> > > > > trimming.
> > > >
> > > > The next time someone sees this behavior, can you please
> > > >
> > > > - enable debug_mon = 20 on all mons (*before* restarting)
> > > >    ceph tell mon.* injectargs '--debug-mon 20'
> > > > - wait for 10 minutes or so to generate some logs
> > > > - add 'debug mon = 20' to ceph.conf (on mons only)
> > > > - restart the monitors
> > > > - wait for them to start trimming
> > > > - remove 'debug mon = 20' from ceph.conf (on mons only)
> > > > - tar up the log files, ceph-post-file them, and share them with ticket
> > > > http://tracker.ceph.com/issues/38322
> > > >
> > > > Thanks!
> > > > sage
> > > >
> > > >
> > > >
> > > >
> > > > > -- dan
> > > > >
> > > > >
> > > > > >
> > > > > > Thanks
> > > > > > Swami
> > > > > >
> > > > > > On Wed, Feb 6, 2019 at 6:24 PM Dan van der Ster <dan-EOCVfBHj35C+XT7JhA+gdA@public.gmane.org> wrote:
> > > > > > >
> > > > > > > Hi,
> > > > > > >
> > > > > > > With HEALTH_OK a mon data dir should be under 2GB for even such a large cluster.
> > > > > > >
> > > > > > > During backfilling scenarios, the mons keep old maps and grow quite
> > > > > > > quickly. So if you have balancing, pg splitting, etc. ongoing for
> > > > > > > awhile, the mon stores will eventually trigger that 15GB alarm.
> > > > > > > But the intended behavior is that once the PGs are all active+clean,
> > > > > > > the old maps should be trimmed and the disk space freed.
> > > > > > >
> > > > > > > However, several people have noted that (at least in luminous
> > > > > > > releases) the old maps are not trimmed until after HEALTH_OK *and* all
> > > > > > > mons are restarted. This ticket seems related:
> > > > > > > http://tracker.ceph.com/issues/37875
> > > > > > >
> > > > > > > (Over here we're restarting mons every ~2-3 weeks, resulting in the
> > > > > > > mon stores dropping from >15GB to ~700MB each time).
> > > > > > >
> > > > > > > -- Dan
> > > > > > >
> > > > > > >
> > > > > > > On Wed, Feb 6, 2019 at 1:26 PM Sage Weil <sage-BnTBU8nroG7k1uMJSBkQmQ@public.gmane.org> wrote:
> > > > > > > >
> > > > > > > > Hi Swami
> > > > > > > >
> > > > > > > > The limit is somewhat arbitrary, based on cluster sizes we had seen when
> > > > > > > > we picked it.  In your case it should be perfectly safe to increase it.
> > > > > > > >
> > > > > > > > sage
> > > > > > > >
> > > > > > > >
> > > > > > > > On Wed, 6 Feb 2019, M Ranga Swami Reddy wrote:
> > > > > > > >
> > > > > > > > > Hello -  Are the any limits for mon_data_size for cluster with 2PB
> > > > > > > > > (with 2000+ OSDs)?
> > > > > > > > >
> > > > > > > > > Currently it set as 15G. What is logic behind this? Can we increase
> > > > > > > > > when we get the mon_data_size_warn messages?
> > > > > > > > >
> > > > > > > > > I am getting the mon_data_size_warn message even though there a ample
> > > > > > > > > of free space on the disk (around 300G free disk)
> > > > > > > > >
> > > > > > > > > Earlier thread on the same discusion:
> > > > > > > > > https://www.spinics.net/lists/ceph-users/msg42456.html
> > > > > > > > >
> > > > > > > > > Thanks
> > > > > > > > > Swami
> > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > _______________________________________________
> > > > > > > > ceph-users mailing list
> > > > > > > > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org
> > > > > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > > >
> > > > >
> > > > >

^ permalink raw reply	[flat|nested] 16+ messages in thread

end of thread, other threads:[~2019-02-18 11:30 UTC | newest]

Thread overview: 16+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-02-06  9:28 ceph mon_data_size_warn limits for large cluster M Ranga Swami Reddy
     [not found] ` <CANA9Uk437gWgTABHq66dw8v2KN6LqGF5DVc+ecaqoV24ABt++w-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-06 12:26   ` Sage Weil
     [not found]     ` <alpine.DEB.2.11.1902061225330.4436-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
2019-02-06 12:54       ` Dan van der Ster
     [not found]         ` <CABZ+qqmBshV3DvSrP2UnvKLo1pSEDtVy_xfgP2YNd8e1ctW=xw-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-07 11:17           ` M Ranga Swami Reddy
     [not found]             ` <CANA9Uk5byzNw8ewi3YHKtnxZnWprMYcmtvB9N3Qq3DaqPtMOwg-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-07 12:59               ` Dan van der Ster
     [not found]                 ` <CABZ+qqn=NtW5oEb-qiqwTku2N4A04PvYBPLGspUnr29wCXqwjg-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-07 15:12                   ` M Ranga Swami Reddy
     [not found]                     ` <CANA9Uk4t05bvPV42=QgczSaewVU5-amxjc7NvFKxDcDWiRUAaw-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-07 15:13                       ` Dan van der Ster
     [not found]                         ` <CABZ+qqn6vEk51S=0zxqxzLj0UgDe8P-bOvwzpM4=0TnAsP8uhQ-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-07 15:37                           ` M Ranga Swami Reddy
2019-02-14 13:31                   ` Sage Weil
     [not found]                     ` <alpine.DEB.2.11.1902141327510.14565-qHenpvqtifaMSRpgCs4c+g@public.gmane.org>
2019-02-14 14:11                       ` M Ranga Swami Reddy
2019-02-15  9:43                       ` M Ranga Swami Reddy
     [not found]                         ` <CANA9Uk4BHvaZk9dQ8v7+a780U+c5usYqcQJB_M6fuNR0nRnzTA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-18 11:11                           ` M Ranga Swami Reddy
     [not found]                             ` <CANA9Uk4yVo65S5gsfzMOxzE0ya80Brqkmhxb3tFVW49a1c_MFA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-18 11:28                               ` Dan van der Ster
     [not found]                                 ` <CABZ+qqktu6RXFSA9-9tHcUGYy2CkuGPJU9s=uvrParrzujoiEA-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org>
2019-02-18 11:30                                   ` M Ranga Swami Reddy
2019-02-18 11:27                       ` Dan van der Ster
2019-02-07 11:18       ` M Ranga Swami Reddy

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.