All of lore.kernel.org
 help / color / mirror / Atom feed
* mon memory issue
@ 2012-08-31  9:02 Xiaopong Tran
  2012-08-31 16:03 ` e: " Sage Weil
  0 siblings, 1 reply; 9+ messages in thread
From: Xiaopong Tran @ 2012-08-31  9:02 UTC (permalink / raw)
  To: ceph-devel

Hi,

Is there any known memory issue with mon? We have 3 mons running, and
on keeps on crashing after 2 or 3 days, and I think it's because mon
sucks up all memory.

Here's mon after starting for 10 minutes:

   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND 

13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon 

  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd 

  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd 

  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd 

  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd 

  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd 

   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 
kworker/4:1
  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 
jbd2/sda3-8
  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 
xfsaild/sdf1
  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd 

  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd 

  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd 


Here's the memory situation of mon on another machine, after mon has
been running for 3 hours:

   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND 

  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon 

  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd 

  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd 

  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd 

  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd 

  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd 

  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd 

  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd 

  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 
jbd2/sda3-8
  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd 

  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd

And here is the situation on a third node, mon has been running
for over a week:

   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND 

  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon 

  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd 

  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd 

  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd 

  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd 

  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd 

  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd 


So, after a while, sooner or later, mon is going to crash, just
a matter of time.

Does anyone see anything like this? This is kinda scary.

OS: Debian Wheezy 3.2.0-3-amd64
Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)

With this issue on hand, I'll have to monitor it closely and
restart mon once in a while, or I will get a crash (which is
still good enough), or a system that does not respond at
all because memory is exhausted, and the whole ceph cluster
is unreachable. We had this problem in the morning, mon on one
node exhausted the memory, none of the ceph command responds
anymore, the only thing left to do is to hard reset the node.
The whole cluster was basically done at that time.

Here is our usage situation:

1) A few applications which read and write data through
librados API, we have about 20-30 connections at any one time.
So far, our apps have no such memory issue, we have been
monitoring them closely.

2) We have a few scripts which pull data from an old storage
system, and use the rados command to put it into ceph.
Basically, just shell script. Each rados command is run
to write one object (one file), and exit. We run about
25 scripts simultaneously, which means at any one time,
there are at most 25 connections.

I don't think this is a very busy system. But this
memory issue is definitely a problem for us.

Thanks for helping.

Xiaopong

^ permalink raw reply	[flat|nested] 9+ messages in thread

* e: mon memory issue
  2012-08-31  9:02 mon memory issue Xiaopong Tran
@ 2012-08-31 16:03 ` Sage Weil
  2012-08-31 18:34   ` Sławomir Skowron
  0 siblings, 1 reply; 9+ messages in thread
From: Sage Weil @ 2012-08-31 16:03 UTC (permalink / raw)
  To: Xiaopong Tran; +Cc: ceph-devel

On Fri, 31 Aug 2012, Xiaopong Tran wrote:

> Hi,
> 
> Is there any known memory issue with mon? We have 3 mons running, and
> on keeps on crashing after 2 or 3 days, and I think it's because mon
> sucks up all memory.
> 
> Here's mon after starting for 10 minutes:
> 
>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND 
> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon 
>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd 
>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd 
>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd 
>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd 
>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd 
>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd 
>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd 
>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd 
> 
> Here's the memory situation of mon on another machine, after mon has
> been running for 3 hours:
> 
>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND 
>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon 
>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd 
>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd 
>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd 
>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd 
>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd 
>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd 
>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd 
>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd 
>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
> 
> And here is the situation on a third node, mon has been running
> for over a week:
> 
>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND 
>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon 
>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd 
>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd 
>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd 
>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd 
>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd 
>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd 
> 
> So, after a while, sooner or later, mon is going to crash, just
> a matter of time.
> 
> Does anyone see anything like this? This is kinda scary.
> 
> OS: Debian Wheezy 3.2.0-3-amd64
> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)

Can you try with 0.48.1argonaut? 

If it still happens, can you run ceph-mon through massif? 

 valgrind --tool=massif ceph-mon -i whatever

That'll generate a massif.out file (make sure it's there; you may need to 
specify the output file for valgrind) over time.  Once ceph-mon starts 
eating ram, send us a copy of the file and we can hopefully see what is 
leaking.

Thanks!
sage


> 
> With this issue on hand, I'll have to monitor it closely and
> restart mon once in a while, or I will get a crash (which is
> still good enough), or a system that does not respond at
> all because memory is exhausted, and the whole ceph cluster
> is unreachable. We had this problem in the morning, mon on one
> node exhausted the memory, none of the ceph command responds
> anymore, the only thing left to do is to hard reset the node.
> The whole cluster was basically done at that time.
> 
> Here is our usage situation:
> 
> 1) A few applications which read and write data through
> librados API, we have about 20-30 connections at any one time.
> So far, our apps have no such memory issue, we have been
> monitoring them closely.
> 
> 2) We have a few scripts which pull data from an old storage
> system, and use the rados command to put it into ceph.
> Basically, just shell script. Each rados command is run
> to write one object (one file), and exit. We run about
> 25 scripts simultaneously, which means at any one time,
> there are at most 25 connections.
> 
> I don't think this is a very busy system. But this
> memory issue is definitely a problem for us.
> 
> Thanks for helping.
> 
> Xiaopong
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: e: mon memory issue
  2012-08-31 16:03 ` e: " Sage Weil
@ 2012-08-31 18:34   ` Sławomir Skowron
  2012-09-04 12:00     ` Sławomir Skowron
  0 siblings, 1 reply; 9+ messages in thread
From: Sławomir Skowron @ 2012-08-31 18:34 UTC (permalink / raw)
  To: Sage Weil; +Cc: Xiaopong Tran, ceph-devel

I have this problem too. My mon's in 0.48.1 cluster have 10GB RAM
each, with 78 osd, and 2k request per minute (max) in radosgw.

Now i have run one via valgrind. I will send output when mon grow up.

On Fri, Aug 31, 2012 at 6:03 PM, Sage Weil <sage@inktank.com> wrote:
> On Fri, 31 Aug 2012, Xiaopong Tran wrote:
>
>> Hi,
>>
>> Is there any known memory issue with mon? We have 3 mons running, and
>> on keeps on crashing after 2 or 3 days, and I think it's because mon
>> sucks up all memory.
>>
>> Here's mon after starting for 10 minutes:
>>
>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon
>>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd
>>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd
>>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd
>>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd
>>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd
>>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
>>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
>>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
>>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd
>>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd
>>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd
>>
>> Here's the memory situation of mon on another machine, after mon has
>> been running for 3 hours:
>>
>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon
>>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd
>>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd
>>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd
>>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd
>>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd
>>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd
>>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd
>>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
>>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd
>>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
>>
>> And here is the situation on a third node, mon has been running
>> for over a week:
>>
>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon
>>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd
>>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd
>>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd
>>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd
>>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd
>>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd
>>
>> So, after a while, sooner or later, mon is going to crash, just
>> a matter of time.
>>
>> Does anyone see anything like this? This is kinda scary.
>>
>> OS: Debian Wheezy 3.2.0-3-amd64
>> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)
>
> Can you try with 0.48.1argonaut?
>
> If it still happens, can you run ceph-mon through massif?
>
>  valgrind --tool=massif ceph-mon -i whatever
>
> That'll generate a massif.out file (make sure it's there; you may need to
> specify the output file for valgrind) over time.  Once ceph-mon starts
> eating ram, send us a copy of the file and we can hopefully see what is
> leaking.
>
> Thanks!
> sage
>
>
>>
>> With this issue on hand, I'll have to monitor it closely and
>> restart mon once in a while, or I will get a crash (which is
>> still good enough), or a system that does not respond at
>> all because memory is exhausted, and the whole ceph cluster
>> is unreachable. We had this problem in the morning, mon on one
>> node exhausted the memory, none of the ceph command responds
>> anymore, the only thing left to do is to hard reset the node.
>> The whole cluster was basically done at that time.
>>
>> Here is our usage situation:
>>
>> 1) A few applications which read and write data through
>> librados API, we have about 20-30 connections at any one time.
>> So far, our apps have no such memory issue, we have been
>> monitoring them closely.
>>
>> 2) We have a few scripts which pull data from an old storage
>> system, and use the rados command to put it into ceph.
>> Basically, just shell script. Each rados command is run
>> to write one object (one file), and exit. We run about
>> 25 scripts simultaneously, which means at any one time,
>> there are at most 25 connections.
>>
>> I don't think this is a very busy system. But this
>> memory issue is definitely a problem for us.
>>
>> Thanks for helping.
>>
>> Xiaopong
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html



-- 
-----
Pozdrawiam

Sławek "sZiBis" Skowron
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: e: mon memory issue
  2012-08-31 18:34   ` Sławomir Skowron
@ 2012-09-04 12:00     ` Sławomir Skowron
  2012-09-05  3:32       ` Sage Weil
  0 siblings, 1 reply; 9+ messages in thread
From: Sławomir Skowron @ 2012-09-04 12:00 UTC (permalink / raw)
  To: Sage Weil; +Cc: Xiaopong Tran, ceph-devel

Valgrind returns nothing.

valgrind --tool=massif --log-file=ceph_mon_valgrind ceph-mon -i 0 > log.txt

==30491== Massif, a heap profiler
==30491== Copyright (C) 2003-2011, and GNU GPL'd, by Nicholas Nethercote
==30491== Using Valgrind-3.7.0 and LibVEX; rerun with -h for copyright info
==30491== Command: ceph-mon -i 0
==30491== Parent PID: 4013
==30491==
==30491==

cat massif.out.26201
desc: (none)
cmd: ceph-mon -i 0
time_unit: i
#-----------
snapshot=0
#-----------
time=0
mem_heap_B=0
mem_heap_extra_B=0
mem_stacks_B=0
heap_tree=empty

What i have done wrong ??

On Fri, Aug 31, 2012 at 8:34 PM, Sławomir Skowron <szibis@gmail.com> wrote:
> I have this problem too. My mon's in 0.48.1 cluster have 10GB RAM
> each, with 78 osd, and 2k request per minute (max) in radosgw.
>
> Now i have run one via valgrind. I will send output when mon grow up.
>
> On Fri, Aug 31, 2012 at 6:03 PM, Sage Weil <sage@inktank.com> wrote:
>> On Fri, 31 Aug 2012, Xiaopong Tran wrote:
>>
>>> Hi,
>>>
>>> Is there any known memory issue with mon? We have 3 mons running, and
>>> on keeps on crashing after 2 or 3 days, and I think it's because mon
>>> sucks up all memory.
>>>
>>> Here's mon after starting for 10 minutes:
>>>
>>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon
>>>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd
>>>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd
>>>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd
>>>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd
>>>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd
>>>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
>>>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
>>>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
>>>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd
>>>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd
>>>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd
>>>
>>> Here's the memory situation of mon on another machine, after mon has
>>> been running for 3 hours:
>>>
>>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon
>>>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd
>>>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd
>>>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd
>>>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd
>>>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd
>>>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd
>>>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd
>>>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
>>>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd
>>>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
>>>
>>> And here is the situation on a third node, mon has been running
>>> for over a week:
>>>
>>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon
>>>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd
>>>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd
>>>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd
>>>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd
>>>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd
>>>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd
>>>
>>> So, after a while, sooner or later, mon is going to crash, just
>>> a matter of time.
>>>
>>> Does anyone see anything like this? This is kinda scary.
>>>
>>> OS: Debian Wheezy 3.2.0-3-amd64
>>> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)
>>
>> Can you try with 0.48.1argonaut?
>>
>> If it still happens, can you run ceph-mon through massif?
>>
>>  valgrind --tool=massif ceph-mon -i whatever
>>
>> That'll generate a massif.out file (make sure it's there; you may need to
>> specify the output file for valgrind) over time.  Once ceph-mon starts
>> eating ram, send us a copy of the file and we can hopefully see what is
>> leaking.
>>
>> Thanks!
>> sage
>>
>>
>>>
>>> With this issue on hand, I'll have to monitor it closely and
>>> restart mon once in a while, or I will get a crash (which is
>>> still good enough), or a system that does not respond at
>>> all because memory is exhausted, and the whole ceph cluster
>>> is unreachable. We had this problem in the morning, mon on one
>>> node exhausted the memory, none of the ceph command responds
>>> anymore, the only thing left to do is to hard reset the node.
>>> The whole cluster was basically done at that time.
>>>
>>> Here is our usage situation:
>>>
>>> 1) A few applications which read and write data through
>>> librados API, we have about 20-30 connections at any one time.
>>> So far, our apps have no such memory issue, we have been
>>> monitoring them closely.
>>>
>>> 2) We have a few scripts which pull data from an old storage
>>> system, and use the rados command to put it into ceph.
>>> Basically, just shell script. Each rados command is run
>>> to write one object (one file), and exit. We run about
>>> 25 scripts simultaneously, which means at any one time,
>>> there are at most 25 connections.
>>>
>>> I don't think this is a very busy system. But this
>>> memory issue is definitely a problem for us.
>>>
>>> Thanks for helping.
>>>
>>> Xiaopong
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> the body of a message to majordomo@vger.kernel.org
>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>>
>>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
>
>
> --
> -----
> Pozdrawiam
>
> Sławek "sZiBis" Skowron



-- 
-----
Pozdrawiam

Sławek "sZiBis" Skowron
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: e: mon memory issue
  2012-09-04 12:00     ` Sławomir Skowron
@ 2012-09-05  3:32       ` Sage Weil
  2012-09-05 10:32         ` Sławomir Skowron
  0 siblings, 1 reply; 9+ messages in thread
From: Sage Weil @ 2012-09-05  3:32 UTC (permalink / raw)
  To: Sławomir Skowron; +Cc: Xiaopong Tran, ceph-devel

On Tue, 4 Sep 2012, S?awomir Skowron wrote:
> Valgrind returns nothing.
> 
> valgrind --tool=massif --log-file=ceph_mon_valgrind ceph-mon -i 0 > log.txt

The fork is probably confusing it.  I usually pass -f to ceph-mon (or 
ceph-osd etc) to keep it in the foreground.  Can you give that a go?  
e.g.,

	valgrind --tool-massif ceph-mon -i 0 -f &

and watch for the massif.out.$pid file.

Thanks!
sage


> 
> ==30491== Massif, a heap profiler
> ==30491== Copyright (C) 2003-2011, and GNU GPL'd, by Nicholas Nethercote
> ==30491== Using Valgrind-3.7.0 and LibVEX; rerun with -h for copyright info
> ==30491== Command: ceph-mon -i 0
> ==30491== Parent PID: 4013
> ==30491==
> ==30491==
> 
> cat massif.out.26201
> desc: (none)
> cmd: ceph-mon -i 0
> time_unit: i
> #-----------
> snapshot=0
> #-----------
> time=0
> mem_heap_B=0
> mem_heap_extra_B=0
> mem_stacks_B=0
> heap_tree=empty
> 
> What i have done wrong ??
> 
> On Fri, Aug 31, 2012 at 8:34 PM, S?awomir Skowron <szibis@gmail.com> wrote:
> > I have this problem too. My mon's in 0.48.1 cluster have 10GB RAM
> > each, with 78 osd, and 2k request per minute (max) in radosgw.
> >
> > Now i have run one via valgrind. I will send output when mon grow up.
> >
> > On Fri, Aug 31, 2012 at 6:03 PM, Sage Weil <sage@inktank.com> wrote:
> >> On Fri, 31 Aug 2012, Xiaopong Tran wrote:
> >>
> >>> Hi,
> >>>
> >>> Is there any known memory issue with mon? We have 3 mons running, and
> >>> on keeps on crashing after 2 or 3 days, and I think it's because mon
> >>> sucks up all memory.
> >>>
> >>> Here's mon after starting for 10 minutes:
> >>>
> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
> >>> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon
> >>>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd
> >>>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd
> >>>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd
> >>>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd
> >>>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd
> >>>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
> >>>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
> >>>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
> >>>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd
> >>>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd
> >>>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd
> >>>
> >>> Here's the memory situation of mon on another machine, after mon has
> >>> been running for 3 hours:
> >>>
> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
> >>>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon
> >>>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd
> >>>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd
> >>>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd
> >>>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd
> >>>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd
> >>>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd
> >>>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd
> >>>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
> >>>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd
> >>>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
> >>>
> >>> And here is the situation on a third node, mon has been running
> >>> for over a week:
> >>>
> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
> >>>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon
> >>>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd
> >>>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd
> >>>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd
> >>>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd
> >>>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd
> >>>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd
> >>>
> >>> So, after a while, sooner or later, mon is going to crash, just
> >>> a matter of time.
> >>>
> >>> Does anyone see anything like this? This is kinda scary.
> >>>
> >>> OS: Debian Wheezy 3.2.0-3-amd64
> >>> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)
> >>
> >> Can you try with 0.48.1argonaut?
> >>
> >> If it still happens, can you run ceph-mon through massif?
> >>
> >>  valgrind --tool=massif ceph-mon -i whatever
> >>
> >> That'll generate a massif.out file (make sure it's there; you may need to
> >> specify the output file for valgrind) over time.  Once ceph-mon starts
> >> eating ram, send us a copy of the file and we can hopefully see what is
> >> leaking.
> >>
> >> Thanks!
> >> sage
> >>
> >>
> >>>
> >>> With this issue on hand, I'll have to monitor it closely and
> >>> restart mon once in a while, or I will get a crash (which is
> >>> still good enough), or a system that does not respond at
> >>> all because memory is exhausted, and the whole ceph cluster
> >>> is unreachable. We had this problem in the morning, mon on one
> >>> node exhausted the memory, none of the ceph command responds
> >>> anymore, the only thing left to do is to hard reset the node.
> >>> The whole cluster was basically done at that time.
> >>>
> >>> Here is our usage situation:
> >>>
> >>> 1) A few applications which read and write data through
> >>> librados API, we have about 20-30 connections at any one time.
> >>> So far, our apps have no such memory issue, we have been
> >>> monitoring them closely.
> >>>
> >>> 2) We have a few scripts which pull data from an old storage
> >>> system, and use the rados command to put it into ceph.
> >>> Basically, just shell script. Each rados command is run
> >>> to write one object (one file), and exit. We run about
> >>> 25 scripts simultaneously, which means at any one time,
> >>> there are at most 25 connections.
> >>>
> >>> I don't think this is a very busy system. But this
> >>> memory issue is definitely a problem for us.
> >>>
> >>> Thanks for helping.
> >>>
> >>> Xiaopong
> >>> --
> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >>> the body of a message to majordomo@vger.kernel.org
> >>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>>
> >>>
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> the body of a message to majordomo@vger.kernel.org
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >
> >
> >
> > --
> > -----
> > Pozdrawiam
> >
> > S?awek "sZiBis" Skowron
> 
> 
> 
> -- 
> -----
> Pozdrawiam
> 
> S?awek "sZiBis" Skowron
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: e: mon memory issue
  2012-09-05  3:32       ` Sage Weil
@ 2012-09-05 10:32         ` Sławomir Skowron
  2012-09-05 15:51           ` Sage Weil
  0 siblings, 1 reply; 9+ messages in thread
From: Sławomir Skowron @ 2012-09-05 10:32 UTC (permalink / raw)
  To: Sage Weil; +Cc: Xiaopong Tran, ceph-devel

Unfortunately here is the problem in my Ubuntu 12.04.1

--9399-- You may be able to write your own handler.
--9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
--9399-- Nevertheless we consider this a bug.  Please report
--9399-- it at http://valgrind.org/support/bug_reports.html.
==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
==9399==    This could cause spurious value errors to appear.
==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
writing a proper wrapper.
--9399-- WARNING: unhandled syscall: 306
--9399-- You may be able to write your own handler.
--9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
--9399-- Nevertheless we consider this a bug.  Please report
--9399-- it at http://valgrind.org/support/bug_reports.html.
==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
==9399==    This could cause spurious value errors to appear.
==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
writing a proper wrapper.
^C2012-09-05 09:13:18.660048 a964700 -1 mon.0@0(leader) e4 *** Got
Signal Interrupt ***
==9399==


On Wed, Sep 5, 2012 at 5:32 AM, Sage Weil <sage@inktank.com> wrote:
> On Tue, 4 Sep 2012, S?awomir Skowron wrote:
>> Valgrind returns nothing.
>>
>> valgrind --tool=massif --log-file=ceph_mon_valgrind ceph-mon -i 0 > log.txt
>
> The fork is probably confusing it.  I usually pass -f to ceph-mon (or
> ceph-osd etc) to keep it in the foreground.  Can you give that a go?
> e.g.,
>
>         valgrind --tool-massif ceph-mon -i 0 -f &
>
> and watch for the massif.out.$pid file.
>
> Thanks!
> sage
>
>
>>
>> ==30491== Massif, a heap profiler
>> ==30491== Copyright (C) 2003-2011, and GNU GPL'd, by Nicholas Nethercote
>> ==30491== Using Valgrind-3.7.0 and LibVEX; rerun with -h for copyright info
>> ==30491== Command: ceph-mon -i 0
>> ==30491== Parent PID: 4013
>> ==30491==
>> ==30491==
>>
>> cat massif.out.26201
>> desc: (none)
>> cmd: ceph-mon -i 0
>> time_unit: i
>> #-----------
>> snapshot=0
>> #-----------
>> time=0
>> mem_heap_B=0
>> mem_heap_extra_B=0
>> mem_stacks_B=0
>> heap_tree=empty
>>
>> What i have done wrong ??
>>
>> On Fri, Aug 31, 2012 at 8:34 PM, S?awomir Skowron <szibis@gmail.com> wrote:
>> > I have this problem too. My mon's in 0.48.1 cluster have 10GB RAM
>> > each, with 78 osd, and 2k request per minute (max) in radosgw.
>> >
>> > Now i have run one via valgrind. I will send output when mon grow up.
>> >
>> > On Fri, Aug 31, 2012 at 6:03 PM, Sage Weil <sage@inktank.com> wrote:
>> >> On Fri, 31 Aug 2012, Xiaopong Tran wrote:
>> >>
>> >>> Hi,
>> >>>
>> >>> Is there any known memory issue with mon? We have 3 mons running, and
>> >>> on keeps on crashing after 2 or 3 days, and I think it's because mon
>> >>> sucks up all memory.
>> >>>
>> >>> Here's mon after starting for 10 minutes:
>> >>>
>> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>> >>> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon
>> >>>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd
>> >>>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd
>> >>>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd
>> >>>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd
>> >>>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd
>> >>>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
>> >>>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
>> >>>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
>> >>>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd
>> >>>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd
>> >>>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd
>> >>>
>> >>> Here's the memory situation of mon on another machine, after mon has
>> >>> been running for 3 hours:
>> >>>
>> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>> >>>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon
>> >>>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd
>> >>>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd
>> >>>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd
>> >>>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd
>> >>>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd
>> >>>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd
>> >>>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd
>> >>>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
>> >>>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd
>> >>>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
>> >>>
>> >>> And here is the situation on a third node, mon has been running
>> >>> for over a week:
>> >>>
>> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>> >>>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon
>> >>>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd
>> >>>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd
>> >>>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd
>> >>>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd
>> >>>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd
>> >>>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd
>> >>>
>> >>> So, after a while, sooner or later, mon is going to crash, just
>> >>> a matter of time.
>> >>>
>> >>> Does anyone see anything like this? This is kinda scary.
>> >>>
>> >>> OS: Debian Wheezy 3.2.0-3-amd64
>> >>> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)
>> >>
>> >> Can you try with 0.48.1argonaut?
>> >>
>> >> If it still happens, can you run ceph-mon through massif?
>> >>
>> >>  valgrind --tool=massif ceph-mon -i whatever
>> >>
>> >> That'll generate a massif.out file (make sure it's there; you may need to
>> >> specify the output file for valgrind) over time.  Once ceph-mon starts
>> >> eating ram, send us a copy of the file and we can hopefully see what is
>> >> leaking.
>> >>
>> >> Thanks!
>> >> sage
>> >>
>> >>
>> >>>
>> >>> With this issue on hand, I'll have to monitor it closely and
>> >>> restart mon once in a while, or I will get a crash (which is
>> >>> still good enough), or a system that does not respond at
>> >>> all because memory is exhausted, and the whole ceph cluster
>> >>> is unreachable. We had this problem in the morning, mon on one
>> >>> node exhausted the memory, none of the ceph command responds
>> >>> anymore, the only thing left to do is to hard reset the node.
>> >>> The whole cluster was basically done at that time.
>> >>>
>> >>> Here is our usage situation:
>> >>>
>> >>> 1) A few applications which read and write data through
>> >>> librados API, we have about 20-30 connections at any one time.
>> >>> So far, our apps have no such memory issue, we have been
>> >>> monitoring them closely.
>> >>>
>> >>> 2) We have a few scripts which pull data from an old storage
>> >>> system, and use the rados command to put it into ceph.
>> >>> Basically, just shell script. Each rados command is run
>> >>> to write one object (one file), and exit. We run about
>> >>> 25 scripts simultaneously, which means at any one time,
>> >>> there are at most 25 connections.
>> >>>
>> >>> I don't think this is a very busy system. But this
>> >>> memory issue is definitely a problem for us.
>> >>>
>> >>> Thanks for helping.
>> >>>
>> >>> Xiaopong
>> >>> --
>> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> >>> the body of a message to majordomo@vger.kernel.org
>> >>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >>>
>> >>>
>> >> --
>> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> >> the body of a message to majordomo@vger.kernel.org
>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >
>> >
>> >
>> > --
>> > -----
>> > Pozdrawiam
>> >
>> > S?awek "sZiBis" Skowron
>>
>>
>>
>> --
>> -----
>> Pozdrawiam
>>
>> S?awek "sZiBis" Skowron
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>>



-- 
-----
Pozdrawiam

Sławek "sZiBis" Skowron
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: e: mon memory issue
  2012-09-05 10:32         ` Sławomir Skowron
@ 2012-09-05 15:51           ` Sage Weil
  2012-09-05 18:44             ` Sławomir Skowron
  0 siblings, 1 reply; 9+ messages in thread
From: Sage Weil @ 2012-09-05 15:51 UTC (permalink / raw)
  To: Sławomir Skowron; +Cc: Xiaopong Tran, ceph-devel

On Wed, 5 Sep 2012, S?awomir Skowron wrote:
> Unfortunately here is the problem in my Ubuntu 12.04.1
> 
> --9399-- You may be able to write your own handler.
> --9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
> --9399-- Nevertheless we consider this a bug.  Please report
> --9399-- it at http://valgrind.org/support/bug_reports.html.
> ==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
> ==9399==    This could cause spurious value errors to appear.
> ==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
> writing a proper wrapper.
> --9399-- WARNING: unhandled syscall: 306
> --9399-- You may be able to write your own handler.
> --9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
> --9399-- Nevertheless we consider this a bug.  Please report
> --9399-- it at http://valgrind.org/support/bug_reports.html.
> ==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
> ==9399==    This could cause spurious value errors to appear.
> ==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
> writing a proper wrapper.

These are harmless; it just doesn't recognize syncfs(2) or one of the 
ioctls, but everything else works.

> ^C2012-09-05 09:13:18.660048 a964700 -1 mon.0@0(leader) e4 *** Got
> Signal Interrupt ***
> ==9399==

Did you hit control-c?

If you leave it running it should gather the memory utilization info we 
need...

sage

> 
> 
> On Wed, Sep 5, 2012 at 5:32 AM, Sage Weil <sage@inktank.com> wrote:
> > On Tue, 4 Sep 2012, S?awomir Skowron wrote:
> >> Valgrind returns nothing.
> >>
> >> valgrind --tool=massif --log-file=ceph_mon_valgrind ceph-mon -i 0 > log.txt
> >
> > The fork is probably confusing it.  I usually pass -f to ceph-mon (or
> > ceph-osd etc) to keep it in the foreground.  Can you give that a go?
> > e.g.,
> >
> >         valgrind --tool-massif ceph-mon -i 0 -f &
> >
> > and watch for the massif.out.$pid file.
> >
> > Thanks!
> > sage
> >
> >
> >>
> >> ==30491== Massif, a heap profiler
> >> ==30491== Copyright (C) 2003-2011, and GNU GPL'd, by Nicholas Nethercote
> >> ==30491== Using Valgrind-3.7.0 and LibVEX; rerun with -h for copyright info
> >> ==30491== Command: ceph-mon -i 0
> >> ==30491== Parent PID: 4013
> >> ==30491==
> >> ==30491==
> >>
> >> cat massif.out.26201
> >> desc: (none)
> >> cmd: ceph-mon -i 0
> >> time_unit: i
> >> #-----------
> >> snapshot=0
> >> #-----------
> >> time=0
> >> mem_heap_B=0
> >> mem_heap_extra_B=0
> >> mem_stacks_B=0
> >> heap_tree=empty
> >>
> >> What i have done wrong ??
> >>
> >> On Fri, Aug 31, 2012 at 8:34 PM, S?awomir Skowron <szibis@gmail.com> wrote:
> >> > I have this problem too. My mon's in 0.48.1 cluster have 10GB RAM
> >> > each, with 78 osd, and 2k request per minute (max) in radosgw.
> >> >
> >> > Now i have run one via valgrind. I will send output when mon grow up.
> >> >
> >> > On Fri, Aug 31, 2012 at 6:03 PM, Sage Weil <sage@inktank.com> wrote:
> >> >> On Fri, 31 Aug 2012, Xiaopong Tran wrote:
> >> >>
> >> >>> Hi,
> >> >>>
> >> >>> Is there any known memory issue with mon? We have 3 mons running, and
> >> >>> on keeps on crashing after 2 or 3 days, and I think it's because mon
> >> >>> sucks up all memory.
> >> >>>
> >> >>> Here's mon after starting for 10 minutes:
> >> >>>
> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
> >> >>> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon
> >> >>>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd
> >> >>>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd
> >> >>>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd
> >> >>>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd
> >> >>>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd
> >> >>>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
> >> >>>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
> >> >>>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
> >> >>>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd
> >> >>>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd
> >> >>>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd
> >> >>>
> >> >>> Here's the memory situation of mon on another machine, after mon has
> >> >>> been running for 3 hours:
> >> >>>
> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
> >> >>>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon
> >> >>>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd
> >> >>>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd
> >> >>>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd
> >> >>>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd
> >> >>>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd
> >> >>>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd
> >> >>>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd
> >> >>>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
> >> >>>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd
> >> >>>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
> >> >>>
> >> >>> And here is the situation on a third node, mon has been running
> >> >>> for over a week:
> >> >>>
> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
> >> >>>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon
> >> >>>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd
> >> >>>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd
> >> >>>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd
> >> >>>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd
> >> >>>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd
> >> >>>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd
> >> >>>
> >> >>> So, after a while, sooner or later, mon is going to crash, just
> >> >>> a matter of time.
> >> >>>
> >> >>> Does anyone see anything like this? This is kinda scary.
> >> >>>
> >> >>> OS: Debian Wheezy 3.2.0-3-amd64
> >> >>> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)
> >> >>
> >> >> Can you try with 0.48.1argonaut?
> >> >>
> >> >> If it still happens, can you run ceph-mon through massif?
> >> >>
> >> >>  valgrind --tool=massif ceph-mon -i whatever
> >> >>
> >> >> That'll generate a massif.out file (make sure it's there; you may need to
> >> >> specify the output file for valgrind) over time.  Once ceph-mon starts
> >> >> eating ram, send us a copy of the file and we can hopefully see what is
> >> >> leaking.
> >> >>
> >> >> Thanks!
> >> >> sage
> >> >>
> >> >>
> >> >>>
> >> >>> With this issue on hand, I'll have to monitor it closely and
> >> >>> restart mon once in a while, or I will get a crash (which is
> >> >>> still good enough), or a system that does not respond at
> >> >>> all because memory is exhausted, and the whole ceph cluster
> >> >>> is unreachable. We had this problem in the morning, mon on one
> >> >>> node exhausted the memory, none of the ceph command responds
> >> >>> anymore, the only thing left to do is to hard reset the node.
> >> >>> The whole cluster was basically done at that time.
> >> >>>
> >> >>> Here is our usage situation:
> >> >>>
> >> >>> 1) A few applications which read and write data through
> >> >>> librados API, we have about 20-30 connections at any one time.
> >> >>> So far, our apps have no such memory issue, we have been
> >> >>> monitoring them closely.
> >> >>>
> >> >>> 2) We have a few scripts which pull data from an old storage
> >> >>> system, and use the rados command to put it into ceph.
> >> >>> Basically, just shell script. Each rados command is run
> >> >>> to write one object (one file), and exit. We run about
> >> >>> 25 scripts simultaneously, which means at any one time,
> >> >>> there are at most 25 connections.
> >> >>>
> >> >>> I don't think this is a very busy system. But this
> >> >>> memory issue is definitely a problem for us.
> >> >>>
> >> >>> Thanks for helping.
> >> >>>
> >> >>> Xiaopong
> >> >>> --
> >> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> >>> the body of a message to majordomo@vger.kernel.org
> >> >>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >>>
> >> >>>
> >> >> --
> >> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> >> the body of a message to majordomo@vger.kernel.org
> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> >
> >> >
> >> >
> >> > --
> >> > -----
> >> > Pozdrawiam
> >> >
> >> > S?awek "sZiBis" Skowron
> >>
> >>
> >>
> >> --
> >> -----
> >> Pozdrawiam
> >>
> >> S?awek "sZiBis" Skowron
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> the body of a message to majordomo@vger.kernel.org
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>
> >>
> 
> 
> 
> -- 
> -----
> Pozdrawiam
> 
> S?awek "sZiBis" Skowron
> 
> 

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: e: mon memory issue
  2012-09-05 15:51           ` Sage Weil
@ 2012-09-05 18:44             ` Sławomir Skowron
  2012-09-10 10:29               ` Sławomir Skowron
  0 siblings, 1 reply; 9+ messages in thread
From: Sławomir Skowron @ 2012-09-05 18:44 UTC (permalink / raw)
  To: Sage Weil; +Cc: Xiaopong Tran, ceph-devel

On Wed, Sep 5, 2012 at 5:51 PM, Sage Weil <sage@inktank.com> wrote:
> On Wed, 5 Sep 2012, S?awomir Skowron wrote:
>> Unfortunately here is the problem in my Ubuntu 12.04.1
>>
>> --9399-- You may be able to write your own handler.
>> --9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
>> --9399-- Nevertheless we consider this a bug.  Please report
>> --9399-- it at http://valgrind.org/support/bug_reports.html.
>> ==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
>> ==9399==    This could cause spurious value errors to appear.
>> ==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
>> writing a proper wrapper.
>> --9399-- WARNING: unhandled syscall: 306
>> --9399-- You may be able to write your own handler.
>> --9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
>> --9399-- Nevertheless we consider this a bug.  Please report
>> --9399-- it at http://valgrind.org/support/bug_reports.html.
>> ==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
>> ==9399==    This could cause spurious value errors to appear.
>> ==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
>> writing a proper wrapper.
>
> These are harmless; it just doesn't recognize syncfs(2) or one of the
> ioctls, but everything else works.
>
>> ^C2012-09-05 09:13:18.660048 a964700 -1 mon.0@0(leader) e4 *** Got
>> Signal Interrupt ***
>> ==9399==
>
> Did you hit control-c?
>
> If you leave it running it should gather the memory utilization info we
> need...

Yes it's running now, and  i will see tomorrow how much memory mon consumes.

>
> sage
>
>>
>>
>> On Wed, Sep 5, 2012 at 5:32 AM, Sage Weil <sage@inktank.com> wrote:
>> > On Tue, 4 Sep 2012, S?awomir Skowron wrote:
>> >> Valgrind returns nothing.
>> >>
>> >> valgrind --tool=massif --log-file=ceph_mon_valgrind ceph-mon -i 0 > log.txt
>> >
>> > The fork is probably confusing it.  I usually pass -f to ceph-mon (or
>> > ceph-osd etc) to keep it in the foreground.  Can you give that a go?
>> > e.g.,
>> >
>> >         valgrind --tool-massif ceph-mon -i 0 -f &
>> >
>> > and watch for the massif.out.$pid file.
>> >
>> > Thanks!
>> > sage
>> >
>> >
>> >>
>> >> ==30491== Massif, a heap profiler
>> >> ==30491== Copyright (C) 2003-2011, and GNU GPL'd, by Nicholas Nethercote
>> >> ==30491== Using Valgrind-3.7.0 and LibVEX; rerun with -h for copyright info
>> >> ==30491== Command: ceph-mon -i 0
>> >> ==30491== Parent PID: 4013
>> >> ==30491==
>> >> ==30491==
>> >>
>> >> cat massif.out.26201
>> >> desc: (none)
>> >> cmd: ceph-mon -i 0
>> >> time_unit: i
>> >> #-----------
>> >> snapshot=0
>> >> #-----------
>> >> time=0
>> >> mem_heap_B=0
>> >> mem_heap_extra_B=0
>> >> mem_stacks_B=0
>> >> heap_tree=empty
>> >>
>> >> What i have done wrong ??
>> >>
>> >> On Fri, Aug 31, 2012 at 8:34 PM, S?awomir Skowron <szibis@gmail.com> wrote:
>> >> > I have this problem too. My mon's in 0.48.1 cluster have 10GB RAM
>> >> > each, with 78 osd, and 2k request per minute (max) in radosgw.
>> >> >
>> >> > Now i have run one via valgrind. I will send output when mon grow up.
>> >> >
>> >> > On Fri, Aug 31, 2012 at 6:03 PM, Sage Weil <sage@inktank.com> wrote:
>> >> >> On Fri, 31 Aug 2012, Xiaopong Tran wrote:
>> >> >>
>> >> >>> Hi,
>> >> >>>
>> >> >>> Is there any known memory issue with mon? We have 3 mons running, and
>> >> >>> on keeps on crashing after 2 or 3 days, and I think it's because mon
>> >> >>> sucks up all memory.
>> >> >>>
>> >> >>> Here's mon after starting for 10 minutes:
>> >> >>>
>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>> >> >>> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon
>> >> >>>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd
>> >> >>>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd
>> >> >>>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd
>> >> >>>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd
>> >> >>>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd
>> >> >>>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
>> >> >>>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
>> >> >>>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
>> >> >>>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd
>> >> >>>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd
>> >> >>>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd
>> >> >>>
>> >> >>> Here's the memory situation of mon on another machine, after mon has
>> >> >>> been running for 3 hours:
>> >> >>>
>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>> >> >>>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon
>> >> >>>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd
>> >> >>>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd
>> >> >>>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd
>> >> >>>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd
>> >> >>>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd
>> >> >>>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd
>> >> >>>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd
>> >> >>>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
>> >> >>>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd
>> >> >>>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
>> >> >>>
>> >> >>> And here is the situation on a third node, mon has been running
>> >> >>> for over a week:
>> >> >>>
>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>> >> >>>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon
>> >> >>>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd
>> >> >>>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd
>> >> >>>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd
>> >> >>>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd
>> >> >>>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd
>> >> >>>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd
>> >> >>>
>> >> >>> So, after a while, sooner or later, mon is going to crash, just
>> >> >>> a matter of time.
>> >> >>>
>> >> >>> Does anyone see anything like this? This is kinda scary.
>> >> >>>
>> >> >>> OS: Debian Wheezy 3.2.0-3-amd64
>> >> >>> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)
>> >> >>
>> >> >> Can you try with 0.48.1argonaut?
>> >> >>
>> >> >> If it still happens, can you run ceph-mon through massif?
>> >> >>
>> >> >>  valgrind --tool=massif ceph-mon -i whatever
>> >> >>
>> >> >> That'll generate a massif.out file (make sure it's there; you may need to
>> >> >> specify the output file for valgrind) over time.  Once ceph-mon starts
>> >> >> eating ram, send us a copy of the file and we can hopefully see what is
>> >> >> leaking.
>> >> >>
>> >> >> Thanks!
>> >> >> sage
>> >> >>
>> >> >>
>> >> >>>
>> >> >>> With this issue on hand, I'll have to monitor it closely and
>> >> >>> restart mon once in a while, or I will get a crash (which is
>> >> >>> still good enough), or a system that does not respond at
>> >> >>> all because memory is exhausted, and the whole ceph cluster
>> >> >>> is unreachable. We had this problem in the morning, mon on one
>> >> >>> node exhausted the memory, none of the ceph command responds
>> >> >>> anymore, the only thing left to do is to hard reset the node.
>> >> >>> The whole cluster was basically done at that time.
>> >> >>>
>> >> >>> Here is our usage situation:
>> >> >>>
>> >> >>> 1) A few applications which read and write data through
>> >> >>> librados API, we have about 20-30 connections at any one time.
>> >> >>> So far, our apps have no such memory issue, we have been
>> >> >>> monitoring them closely.
>> >> >>>
>> >> >>> 2) We have a few scripts which pull data from an old storage
>> >> >>> system, and use the rados command to put it into ceph.
>> >> >>> Basically, just shell script. Each rados command is run
>> >> >>> to write one object (one file), and exit. We run about
>> >> >>> 25 scripts simultaneously, which means at any one time,
>> >> >>> there are at most 25 connections.
>> >> >>>
>> >> >>> I don't think this is a very busy system. But this
>> >> >>> memory issue is definitely a problem for us.
>> >> >>>
>> >> >>> Thanks for helping.
>> >> >>>
>> >> >>> Xiaopong
>> >> >>> --
>> >> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> >> >>> the body of a message to majordomo@vger.kernel.org
>> >> >>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> >>>
>> >> >>>
>> >> >> --
>> >> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> >> >> the body of a message to majordomo@vger.kernel.org
>> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >> >
>> >> >
>> >> >
>> >> > --
>> >> > -----
>> >> > Pozdrawiam
>> >> >
>> >> > S?awek "sZiBis" Skowron
>> >>
>> >>
>> >>
>> >> --
>> >> -----
>> >> Pozdrawiam
>> >>
>> >> S?awek "sZiBis" Skowron
>> >> --
>> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> >> the body of a message to majordomo@vger.kernel.org
>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>> >>
>> >>
>>
>>
>>
>> --
>> -----
>> Pozdrawiam
>>
>> S?awek "sZiBis" Skowron
>>
>>



-- 
-----
Pozdrawiam

Sławek "sZiBis" Skowron
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: e: mon memory issue
  2012-09-05 18:44             ` Sławomir Skowron
@ 2012-09-10 10:29               ` Sławomir Skowron
  0 siblings, 0 replies; 9+ messages in thread
From: Sławomir Skowron @ 2012-09-10 10:29 UTC (permalink / raw)
  To: Sage Weil; +Cc: Xiaopong Tran, ceph-devel

I try with radosgw and it's reporting very nice output from valgrind,
but still nothing from mon.

desc: (none)
cmd: /usr/bin/ceph-mon -i 0 --pid-file /var/run/ceph/mon.0.pid -c
/etc/ceph/ceph.conf -f
time_unit: i
#-----------
snapshot=0
#-----------
time=0
mem_heap_B=0
mem_heap_extra_B=0
mem_stacks_B=0
heap_tree=empty

On Wed, Sep 5, 2012 at 8:44 PM, Sławomir Skowron <szibis@gmail.com> wrote:
> On Wed, Sep 5, 2012 at 5:51 PM, Sage Weil <sage@inktank.com> wrote:
>> On Wed, 5 Sep 2012, S?awomir Skowron wrote:
>>> Unfortunately here is the problem in my Ubuntu 12.04.1
>>>
>>> --9399-- You may be able to write your own handler.
>>> --9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
>>> --9399-- Nevertheless we consider this a bug.  Please report
>>> --9399-- it at http://valgrind.org/support/bug_reports.html.
>>> ==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
>>> ==9399==    This could cause spurious value errors to appear.
>>> ==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
>>> writing a proper wrapper.
>>> --9399-- WARNING: unhandled syscall: 306
>>> --9399-- You may be able to write your own handler.
>>> --9399-- Read the file README_MISSING_SYSCALL_OR_IOCTL.
>>> --9399-- Nevertheless we consider this a bug.  Please report
>>> --9399-- it at http://valgrind.org/support/bug_reports.html.
>>> ==9399== Warning: noted but unhandled ioctl 0x9408 with no size/direction hints
>>> ==9399==    This could cause spurious value errors to appear.
>>> ==9399==    See README_MISSING_SYSCALL_OR_IOCTL for guidance on
>>> writing a proper wrapper.
>>
>> These are harmless; it just doesn't recognize syncfs(2) or one of the
>> ioctls, but everything else works.
>>
>>> ^C2012-09-05 09:13:18.660048 a964700 -1 mon.0@0(leader) e4 *** Got
>>> Signal Interrupt ***
>>> ==9399==
>>
>> Did you hit control-c?
>>
>> If you leave it running it should gather the memory utilization info we
>> need...
>
> Yes it's running now, and  i will see tomorrow how much memory mon consumes.
>
>>
>> sage
>>
>>>
>>>
>>> On Wed, Sep 5, 2012 at 5:32 AM, Sage Weil <sage@inktank.com> wrote:
>>> > On Tue, 4 Sep 2012, S?awomir Skowron wrote:
>>> >> Valgrind returns nothing.
>>> >>
>>> >> valgrind --tool=massif --log-file=ceph_mon_valgrind ceph-mon -i 0 > log.txt
>>> >
>>> > The fork is probably confusing it.  I usually pass -f to ceph-mon (or
>>> > ceph-osd etc) to keep it in the foreground.  Can you give that a go?
>>> > e.g.,
>>> >
>>> >         valgrind --tool-massif ceph-mon -i 0 -f &
>>> >
>>> > and watch for the massif.out.$pid file.
>>> >
>>> > Thanks!
>>> > sage
>>> >
>>> >
>>> >>
>>> >> ==30491== Massif, a heap profiler
>>> >> ==30491== Copyright (C) 2003-2011, and GNU GPL'd, by Nicholas Nethercote
>>> >> ==30491== Using Valgrind-3.7.0 and LibVEX; rerun with -h for copyright info
>>> >> ==30491== Command: ceph-mon -i 0
>>> >> ==30491== Parent PID: 4013
>>> >> ==30491==
>>> >> ==30491==
>>> >>
>>> >> cat massif.out.26201
>>> >> desc: (none)
>>> >> cmd: ceph-mon -i 0
>>> >> time_unit: i
>>> >> #-----------
>>> >> snapshot=0
>>> >> #-----------
>>> >> time=0
>>> >> mem_heap_B=0
>>> >> mem_heap_extra_B=0
>>> >> mem_stacks_B=0
>>> >> heap_tree=empty
>>> >>
>>> >> What i have done wrong ??
>>> >>
>>> >> On Fri, Aug 31, 2012 at 8:34 PM, S?awomir Skowron <szibis@gmail.com> wrote:
>>> >> > I have this problem too. My mon's in 0.48.1 cluster have 10GB RAM
>>> >> > each, with 78 osd, and 2k request per minute (max) in radosgw.
>>> >> >
>>> >> > Now i have run one via valgrind. I will send output when mon grow up.
>>> >> >
>>> >> > On Fri, Aug 31, 2012 at 6:03 PM, Sage Weil <sage@inktank.com> wrote:
>>> >> >> On Fri, 31 Aug 2012, Xiaopong Tran wrote:
>>> >> >>
>>> >> >>> Hi,
>>> >> >>>
>>> >> >>> Is there any known memory issue with mon? We have 3 mons running, and
>>> >> >>> on keeps on crashing after 2 or 3 days, and I think it's because mon
>>> >> >>> sucks up all memory.
>>> >> >>>
>>> >> >>> Here's mon after starting for 10 minutes:
>>> >> >>>
>>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>> >> >>> 13700 root      20   0  163m  32m 3712 S   4.3  0.1   0:05.15 ceph-mon
>>> >> >>>  2595 root      20   0 1672m 523m    0 S   1.7  1.6 954:33.56 ceph-osd
>>> >> >>>  1941 root      20   0 1292m 220m    0 S   0.7  0.7 946:40.69 ceph-osd
>>> >> >>>  2316 root      20   0 1169m 198m    0 S   0.7  0.6 420:26.74 ceph-osd
>>> >> >>>  2395 root      20   0 1149m 184m    0 S   0.7  0.6 364:29.08 ceph-osd
>>> >> >>>  2487 root      20   0 1354m 373m    0 S   0.7  1.2 401:13.97 ceph-osd
>>> >> >>>   235 root      20   0     0    0    0 S   0.3  0.0   0:37.68 kworker/4:1
>>> >> >>>  1304 root      20   0     0    0    0 S   0.3  0.0   0:00.16 jbd2/sda3-8
>>> >> >>>  1327 root      20   0     0    0    0 S   0.3  0.0  13:07.00 xfsaild/sdf1
>>> >> >>>  2011 root      20   0 1240m 177m    0 S   0.3  0.6 411:52.91 ceph-osd
>>> >> >>>  2153 root      20   0 1095m 166m    0 S   0.3  0.5 370:56.01 ceph-osd
>>> >> >>>  2725 root      20   0 1214m 186m    0 S   0.3  0.6 378:16.59 ceph-osd
>>> >> >>>
>>> >> >>> Here's the memory situation of mon on another machine, after mon has
>>> >> >>> been running for 3 hours:
>>> >> >>>
>>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>> >> >>>  1716 root      20   0 1923m 1.6g 4028 S   7.6  5.2   8:45.82 ceph-mon
>>> >> >>>  1923 root      20   0  774m 138m 5052 S   0.7  0.4   1:28.56 ceph-osd
>>> >> >>>  2114 root      20   0  836m 143m 4864 S   0.7  0.4   1:20.14 ceph-osd
>>> >> >>>  2304 root      20   0  863m 176m 4988 S   0.7  0.5   1:13.30 ceph-osd
>>> >> >>>  2578 root      20   0  823m 150m 5056 S   0.7  0.5   1:24.55 ceph-osd
>>> >> >>>  2781 root      20   0  819m 131m 4900 S   0.7  0.4   1:12.14 ceph-osd
>>> >> >>>  2995 root      20   0  863m 179m 5024 S   0.7  0.6   1:41.96 ceph-osd
>>> >> >>>  3474 root      20   0  888m 208m 5608 S   0.7  0.6   7:08.08 ceph-osd
>>> >> >>>  1228 root      20   0     0    0    0 S   0.3  0.0   0:07.01 jbd2/sda3-8
>>> >> >>>  1853 root      20   0  859m 176m 4820 S   0.3  0.5   1:17.01 ceph-osd
>>> >> >>>  3373 root      20   0  789m 118m 4916 S   0.3  0.4   1:06.26 ceph-osd
>>> >> >>>
>>> >> >>> And here is the situation on a third node, mon has been running
>>> >> >>> for over a week:
>>> >> >>>
>>> >> >>>   PID USER      PR  NI  VIRT  RES  SHR S  %CPU %MEM    TIME+  COMMAND
>>> >> >>>  1717 root      20   0 68.8g  26g 2044 S  91.5 84.1   9220:40 ceph-mon
>>> >> >>>  1986 root      20   0 1281m 226m    0 S   1.7  0.7   1225:28 ceph-osd
>>> >> >>>  2196 root      20   0 1501m 538m    0 S   1.0  1.7   1221:54 ceph-osd
>>> >> >>>  2266 root      20   0 1121m 176m    0 S   0.7  0.5 399:23.70 ceph-osd
>>> >> >>>  2056 root      20   0 1072m 167m    0 S   0.3  0.5 403:49.76 ceph-osd
>>> >> >>>  2126 root      20   0 1412m 458m    0 S   0.3  1.4   1215:48 ceph-osd
>>> >> >>>  2337 root      20   0 1128m 188m    0 S   0.3  0.6 408:31.88 ceph-osd
>>> >> >>>
>>> >> >>> So, after a while, sooner or later, mon is going to crash, just
>>> >> >>> a matter of time.
>>> >> >>>
>>> >> >>> Does anyone see anything like this? This is kinda scary.
>>> >> >>>
>>> >> >>> OS: Debian Wheezy 3.2.0-3-amd64
>>> >> >>> Ceph: 0.48argonaut (commit:c2b20ca74249892c8e5e40c12aa14446a2bf2030)
>>> >> >>
>>> >> >> Can you try with 0.48.1argonaut?
>>> >> >>
>>> >> >> If it still happens, can you run ceph-mon through massif?
>>> >> >>
>>> >> >>  valgrind --tool=massif ceph-mon -i whatever
>>> >> >>
>>> >> >> That'll generate a massif.out file (make sure it's there; you may need to
>>> >> >> specify the output file for valgrind) over time.  Once ceph-mon starts
>>> >> >> eating ram, send us a copy of the file and we can hopefully see what is
>>> >> >> leaking.
>>> >> >>
>>> >> >> Thanks!
>>> >> >> sage
>>> >> >>
>>> >> >>
>>> >> >>>
>>> >> >>> With this issue on hand, I'll have to monitor it closely and
>>> >> >>> restart mon once in a while, or I will get a crash (which is
>>> >> >>> still good enough), or a system that does not respond at
>>> >> >>> all because memory is exhausted, and the whole ceph cluster
>>> >> >>> is unreachable. We had this problem in the morning, mon on one
>>> >> >>> node exhausted the memory, none of the ceph command responds
>>> >> >>> anymore, the only thing left to do is to hard reset the node.
>>> >> >>> The whole cluster was basically done at that time.
>>> >> >>>
>>> >> >>> Here is our usage situation:
>>> >> >>>
>>> >> >>> 1) A few applications which read and write data through
>>> >> >>> librados API, we have about 20-30 connections at any one time.
>>> >> >>> So far, our apps have no such memory issue, we have been
>>> >> >>> monitoring them closely.
>>> >> >>>
>>> >> >>> 2) We have a few scripts which pull data from an old storage
>>> >> >>> system, and use the rados command to put it into ceph.
>>> >> >>> Basically, just shell script. Each rados command is run
>>> >> >>> to write one object (one file), and exit. We run about
>>> >> >>> 25 scripts simultaneously, which means at any one time,
>>> >> >>> there are at most 25 connections.
>>> >> >>>
>>> >> >>> I don't think this is a very busy system. But this
>>> >> >>> memory issue is definitely a problem for us.
>>> >> >>>
>>> >> >>> Thanks for helping.
>>> >> >>>
>>> >> >>> Xiaopong
>>> >> >>> --
>>> >> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> >> >>> the body of a message to majordomo@vger.kernel.org
>>> >> >>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> >> >>>
>>> >> >>>
>>> >> >> --
>>> >> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> >> >> the body of a message to majordomo@vger.kernel.org
>>> >> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> >> >
>>> >> >
>>> >> >
>>> >> > --
>>> >> > -----
>>> >> > Pozdrawiam
>>> >> >
>>> >> > S?awek "sZiBis" Skowron
>>> >>
>>> >>
>>> >>
>>> >> --
>>> >> -----
>>> >> Pozdrawiam
>>> >>
>>> >> S?awek "sZiBis" Skowron
>>> >> --
>>> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>>> >> the body of a message to majordomo@vger.kernel.org
>>> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>> >>
>>> >>
>>>
>>>
>>>
>>> --
>>> -----
>>> Pozdrawiam
>>>
>>> S?awek "sZiBis" Skowron
>>>
>>>
>
>
>
> --
> -----
> Pozdrawiam
>
> Sławek "sZiBis" Skowron



-- 
-----
Pozdrawiam

Sławek "sZiBis" Skowron
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 9+ messages in thread

end of thread, other threads:[~2012-09-10 10:29 UTC | newest]

Thread overview: 9+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2012-08-31  9:02 mon memory issue Xiaopong Tran
2012-08-31 16:03 ` e: " Sage Weil
2012-08-31 18:34   ` Sławomir Skowron
2012-09-04 12:00     ` Sławomir Skowron
2012-09-05  3:32       ` Sage Weil
2012-09-05 10:32         ` Sławomir Skowron
2012-09-05 15:51           ` Sage Weil
2012-09-05 18:44             ` Sławomir Skowron
2012-09-10 10:29               ` Sławomir Skowron

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.