All of lore.kernel.org
 help / color / mirror / Atom feed
* KVM/RBD Block device hangs
@ 2012-05-21 14:44 Stefan Priebe - Profihost AG
  2012-05-21 14:47 ` Wido den Hollander
  0 siblings, 1 reply; 7+ messages in thread
From: Stefan Priebe - Profihost AG @ 2012-05-21 14:44 UTC (permalink / raw)
  To: ceph-devel

Hi List,

i'm testing the ceph / rbd block device with qemu and KVM. The creation
of a new disk went fine and i can also access the disk from the KVM client.

But after a short time while installing debian squeeze the osd seems to
crash and i see /core files on my machines.

What's next / wrong?

Greets
Stefan

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: KVM/RBD Block device hangs
  2012-05-21 14:44 KVM/RBD Block device hangs Stefan Priebe - Profihost AG
@ 2012-05-21 14:47 ` Wido den Hollander
  2012-05-21 14:50   ` Stefan Priebe - Profihost AG
  0 siblings, 1 reply; 7+ messages in thread
From: Wido den Hollander @ 2012-05-21 14:47 UTC (permalink / raw)
  To: Stefan Priebe - Profihost AG; +Cc: ceph-devel

Hi,

On 05/21/2012 04:44 PM, Stefan Priebe - Profihost AG wrote:
> Hi List,
>
> i'm testing the ceph / rbd block device with qemu and KVM. The creation
> of a new disk went fine and i can also access the disk from the KVM client.
>
> But after a short time while installing debian squeeze the osd seems to
> crash and i see /core files on my machines.

What do you mean the OSD crashes? I assume you have multiple OSD's in 
your cluster and one of them crashes? Or do ALL of them do go down?

>
> What's next / wrong?

Could you fetch the last 100 line from the log file and also tell us 
which version you are running?

Also, the output of "ceph -s" might be useful.

Wido

>
> Greets
> Stefan
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: KVM/RBD Block device hangs
  2012-05-21 14:47 ` Wido den Hollander
@ 2012-05-21 14:50   ` Stefan Priebe - Profihost AG
  2012-05-21 14:59     ` Wido den Hollander
  0 siblings, 1 reply; 7+ messages in thread
From: Stefan Priebe - Profihost AG @ 2012-05-21 14:50 UTC (permalink / raw)
  To: Wido den Hollander; +Cc: ceph-devel

Am 21.05.2012 16:47, schrieb Wido den Hollander:
> What do you mean the OSD crashes? I assume you have multiple OSD's in
> your cluster and one of them crashes? Or do ALL of them do go down?
2 of them went down (1st then 10 minutes later the 2nd)

>> What's next / wrong?
> 
> Could you fetch the last 100 line from the log file and also tell us
> which version you are running?
i'm upgrading right now to v0.47 and will try to reproduce.

> Also, the output of "ceph -s" might be useful.
before or after the crash?

Stefan

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: KVM/RBD Block device hangs
  2012-05-21 14:50   ` Stefan Priebe - Profihost AG
@ 2012-05-21 14:59     ` Wido den Hollander
  2012-05-21 20:51       ` Stefan Priebe
  0 siblings, 1 reply; 7+ messages in thread
From: Wido den Hollander @ 2012-05-21 14:59 UTC (permalink / raw)
  To: Stefan Priebe - Profihost AG; +Cc: ceph-devel

On 05/21/2012 04:50 PM, Stefan Priebe - Profihost AG wrote:
> Am 21.05.2012 16:47, schrieb Wido den Hollander:
>> What do you mean the OSD crashes? I assume you have multiple OSD's in
>> your cluster and one of them crashes? Or do ALL of them do go down?
> 2 of them went down (1st then 10 minutes later the 2nd)
>
>>> What's next / wrong?
>>
>> Could you fetch the last 100 line from the log file and also tell us
>> which version you are running?
> i'm upgrading right now to v0.47 and will try to reproduce.
>
>> Also, the output of "ceph -s" might be useful.
> before or after the crash?

Probably after, but both is fine. I just want to know how your cluster 
is doing and what the PG states are.

Wido

>
> Stefan
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: KVM/RBD Block device hangs
  2012-05-21 14:59     ` Wido den Hollander
@ 2012-05-21 20:51       ` Stefan Priebe
  2012-05-21 20:57         ` Gregory Farnum
  0 siblings, 1 reply; 7+ messages in thread
From: Stefan Priebe @ 2012-05-21 20:51 UTC (permalink / raw)
  To: Wido den Hollander; +Cc: ceph-devel

Am 21.05.2012 16:59, schrieb Wido den Hollander:
> Probably after, but both is fine. I just want to know how your cluster
> is doing and what the PG states are.
will do so.

dmesg says at least:
[ 4975.850765] ceph-osd[6226]: segfault at 439d34000 ip 00007f0db39bfb23 
sp 00007f0da8c5e240 error 4 in libtcmalloc.so.0.0.0[7f0db399b000+3d000]

Stefan

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: KVM/RBD Block device hangs
  2012-05-21 20:51       ` Stefan Priebe
@ 2012-05-21 20:57         ` Gregory Farnum
  2012-05-22 12:47           ` Stefan Priebe - Profihost AG
  0 siblings, 1 reply; 7+ messages in thread
From: Gregory Farnum @ 2012-05-21 20:57 UTC (permalink / raw)
  To: Stefan Priebe; +Cc: Wido den Hollander, ceph-devel

On Mon, May 21, 2012 at 1:51 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
> Am 21.05.2012 16:59, schrieb Wido den Hollander:
>
>> Probably after, but both is fine. I just want to know how your cluster
>> is doing and what the PG states are.
>
> will do so.
>
> dmesg says at least:
> [ 4975.850765] ceph-osd[6226]: segfault at 439d34000 ip 00007f0db39bfb23 sp
> 00007f0da8c5e240 error 4 in libtcmalloc.so.0.0.0[7f0db399b000+3d000]

This is not an error we've seen before, and it's a little scary. Can you
1) run gdb and get a backtrace and a printout of each frame in the trace?
2) reproduce with high-level OSD debug logging enabled?

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: KVM/RBD Block device hangs
  2012-05-21 20:57         ` Gregory Farnum
@ 2012-05-22 12:47           ` Stefan Priebe - Profihost AG
  0 siblings, 0 replies; 7+ messages in thread
From: Stefan Priebe - Profihost AG @ 2012-05-22 12:47 UTC (permalink / raw)
  To: Gregory Farnum; +Cc: Wido den Hollander, ceph-devel

Am 21.05.2012 22:57, schrieb Gregory Farnum:
> On Mon, May 21, 2012 at 1:51 PM, Stefan Priebe <s.priebe@profihost.ag> wrote:
>> Am 21.05.2012 16:59, schrieb Wido den Hollander:
>>
>>> Probably after, but both is fine. I just want to know how your cluster
>>> is doing and what the PG states are.
>>
>> will do so.
>>
>> dmesg says at least:
>> [ 4975.850765] ceph-osd[6226]: segfault at 439d34000 ip 00007f0db39bfb23 sp
>> 00007f0da8c5e240 error 4 in libtcmalloc.so.0.0.0[7f0db399b000+3d000]
> 
> This is not an error we've seen before, and it's a little scary. Can you
> 1) run gdb and get a backtrace and a printout of each frame in the trace?
> 2) reproduce with high-level OSD debug logging enabled?

I haven't seen this problem again with v0.47 i will report back if i see
it again.

Stefan

^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2012-05-22 12:47 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2012-05-21 14:44 KVM/RBD Block device hangs Stefan Priebe - Profihost AG
2012-05-21 14:47 ` Wido den Hollander
2012-05-21 14:50   ` Stefan Priebe - Profihost AG
2012-05-21 14:59     ` Wido den Hollander
2012-05-21 20:51       ` Stefan Priebe
2012-05-21 20:57         ` Gregory Farnum
2012-05-22 12:47           ` Stefan Priebe - Profihost AG

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.