From mboxrd@z Thu Jan 1 00:00:00 1970 From: Damien Churchill Subject: Re: pgs stuck inactive Date: Fri, 6 Apr 2012 17:55:04 +0100 Message-ID: References: Mime-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: QUOTED-PRINTABLE Return-path: Received: from mail-gy0-f174.google.com ([209.85.160.174]:60694 "EHLO mail-gy0-f174.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755703Ab2DFQzF convert rfc822-to-8bit (ORCPT ); Fri, 6 Apr 2012 12:55:05 -0400 Received: by ghrr11 with SMTP id r11so1324514ghr.19 for ; Fri, 06 Apr 2012 09:55:04 -0700 (PDT) In-Reply-To: Sender: ceph-devel-owner@vger.kernel.org List-ID: To: Samuel Just Cc: ceph-devel I've got that directory on 3 of the osds: 0, 3 and 4. Do you want the logs to all 3 of them? On 6 April 2012 17:50, Samuel Just wrote: > Is there a 0.138_head directory under current/ on any of your osds? > If so, can you post the log to that osd? =C2=A0I could also use the o= sd.0 > log. > -Sam > > On Wed, Apr 4, 2012 at 2:44 PM, Damien Churchill w= rote: >> I've uploaded them to: >> >> http://damoxc.net/ceph/osdmap >> http://damoxc.net/ceph/pg_dump >> >> Thanks >> >> On 4 April 2012 21:51, Samuel Just wrote: >>> Can you post a copy of your osd map and the output of 'ceph pg dump= ' ? >>> =C2=A0You can get the osdmap via 'ceph osd getmap -o '. >>> -Sam >>> >>> On Wed, Apr 4, 2012 at 1:12 AM, Damien Churchill = wrote: >>>> Hi, >>>> >>>> I'm having some trouble getting some pgs to stop being inactive. T= he >>>> cluster is running 0.44.1 and the kernel version is 3.2.x. >>>> >>>> ceph -s reports: >>>> 2012-04-04 09:08:57.816029 =C2=A0 =C2=A0pg v188540: 990 pgs: 223 i= nactive, 767 >>>> active+clean; 205 GB data, 1013 GB used, 8204 GB / 9315 GB avail >>>> 2012-04-04 09:08:57.817970 =C2=A0 mds e2198: 1/1/1 up {0=3Dnode24=3D= up:active}, >>>> 4 up:standby >>>> 2012-04-04 09:08:57.818024 =C2=A0 osd e5910: 5 osds: 5 up, 5 in >>>> 2012-04-04 09:08:57.818201 =C2=A0 log 2012-04-04 09:04:03.838358 o= sd.3 >>>> 172.22.10.24:6801/30000 159 : [INF] 0.13d scrub ok >>>> 2012-04-04 09:08:57.818280 =C2=A0 mon e7: 3 mons at >>>> {node21=3D172.22.10.21:6789/0,node22=3D172.22.10.22:6789/0,node23=3D= 172.22.10.23:6789/0} >>>> >>>> ceph health says: >>>> 2012-04-04 09:09:01.651053 mon <- [health] >>>> 2012-04-04 09:09:01.666585 mon.1 -> 'HEALTH_WARN 223 pgs stuck >>>> inactive; 223 pgs stuck unclean' (0) >>>> >>>> I was wondering if anyone has any suggestions about how to resolve >>>> this, or things to look for. I've tried restarted the ceph daemons= on >>>> the various nodes a few times to no-avail. I don't think that ther= e is >>>> anything wrong with any of the nodes either. >>>> >>>> Thanks in advance, >>>> Damien >>>> -- >>>> To unsubscribe from this list: send the line "unsubscribe ceph-dev= el" in >>>> the body of a message to majordomo@vger.kernel.org >>>> More majordomo info at =C2=A0http://vger.kernel.org/majordomo-info= =2Ehtml -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" i= n the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html