All of lore.kernel.org
 help / color / mirror / Atom feed
From: Willem Jan Withagen <wjw@digiware.nl>
To: kefu chai <tchaikov@gmail.com>
Cc: Ceph Development <ceph-devel@vger.kernel.org>
Subject: Re: Toying with a FreeBSD cluster results in a crash
Date: Sat, 8 Apr 2017 14:43:37 +0200	[thread overview]
Message-ID: <3fd7e04e-cf9d-5439-a932-33dcef685c88@digiware.nl> (raw)
In-Reply-To: <CAJE9aONW_n1k+_5by=JLs_7kkpka8xJCkcsPkM0YMchQottA_Q@mail.gmail.com>

On 08-04-17 05:33, kefu chai wrote:
> On Fri, Apr 7, 2017 at 10:34 PM, Willem Jan Withagen <wjw@digiware.nl> wrote:
>> Hi,
>>
>> I'm playing with my/a FreeBSD test cluster.
>> It is full with different types of disks, and sometimes they are not
>> very new.
>>
>> The deepscrub on it showed things like:
>>  filestore(/var/lib/ceph/osd/osd.7) error creating
>> #-1:4962ce63:::inc_osdmap.705:0#
>> (/var/lib/ceph/osd/osd.7/current/meta/inc\uosdmap
>> .705__0_C6734692__none) in index: (87) Attribute not found
> 
> filestore stores subdir states using xattr, could you check the xattr
> of your meta collection using something like:
> 
>  lsextattr user /var/lib/ceph/osd/osd.7/current/meta
> 
> if nothing shows up, did you enable the xattr on the mounted fs in
> which /var/lib/ceph/osd/osd.7/current/meta is located?

This is on ZFS, and there attributes are on per default.

I checked other parts of the OSD files and several other did have
attributes. But everything in this directory had nothing set.

Now the trick question is if it can recover from a crash like this:
>> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: 3360:
>> FAILED assert(0 == "Missing map in load_pgs")

If it depends on infomation in its local tree, things might be too
corrupt to restart it...
But if it needs to be fetched from the rest of the cluster, I had a
different type of problem?

--WjW
> 
>>
>>
>> I've build the cluster with:
>>         osd pool default size      = 1
>>
>> Created some pools, and then increased
>>         osd pool default size      = 3
>>
>> Restarted the pools, but 1 pool does not want to reboot, so now I wonder
>> if the restarting problem is due to issue like quoted above?
>>
>> And how do I cleanup this mess, without wiping the cluster and
>> restarting. :) Note that it is just practice for me doing somewhat more
>> tricky work.
>>
>> Thanx,
>> --WjW
>>
>>
>>     -6> 2017-04-07 16:04:57.530301 806e16000  0 osd.7 733 crush map has
>> features 2200130813952, adjusting msgr requires for clients
>>     -5> 2017-04-07 16:04:57.530314 806e16000  0 osd.7 733 crush map has
>> features 2200130813952 was 8705, adjusting msgr requires for mons
>>     -4> 2017-04-07 16:04:57.530321 806e16000  0 osd.7 733 crush map has
>> features 2200130813952, adjusting msgr requires for osds
>>     -3> 2017-04-07 16:04:57.552968 806e16000  0 osd.7 733 load_pgs
>>     -2> 2017-04-07 16:04:57.553479 806e16000 -1 osd.7 0 failed to load
>> OSD map for epoch 714, got 0 bytes
>>     -1> 2017-04-07 16:04:57.553493 806e16000 -1 osd.7 733 load_pgs: have
>> pgid 8.e9 at epoch 714, but missing map.  Crashing.
>>      0> 2017-04-07 16:04:57.554157 806e16000 -1
>> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: In function
>> 'void OSD::load_pgs()' thread 806e16000 time 2017-04-0
>> 7 16:04:57.553497
>> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: 3360: FAILED
>> assert(0 == "Missing map in load_pgs")
>>
>> Most of the pools are in "oke" state:
>> [/var/log/ceph] wjw@cephtest> ceph -s
>>     cluster 746e196d-e344-11e6-b4b7-0025903744dc
>>      health HEALTH_ERR
>>             45 pgs are stuck inactive for more than 300 seconds
>>             7 pgs down
>>             38 pgs stale
>>             7 pgs stuck inactive
>>             38 pgs stuck stale
>>             7 pgs stuck unclean
>>             pool cephfsdata has many more objects per pg than average
>> (too few pgs?)
>>      monmap e5: 3 mons at
>> {a=192.168.10.70:6789/0,b=192.168.9.79:6789/0,c=192.168.8.79:6789/0}
>>             election epoch 114, quorum 0,1,2 c,b,a
>>       fsmap e755: 1/1/1 up {0=alpha=up:active}
>>         mgr active: admin
>>      osdmap e877: 8 osds: 7 up, 7 in; 6 remapped pgs
>>             flags sortbitwise,require_jewel_osds,require_kraken_osds
>>       pgmap v681735: 1864 pgs, 7 pools, 12416 MB data, 354 kobjects
>>             79963 MB used, 7837 GB / 7915 GB avail
>>                 1819 active+clean
>>                   38 stale+active+clean
>>                    6 down
>>                    1 down+remapped
>>
>> Just the ones that were only on the OSD that doesn't want to come up.
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
> 


  reply	other threads:[~2017-04-08 12:43 UTC|newest]

Thread overview: 5+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2017-04-07 14:34 Toying with a FreeBSD cluster results in a crash Willem Jan Withagen
2017-04-08  3:33 ` kefu chai
2017-04-08 12:43   ` Willem Jan Withagen [this message]
2017-04-10  8:12     ` kefu chai
2017-04-10 12:32       ` Willem Jan Withagen

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=3fd7e04e-cf9d-5439-a932-33dcef685c88@digiware.nl \
    --to=wjw@digiware.nl \
    --cc=ceph-devel@vger.kernel.org \
    --cc=tchaikov@gmail.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.