All of lore.kernel.org
 help / color / mirror / Atom feed
* Toying with a FreeBSD cluster results in a crash
@ 2017-04-07 14:34 Willem Jan Withagen
  2017-04-08  3:33 ` kefu chai
  0 siblings, 1 reply; 5+ messages in thread
From: Willem Jan Withagen @ 2017-04-07 14:34 UTC (permalink / raw)
  To: Ceph Development

Hi,

I'm playing with my/a FreeBSD test cluster.
It is full with different types of disks, and sometimes they are not
very new.

The deepscrub on it showed things like:
 filestore(/var/lib/ceph/osd/osd.7) error creating
#-1:4962ce63:::inc_osdmap.705:0#
(/var/lib/ceph/osd/osd.7/current/meta/inc\uosdmap
.705__0_C6734692__none) in index: (87) Attribute not found


I've build the cluster with:
	osd pool default size      = 1

Created some pools, and then increased
	osd pool default size      = 3

Restarted the pools, but 1 pool does not want to reboot, so now I wonder
if the restarting problem is due to issue like quoted above?

And how do I cleanup this mess, without wiping the cluster and
restarting. :) Note that it is just practice for me doing somewhat more
tricky work.

Thanx,
--WjW


    -6> 2017-04-07 16:04:57.530301 806e16000  0 osd.7 733 crush map has
features 2200130813952, adjusting msgr requires for clients
    -5> 2017-04-07 16:04:57.530314 806e16000  0 osd.7 733 crush map has
features 2200130813952 was 8705, adjusting msgr requires for mons
    -4> 2017-04-07 16:04:57.530321 806e16000  0 osd.7 733 crush map has
features 2200130813952, adjusting msgr requires for osds
    -3> 2017-04-07 16:04:57.552968 806e16000  0 osd.7 733 load_pgs
    -2> 2017-04-07 16:04:57.553479 806e16000 -1 osd.7 0 failed to load
OSD map for epoch 714, got 0 bytes
    -1> 2017-04-07 16:04:57.553493 806e16000 -1 osd.7 733 load_pgs: have
pgid 8.e9 at epoch 714, but missing map.  Crashing.
     0> 2017-04-07 16:04:57.554157 806e16000 -1
/usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: In function
'void OSD::load_pgs()' thread 806e16000 time 2017-04-0
7 16:04:57.553497
/usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: 3360: FAILED
assert(0 == "Missing map in load_pgs")

Most of the pools are in "oke" state:
[/var/log/ceph] wjw@cephtest> ceph -s
    cluster 746e196d-e344-11e6-b4b7-0025903744dc
     health HEALTH_ERR
            45 pgs are stuck inactive for more than 300 seconds
            7 pgs down
            38 pgs stale
            7 pgs stuck inactive
            38 pgs stuck stale
            7 pgs stuck unclean
            pool cephfsdata has many more objects per pg than average
(too few pgs?)
     monmap e5: 3 mons at
{a=192.168.10.70:6789/0,b=192.168.9.79:6789/0,c=192.168.8.79:6789/0}
            election epoch 114, quorum 0,1,2 c,b,a
      fsmap e755: 1/1/1 up {0=alpha=up:active}
        mgr active: admin
     osdmap e877: 8 osds: 7 up, 7 in; 6 remapped pgs
            flags sortbitwise,require_jewel_osds,require_kraken_osds
      pgmap v681735: 1864 pgs, 7 pools, 12416 MB data, 354 kobjects
            79963 MB used, 7837 GB / 7915 GB avail
                1819 active+clean
                  38 stale+active+clean
                   6 down
                   1 down+remapped

Just the ones that were only on the OSD that doesn't want to come up.

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: Toying with a FreeBSD cluster results in a crash
  2017-04-07 14:34 Toying with a FreeBSD cluster results in a crash Willem Jan Withagen
@ 2017-04-08  3:33 ` kefu chai
  2017-04-08 12:43   ` Willem Jan Withagen
  0 siblings, 1 reply; 5+ messages in thread
From: kefu chai @ 2017-04-08  3:33 UTC (permalink / raw)
  To: Willem Jan Withagen; +Cc: Ceph Development

On Fri, Apr 7, 2017 at 10:34 PM, Willem Jan Withagen <wjw@digiware.nl> wrote:
> Hi,
>
> I'm playing with my/a FreeBSD test cluster.
> It is full with different types of disks, and sometimes they are not
> very new.
>
> The deepscrub on it showed things like:
>  filestore(/var/lib/ceph/osd/osd.7) error creating
> #-1:4962ce63:::inc_osdmap.705:0#
> (/var/lib/ceph/osd/osd.7/current/meta/inc\uosdmap
> .705__0_C6734692__none) in index: (87) Attribute not found

filestore stores subdir states using xattr, could you check the xattr
of your meta collection using something like:

 lsextattr user /var/lib/ceph/osd/osd.7/current/meta

if nothing shows up, did you enable the xattr on the mounted fs in
which /var/lib/ceph/osd/osd.7/current/meta is located?

>
>
> I've build the cluster with:
>         osd pool default size      = 1
>
> Created some pools, and then increased
>         osd pool default size      = 3
>
> Restarted the pools, but 1 pool does not want to reboot, so now I wonder
> if the restarting problem is due to issue like quoted above?
>
> And how do I cleanup this mess, without wiping the cluster and
> restarting. :) Note that it is just practice for me doing somewhat more
> tricky work.
>
> Thanx,
> --WjW
>
>
>     -6> 2017-04-07 16:04:57.530301 806e16000  0 osd.7 733 crush map has
> features 2200130813952, adjusting msgr requires for clients
>     -5> 2017-04-07 16:04:57.530314 806e16000  0 osd.7 733 crush map has
> features 2200130813952 was 8705, adjusting msgr requires for mons
>     -4> 2017-04-07 16:04:57.530321 806e16000  0 osd.7 733 crush map has
> features 2200130813952, adjusting msgr requires for osds
>     -3> 2017-04-07 16:04:57.552968 806e16000  0 osd.7 733 load_pgs
>     -2> 2017-04-07 16:04:57.553479 806e16000 -1 osd.7 0 failed to load
> OSD map for epoch 714, got 0 bytes
>     -1> 2017-04-07 16:04:57.553493 806e16000 -1 osd.7 733 load_pgs: have
> pgid 8.e9 at epoch 714, but missing map.  Crashing.
>      0> 2017-04-07 16:04:57.554157 806e16000 -1
> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: In function
> 'void OSD::load_pgs()' thread 806e16000 time 2017-04-0
> 7 16:04:57.553497
> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: 3360: FAILED
> assert(0 == "Missing map in load_pgs")
>
> Most of the pools are in "oke" state:
> [/var/log/ceph] wjw@cephtest> ceph -s
>     cluster 746e196d-e344-11e6-b4b7-0025903744dc
>      health HEALTH_ERR
>             45 pgs are stuck inactive for more than 300 seconds
>             7 pgs down
>             38 pgs stale
>             7 pgs stuck inactive
>             38 pgs stuck stale
>             7 pgs stuck unclean
>             pool cephfsdata has many more objects per pg than average
> (too few pgs?)
>      monmap e5: 3 mons at
> {a=192.168.10.70:6789/0,b=192.168.9.79:6789/0,c=192.168.8.79:6789/0}
>             election epoch 114, quorum 0,1,2 c,b,a
>       fsmap e755: 1/1/1 up {0=alpha=up:active}
>         mgr active: admin
>      osdmap e877: 8 osds: 7 up, 7 in; 6 remapped pgs
>             flags sortbitwise,require_jewel_osds,require_kraken_osds
>       pgmap v681735: 1864 pgs, 7 pools, 12416 MB data, 354 kobjects
>             79963 MB used, 7837 GB / 7915 GB avail
>                 1819 active+clean
>                   38 stale+active+clean
>                    6 down
>                    1 down+remapped
>
> Just the ones that were only on the OSD that doesn't want to come up.
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html



-- 
Regards
Kefu Chai

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: Toying with a FreeBSD cluster results in a crash
  2017-04-08  3:33 ` kefu chai
@ 2017-04-08 12:43   ` Willem Jan Withagen
  2017-04-10  8:12     ` kefu chai
  0 siblings, 1 reply; 5+ messages in thread
From: Willem Jan Withagen @ 2017-04-08 12:43 UTC (permalink / raw)
  To: kefu chai; +Cc: Ceph Development

On 08-04-17 05:33, kefu chai wrote:
> On Fri, Apr 7, 2017 at 10:34 PM, Willem Jan Withagen <wjw@digiware.nl> wrote:
>> Hi,
>>
>> I'm playing with my/a FreeBSD test cluster.
>> It is full with different types of disks, and sometimes they are not
>> very new.
>>
>> The deepscrub on it showed things like:
>>  filestore(/var/lib/ceph/osd/osd.7) error creating
>> #-1:4962ce63:::inc_osdmap.705:0#
>> (/var/lib/ceph/osd/osd.7/current/meta/inc\uosdmap
>> .705__0_C6734692__none) in index: (87) Attribute not found
> 
> filestore stores subdir states using xattr, could you check the xattr
> of your meta collection using something like:
> 
>  lsextattr user /var/lib/ceph/osd/osd.7/current/meta
> 
> if nothing shows up, did you enable the xattr on the mounted fs in
> which /var/lib/ceph/osd/osd.7/current/meta is located?

This is on ZFS, and there attributes are on per default.

I checked other parts of the OSD files and several other did have
attributes. But everything in this directory had nothing set.

Now the trick question is if it can recover from a crash like this:
>> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: 3360:
>> FAILED assert(0 == "Missing map in load_pgs")

If it depends on infomation in its local tree, things might be too
corrupt to restart it...
But if it needs to be fetched from the rest of the cluster, I had a
different type of problem?

--WjW
> 
>>
>>
>> I've build the cluster with:
>>         osd pool default size      = 1
>>
>> Created some pools, and then increased
>>         osd pool default size      = 3
>>
>> Restarted the pools, but 1 pool does not want to reboot, so now I wonder
>> if the restarting problem is due to issue like quoted above?
>>
>> And how do I cleanup this mess, without wiping the cluster and
>> restarting. :) Note that it is just practice for me doing somewhat more
>> tricky work.
>>
>> Thanx,
>> --WjW
>>
>>
>>     -6> 2017-04-07 16:04:57.530301 806e16000  0 osd.7 733 crush map has
>> features 2200130813952, adjusting msgr requires for clients
>>     -5> 2017-04-07 16:04:57.530314 806e16000  0 osd.7 733 crush map has
>> features 2200130813952 was 8705, adjusting msgr requires for mons
>>     -4> 2017-04-07 16:04:57.530321 806e16000  0 osd.7 733 crush map has
>> features 2200130813952, adjusting msgr requires for osds
>>     -3> 2017-04-07 16:04:57.552968 806e16000  0 osd.7 733 load_pgs
>>     -2> 2017-04-07 16:04:57.553479 806e16000 -1 osd.7 0 failed to load
>> OSD map for epoch 714, got 0 bytes
>>     -1> 2017-04-07 16:04:57.553493 806e16000 -1 osd.7 733 load_pgs: have
>> pgid 8.e9 at epoch 714, but missing map.  Crashing.
>>      0> 2017-04-07 16:04:57.554157 806e16000 -1
>> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: In function
>> 'void OSD::load_pgs()' thread 806e16000 time 2017-04-0
>> 7 16:04:57.553497
>> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: 3360: FAILED
>> assert(0 == "Missing map in load_pgs")
>>
>> Most of the pools are in "oke" state:
>> [/var/log/ceph] wjw@cephtest> ceph -s
>>     cluster 746e196d-e344-11e6-b4b7-0025903744dc
>>      health HEALTH_ERR
>>             45 pgs are stuck inactive for more than 300 seconds
>>             7 pgs down
>>             38 pgs stale
>>             7 pgs stuck inactive
>>             38 pgs stuck stale
>>             7 pgs stuck unclean
>>             pool cephfsdata has many more objects per pg than average
>> (too few pgs?)
>>      monmap e5: 3 mons at
>> {a=192.168.10.70:6789/0,b=192.168.9.79:6789/0,c=192.168.8.79:6789/0}
>>             election epoch 114, quorum 0,1,2 c,b,a
>>       fsmap e755: 1/1/1 up {0=alpha=up:active}
>>         mgr active: admin
>>      osdmap e877: 8 osds: 7 up, 7 in; 6 remapped pgs
>>             flags sortbitwise,require_jewel_osds,require_kraken_osds
>>       pgmap v681735: 1864 pgs, 7 pools, 12416 MB data, 354 kobjects
>>             79963 MB used, 7837 GB / 7915 GB avail
>>                 1819 active+clean
>>                   38 stale+active+clean
>>                    6 down
>>                    1 down+remapped
>>
>> Just the ones that were only on the OSD that doesn't want to come up.
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
> 


^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: Toying with a FreeBSD cluster results in a crash
  2017-04-08 12:43   ` Willem Jan Withagen
@ 2017-04-10  8:12     ` kefu chai
  2017-04-10 12:32       ` Willem Jan Withagen
  0 siblings, 1 reply; 5+ messages in thread
From: kefu chai @ 2017-04-10  8:12 UTC (permalink / raw)
  To: Willem Jan Withagen; +Cc: Ceph Development

On Sat, Apr 8, 2017 at 8:43 PM, Willem Jan Withagen <wjw@digiware.nl> wrote:
> On 08-04-17 05:33, kefu chai wrote:
>> On Fri, Apr 7, 2017 at 10:34 PM, Willem Jan Withagen <wjw@digiware.nl> wrote:
>>> Hi,
>>>
>>> I'm playing with my/a FreeBSD test cluster.
>>> It is full with different types of disks, and sometimes they are not
>>> very new.
>>>
>>> The deepscrub on it showed things like:
>>>  filestore(/var/lib/ceph/osd/osd.7) error creating
>>> #-1:4962ce63:::inc_osdmap.705:0#
>>> (/var/lib/ceph/osd/osd.7/current/meta/inc\uosdmap
>>> .705__0_C6734692__none) in index: (87) Attribute not found
>>
>> filestore stores subdir states using xattr, could you check the xattr
>> of your meta collection using something like:
>>
>>  lsextattr user /var/lib/ceph/osd/osd.7/current/meta
>>
>> if nothing shows up, did you enable the xattr on the mounted fs in
>> which /var/lib/ceph/osd/osd.7/current/meta is located?
>
> This is on ZFS, and there attributes are on per default.
>
> I checked other parts of the OSD files and several other did have
> attributes. But everything in this directory had nothing set.

that sounds weird. looks like a bug. can you reproduce it?

>
> Now the trick question is if it can recover from a crash like this:
>>> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: 3360:
>>> FAILED assert(0 == "Missing map in load_pgs")
>
> If it depends on infomation in its local tree, things might be too
> corrupt to restart it...

yes, the osdmap is stored in the meta collection, which in turn misses
the subdir states stored in the xattr.

> But if it needs to be fetched from the rest of the cluster, I had a
> different type of problem?

maybe you can rebuild that osd? the meta collection is different from
one osd to another, i am not sure if we can copy it over from anther
osd.

-- 
Regards
Kefu Chai

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: Toying with a FreeBSD cluster results in a crash
  2017-04-10  8:12     ` kefu chai
@ 2017-04-10 12:32       ` Willem Jan Withagen
  0 siblings, 0 replies; 5+ messages in thread
From: Willem Jan Withagen @ 2017-04-10 12:32 UTC (permalink / raw)
  To: kefu chai; +Cc: Ceph Development

On 10-4-2017 10:12, kefu chai wrote:
> On Sat, Apr 8, 2017 at 8:43 PM, Willem Jan Withagen <wjw@digiware.nl> wrote:
>> On 08-04-17 05:33, kefu chai wrote:
>>> On Fri, Apr 7, 2017 at 10:34 PM, Willem Jan Withagen <wjw@digiware.nl> wrote:
>>>> Hi,
>>>>
>>>> I'm playing with my/a FreeBSD test cluster.
>>>> It is full with different types of disks, and sometimes they are not
>>>> very new.
>>>>
>>>> The deepscrub on it showed things like:
>>>>  filestore(/var/lib/ceph/osd/osd.7) error creating
>>>> #-1:4962ce63:::inc_osdmap.705:0#
>>>> (/var/lib/ceph/osd/osd.7/current/meta/inc\uosdmap
>>>> .705__0_C6734692__none) in index: (87) Attribute not found
>>>
>>> filestore stores subdir states using xattr, could you check the xattr
>>> of your meta collection using something like:
>>>
>>>  lsextattr user /var/lib/ceph/osd/osd.7/current/meta
>>>
>>> if nothing shows up, did you enable the xattr on the mounted fs in
>>> which /var/lib/ceph/osd/osd.7/current/meta is located?
>>
>> This is on ZFS, and there attributes are on per default.
>>
>> I checked other parts of the OSD files and several other did have
>> attributes. But everything in this directory had nothing set.
> 
> that sounds weird. looks like a bug. can you reproduce it?

It is a long standing (> 3 months) test cluster that has been severely
abused by me learning Ceph things and testing ceph-fuse.

>> Now the trick question is if it can recover from a crash like this:
>>>> /usr/ports/net/ceph/work/ceph-wip.FreeBSD/src/osd/OSD.cc: 3360:
>>>> FAILED assert(0 == "Missing map in load_pgs")
>>
>> If it depends on infomation in its local tree, things might be too
>> corrupt to restart it...
> 
> yes, the osdmap is stored in the meta collection, which in turn misses
> the subdir states stored in the xattr.
> 
>> But if it needs to be fetched from the rest of the cluster, I had a
>> different type of problem?
> 
> maybe you can rebuild that osd? the meta collection is different from
> one osd to another, i am not sure if we can copy it over from anther
> osd.

Well trouble started when I wanted to grow from 1 replica to 3 replica.
(So what is lost on osd.7 is not available elsewhere.)

This I did to see how that this growing behaves, and then this osd.7
crashed. From what I could find in the logfiles it complained about the
missing maps or attributes.

I guess I'll just scrap the cluster, and start fresh, because this could
very well be a FreeBSD-ish problem in combination with me doing silly
things. And I do not have to keep the cluster around, because it has
grown to a size that eats all my diskspace on the test hardware.

--WjW




^ permalink raw reply	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2017-04-10 12:32 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2017-04-07 14:34 Toying with a FreeBSD cluster results in a crash Willem Jan Withagen
2017-04-08  3:33 ` kefu chai
2017-04-08 12:43   ` Willem Jan Withagen
2017-04-10  8:12     ` kefu chai
2017-04-10 12:32       ` Willem Jan Withagen

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.