All of lore.kernel.org
 help / color / mirror / Atom feed
From: Sage Weil <sage@newdream.net>
To: "Łukasz Chrustek" <skidoo@tlen.pl>
Cc: ceph-devel@vger.kernel.org
Subject: Re: Problem with query and any operation on PGs
Date: Wed, 24 May 2017 14:47:23 +0000 (UTC)	[thread overview]
Message-ID: <alpine.DEB.2.11.1705241444150.3646@piezo.novalocal> (raw)
In-Reply-To: <379087365.20170524161815@tlen.pl>

[-- Attachment #1: Type: TEXT/PLAIN, Size: 9357 bytes --]

On Wed, 24 May 2017, Łukasz Chrustek wrote:
> Cześć,
> 
> > On Wed, 24 May 2017, Łukasz Chrustek wrote:
> >> Cześć,
> >> 
> >> > On Wed, 24 May 2017, Łukasz Chrustek wrote:
> >> >> Cześć,
> >> >> 
> >> >> > On Tue, 23 May 2017, Łukasz Chrustek wrote:
> >> >> >> Cześć,
> >> >> >> 
> >> >> >> > On Tue, 23 May 2017, Łukasz Chrustek wrote:
> >> >> >> >> I'm  not  sleeping for over 30 hours, and still can't find solution. I
> >> >> >> >> did,      as      You      wrote,     but     turning     off     this
> >> >> >> >> (https://pastebin.com/1npBXeMV) osds didn't resolve issue...
> >> >> >> 
> >> >> >> > The important bit is:
> >> >> >> 
> >> >> >> >             "blocked": "peering is blocked due to down osds",
> >> >> >> >             "down_osds_we_would_probe": [
> >> >> >> >                 6,
> >> >> >> >                 10,
> >> >> >> >                 33,
> >> >> >> >                 37,
> >> >> >> >                 72
> >> >> >> >             ],
> >> >> >> >             "peering_blocked_by": [
> >> >> >> >                 {
> >> >> >> >                     "osd": 6,
> >> >> >> >                     "current_lost_at": 0,
> >> >> >> >                     "comment": "starting or marking this osd lost may let
> >> >> >> > us proceed"
> >> >> >> >                 },
> >> >> >> >                 {
> >> >> >> >                     "osd": 10,
> >> >> >> >                     "current_lost_at": 0,
> >> >> >> >                     "comment": "starting or marking this osd lost may let
> >> >> >> > us proceed"
> >> >> >> >                 },
> >> >> >> >                 {
> >> >> >> >                     "osd": 37,
> >> >> >> >                     "current_lost_at": 0,
> >> >> >> >                     "comment": "starting or marking this osd lost may let
> >> >> >> > us proceed"
> >> >> >> >                 },
> >> >> >> >                 {
> >> >> >> >                     "osd": 72,
> >> >> >> >                     "current_lost_at": 113771,
> >> >> >> >                     "comment": "starting or marking this osd lost may let
> >> >> >> > us proceed"
> 
> > These are the osds (6, 10, 37, 72).
> 
> >> >> >> >                 }
> >> >> >> >             ]
> >> >> >> >         },
> >> >> >> 
> >> >> >> > Are any of those OSDs startable?
> 
> > This
> 
> osd 6 - isn't startable

Disk completely 100% dead, or just borken enough that ceph-osd won't 
start?  ceph-objectstore-tool can be used to extract a copy of the 2 pgs 
from this osd to recover any important writes on that osd.

> osd 10, 37, 72 are startable

With those started, I'd repeat the original sequence and get a fresh pg 
query to confirm that it still wants just osd.6.

use ceph-objectstore-tool to export the pg from osd.6, stop some other 
ranodm osd (not one of these ones), import the pg into that osd, and start 
again.  once it is up, 'ceph osd lost 6'.  the pg *should* peer at that 
point.  repeat with the same basic process with the other pg.

s


> 
> >> >> >> 
> >> >> >> They were all up and running - but I decided to shut them down and out
> >> >> >> them  from  ceph, now it looks like ceph working ok, but still two PGs
> >> >> >> are in down state, how to get rid of it ?
> >> >> 
> >> >> > If you haven't deleted the data, you should start the OSDs back up.
> 
> > This
> 
> By OSDs backup You mean copy /var/lib/ceph/osd/ceph-72/* to some other
> (non ceph) disk ?
> 
> >> >> 
> >> >> > If they are partially damanged you can use ceph-objectstore-tool to 
> >> >> > extract just the PGs in question to make sure you haven't lost anything,
> >> >> > inject them on some other OSD(s) and restart those, and *then* mark the
> >> >> > bad OSDs as 'lost'.
> >> >> 
> >> >> > If all else fails, you can just mark those OSDs 'lost', but in doing so
> >> >> > you might be telling the cluster to lose data.
> >> >> 
> >> >> > The best thing to do is definitely to get those OSDs started again.
> 
> > This
> 
> There were actions on this PGs, that make them destroy. I started this
> osds   (these  three,  which  are  startable)  -  this  dosn't  solved
> situation.  I  need to add, that on this cluster are other pools, only
> with pool with broken/down PGs is problem.
> >> >> 
> >> >> Now situation looks like this:
> >> >> 
> >> >> [root@cc1 ~]# rbd info volumes/volume-ccc5d976-cecf-4938-a452-1bee6188987b
> >> >> rbd image 'volume-ccc5d976-cecf-4938-a452-1bee6188987b':
> >> >>         size 500 GB in 128000 objects
> >> >>         order 22 (4096 kB objects)
> >> >>         block_name_prefix: rbd_data.ed9d394a851426
> >> >>         format: 2
> >> >>         features: layering
> >> >>         flags:
> >> >> 
> >> >> [root@cc1 ~]# rados -p volumes ls | grep rbd_data.ed9d394a851426
> >> >> (output cutted)
> >> >> rbd_data.ed9d394a851426.000000000000447c
> >> >> rbd_data.ed9d394a851426.0000000000010857
> >> >> rbd_data.ed9d394a851426.000000000000ec8b
> >> >> rbd_data.ed9d394a851426.000000000000fa43
> >> >> rbd_data.ed9d394a851426.000000000001ef2d
> >> >> ^C
> >> >> 
> >> >> it hangs on this object and isn't going further. rbd cp also hangs...
> >> >> rbd map - also...
> >> >> 
> >> >> can  You advice what can be solution for this case ?
> >> 
> >> > The hang is due to OSD throttling (see my first reply for how to wrok 
> >> > around that and get a pg query).  But you already did that and the cluster
> >> > told you which OSDs it needs to see up in order for it to peer and 
> >> > recover.  If you haven't destroyed those disks, you should start those
> 
> >> > osds and it shoudl be fine.  If you've destroyed the data or the disks are
> >> > truly broken and dead, then you can mark those OSDs lost and the cluster
> >> > *maybe* recover (but hard to say given the information you've shared).
> 
> > This
> 
> 
> [root@cc1 ~]# ceph osd lost 10 --yes-i-really-mean-it
> marked osd lost in epoch 115310
> [root@cc1 ~]# ceph osd lost 37 --yes-i-really-mean-it
> marked osd lost in epoch 115314
> [root@cc1 ~]# ceph osd lost 72 --yes-i-really-mean-it
> marked osd lost in epoch 115317
> [root@cc1 ~]# ceph -s
>     cluster 8cdfbff9-b7be-46de-85bd-9d49866fcf60
>      health HEALTH_WARN
>             2 pgs down
>             2 pgs peering
>             2 pgs stuck inactive
>      monmap e1: 3 mons at {cc1=192.168.128.1:6789/0,cc2=192.168.128.2:6789/0,cc3=192.168.128.3:6789/0}
>             election epoch 872, quorum 0,1,2 cc1,cc2,cc3
>      osdmap e115434: 100 osds: 89 up, 86 in; 1 remapped pgs
>       pgmap v67642483: 4032 pgs, 18 pools, 26713 GB data, 4857 kobjects
>             76718 GB used, 107 TB / 182 TB avail
>                 4030 active+clean
>                    1 down+remapped+peering
>                    1 down+peering
>   client io 14624 kB/s rd, 31619 kB/s wr, 382 op/s rd, 228 op/s wr
> [root@cc1 ~]# ceph -s
>     cluster 8cdfbff9-b7be-46de-85bd-9d49866fcf60
>      health HEALTH_WARN
>             2 pgs down
>             2 pgs peering
>             2 pgs stuck inactive
>      monmap e1: 3 mons at {cc1=192.168.128.1:6789/0,cc2=192.168.128.2:6789/0,cc3=192.168.128.3:6789/0}
>             election epoch 872, quorum 0,1,2 cc1,cc2,cc3
>      osdmap e115434: 100 osds: 89 up, 86 in; 1 remapped pgs
>       pgmap v67642485: 4032 pgs, 18 pools, 26713 GB data, 4857 kobjects
>             76718 GB used, 107 TB / 182 TB avail
>                 4030 active+clean
>                    1 down+remapped+peering
>                    1 down+peering
>   client io 17805 kB/s rd, 18787 kB/s wr, 215 op/s rd, 107 op/s wr
> 
> >> 
> >> > sage
> >> 
> >> What information I can bring to You to say it is recoverable ?
> >> 
> >> here are ceph -s and ceph health detail:
> >> 
> >> [root@cc1 ~]# ceph -s
> >>     cluster 8cdfbff9-b7be-46de-85bd-9d49866fcf60
> >>      health HEALTH_WARN
> >>             2 pgs down
> >>             2 pgs peering
> >>             2 pgs stuck inactive
> >>      monmap e1: 3 mons at {cc1=192.168.128.1:6789/0,cc2=192.168.128.2:6789/0,cc3=192.168.128.3:6789/0}
> >>             election epoch 872, quorum 0,1,2 cc1,cc2,cc3
> >>      osdmap e115431: 100 osds: 89 up, 86 in; 1 remapped pgs
> >>       pgmap v67641261: 4032 pgs, 18 pools, 26706 GB data, 4855 kobjects
> >>             76705 GB used, 107 TB / 182 TB avail
> >>                 4030 active+clean
> >>                    1 down+remapped+peering
> >>                    1 down+peering
> >>   client io 5704 kB/s rd, 24685 kB/s wr, 49 op/s rd, 165 op/s wr
> >> [root@cc1 ~]# ceph health detail
> >> HEALTH_WARN 2 pgs down; 2 pgs peering; 2 pgs stuck inactive
> >> pg 1.165 is stuck inactive since forever, current state down+peering, last acting [67,88,48]
> >> pg 1.60 is stuck inactive since forever, current state down+remapped+peering, last acting [66,40]
> >> pg 1.60 is down+remapped+peering, acting [66,40]
> >> pg 1.165 is down+peering, acting [67,88,48]
> >> [root@cc1 ~]#
> >> 
> >> -- 
> >> Regards,
> >>  Łukasz Chrustek
> >> 
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> the body of a message to majordomo@vger.kernel.org
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >> 
> >> 
> 
> 
> 
> -- 
> Pozdrowienia,
>  Łukasz Chrustek
> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 

  reply	other threads:[~2017-05-24 14:47 UTC|newest]

Thread overview: 35+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <175484591.20170523135449@tlen.pl>
2017-05-23 12:48 ` Problem with query and any operation on PGs Łukasz Chrustek
2017-05-23 14:17   ` Sage Weil
2017-05-23 14:43     ` Łukasz Chrustek
     [not found]     ` <1464688590.20170523185052@tlen.pl>
2017-05-23 17:40       ` Sage Weil
2017-05-23 21:43         ` Łukasz Chrustek
2017-05-23 21:48           ` Sage Weil
2017-05-24 13:19             ` Łukasz Chrustek
2017-05-24 13:37               ` Sage Weil
2017-05-24 13:58                 ` Łukasz Chrustek
2017-05-24 14:02                   ` Sage Weil
2017-05-24 14:18                     ` Łukasz Chrustek
2017-05-24 14:47                       ` Sage Weil [this message]
2017-05-24 15:00                         ` Łukasz Chrustek
2017-05-24 15:07                           ` Łukasz Chrustek
2017-05-24 15:11                           ` Sage Weil
2017-05-24 15:24                             ` Łukasz Chrustek
2017-05-24 15:54                             ` Łukasz Chrustek
2017-05-24 16:02                               ` Łukasz Chrustek
2017-05-24 17:07                                 ` Łukasz Chrustek
2017-05-24 17:16                                   ` Sage Weil
2017-05-24 17:28                                     ` Łukasz Chrustek
2017-05-24 18:16                                       ` Sage Weil
2017-05-24 19:47                                         ` Łukasz Chrustek
2017-05-24 17:30                                     ` Łukasz Chrustek
2017-05-24 17:35                                       ` Łukasz Chrustek
2017-05-24 21:38                         ` Łukasz Chrustek
2017-05-24 21:53                           ` Sage Weil
2017-05-24 22:09                             ` Łukasz Chrustek
2017-05-24 22:27                               ` Sage Weil
2017-05-24 22:46                                 ` Łukasz Chrustek
2017-05-25  2:06                                   ` Sage Weil
2017-05-25 11:22                                     ` Łukasz Chrustek
2017-05-29 15:31                                       ` Łukasz Chrustek
2017-05-30 13:21                                   ` Sage Weil
2017-06-10 22:45                                     ` Łukasz Chrustek

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=alpine.DEB.2.11.1705241444150.3646@piezo.novalocal \
    --to=sage@newdream.net \
    --cc=ceph-devel@vger.kernel.org \
    --cc=skidoo@tlen.pl \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.