All of lore.kernel.org
 help / color / mirror / Atom feed
* Iusse: Ceph osd rm one osd cause 30% objects degraded
@ 2014-11-19 11:29 Qiang
  2014-11-19 15:30 ` Issue: " Qiang
  0 siblings, 1 reply; 2+ messages in thread
From: Qiang @ 2014-11-19 11:29 UTC (permalink / raw)
  To: ceph-devel

Hi, Dear ceph-devel

I met a issue:  Ceph osd rm one osd cause 30% objects degraded.

Step 1:
#created a ssd root
ceph osd crush add-bucket ssd root

Step 2: Installed a osd.100 failed:

94	1			osd.94	up	1	
95	1			osd.95	up	1	
96	1			osd.96	up	1	
97	1			osd.97	up	1	
98	1			osd.98	up	1	
99	1			osd.99	up	1	
100	0	osd.100	down	0

Step 3: Installed a osd.101 again successfully.  Move a host into root=ssd.
-12	1	root ssd
-13	1		host ssd-cephnode1
101	1			osd.101	up	1	
-1	100	root default
-2	10		host cephnode1
0	1			osd.0	up	1	
1	1			osd.1	up	1	
2	1			osd.2	up	1

Step 4: Then I ceph osd rm 100, but the the ceph health turned into 30% 
objects degraded. Then the io performance downgrade to very slow (1MB/s 
each clients).

Anybody know what is the root cause? Or some suggestions to finger it out?

Thank you very much.

^ permalink raw reply	[flat|nested] 2+ messages in thread

* Re: Issue: Ceph osd rm one osd cause 30% objects degraded
  2014-11-19 11:29 Iusse: Ceph osd rm one osd cause 30% objects degraded Qiang
@ 2014-11-19 15:30 ` Qiang
  0 siblings, 0 replies; 2+ messages in thread
From: Qiang @ 2014-11-19 15:30 UTC (permalink / raw)
  To: ceph-devel


Add more information:

After step4, there are many "restarting backfill on osd.x" in ceph.log

2014-11-19 16:03:37.766787 mon.0 10.16.40.40:6789/0 2460367 : [INF] 
pgmap v9995708: 8192 pgs: 10 inactive, 15 peering, 8167 active+clean; 
21280 GB data, 63334 GB used, 209 TB / 270 TB avail; 174 kB/s wr, 26 op/s
2014-11-19 16:03:38.446557 osd.39 10.16.40.53:6802/38684 1310 : [INF] 
3.42a restarting backfill on osd.34 from (0'0,0'0] MAX to 1528'608742
2014-11-19 16:03:38.451568 osd.39 10.16.40.53:6802/38684 1311 : [INF] 
3.b0a restarting backfill on osd.72 from (0'0,0'0] MAX to 1528'837511
2014-11-19 16:03:38.481297 osd.39 10.16.40.53:6802/38684 1312 : [INF] 
3.375 restarting backfill on osd.22 from (0'0,0'0] MAX to 1529'103924
2014-11-19 16:03:38.484977 osd.39 10.16.40.53:6802/38684 1313 : [INF] 
3.b0a restarting backfill on osd.87 from (0'0,0'0] MAX to 1528'837511
2014-11-19 16:03:38.541612 osd.39 10.16.40.53:6802/38684 1314 : [INF] 
3.b54 restarting backfill on osd.80 from (0'0,0'0] MAX to 1529'598339

Then (28.190%) objects degraded
2014-11-19 16:07:40.324423 mon.1 10.16.40.41:6789/0 12 : [INF] mon.xx 
calling new monitor election
2014-11-19 16:07:51.003344 mon.0 10.16.40.40:6789/0 2460469 : [INF] 
pgmap v9995757: 8192 pgs: 4939 active+remapped+wait_backfill, 2 
active+remapped, 21 active+remapped+backfilling, 765 
active+recovery_wait, 2122 active+clean, 343 active+recovering; 21281 GB 
data, 64164 GB used, 208 TB / 270 TB avail; 4888 kB/s rd, 2120 kB/s wr, 
398 op/s; 6032032/21397704 objects degraded (28.190%); 2917 MB/s, 18 
objects/s recovering

Thanks very much.

On 2014年11月19日 19:29, Qiang wrote:
> Hi, Dear ceph-devel
>
> I met a issue:  Ceph osd rm one osd cause 30% objects degraded.
>
> Step 1:
> #created a ssd root
> ceph osd crush add-bucket ssd root
>
> Step 2: Installed a osd.100 failed:
>
> 94    1            osd.94    up    1
> 95    1            osd.95    up    1
> 96    1            osd.96    up    1
> 97    1            osd.97    up    1
> 98    1            osd.98    up    1
> 99    1            osd.99    up    1
> 100    0    osd.100    down    0
>
> Step 3: Installed a osd.101 again successfully.  Move a host into root=ssd.
> -12    1    root ssd
> -13    1        host ssd-cephnode1
> 101    1            osd.101    up    1
> -1    100    root default
> -2    10        host cephnode1
> 0    1            osd.0    up    1
> 1    1            osd.1    up    1
> 2    1            osd.2    up    1
>
> Step 4: Then I ceph osd rm 100, but the the ceph health turned into 30%
> objects degraded. Then the io performance downgrade to very slow (1MB/s
> each clients).
>
> Anybody know what is the root cause? Or some suggestions to finger it out?
>
> Thank you very much.
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2014-11-19 15:30 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2014-11-19 11:29 Iusse: Ceph osd rm one osd cause 30% objects degraded Qiang
2014-11-19 15:30 ` Issue: " Qiang

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.