From mboxrd@z Thu Jan 1 00:00:00 1970 From: David Turner Subject: Re: Ceph cluster stability Date: Tue, 19 Feb 2019 12:26:41 -0500 Message-ID: References: Mime-Version: 1.0 Content-Type: multipart/mixed; boundary="===============8817465392738494558==" Return-path: In-Reply-To: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: ceph-users-bounces-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org Sender: "ceph-users" To: M Ranga Swami Reddy Cc: ceph-users , ceph-devel List-Id: ceph-devel.vger.kernel.org --===============8817465392738494558== Content-Type: multipart/alternative; boundary="000000000000fe65f70582428b35" --000000000000fe65f70582428b35 Content-Type: text/plain; charset="UTF-8" With a RACK failure domain, you should be able to have an entire rack powered down without noticing any major impact on the clients. I regularly take down OSDs and nodes for maintenance and upgrades without seeing any problems with client IO. On Tue, Feb 12, 2019 at 5:01 AM M Ranga Swami Reddy wrote: > Hello - I have a couple of questions on ceph cluster stability, even > we follow all recommendations as below: > - Having separate replication n/w and data n/w > - RACK is the failure domain > - Using SSDs for journals (1:4ratio) > > Q1 - If one OSD down, cluster IO down drastically and customer Apps > impacted. > Q2 - what is stability ratio, like with above, is ceph cluster > workable condition, if one osd down or one node down,etc. > > Thanks > Swami > _______________________________________________ > ceph-users mailing list > ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > --000000000000fe65f70582428b35 Content-Type: text/html; charset="UTF-8" Content-Transfer-Encoding: quoted-printable
With a RACK failure domain, you should be able to have an = entire rack powered down without noticing any major impact on the clients.= =C2=A0 I regularly take down OSDs and nodes for maintenance and upgrades wi= thout seeing any problems with client IO.

On Tue, Feb 12, 2019 at 5:01 AM M = Ranga Swami Reddy <swamireddy@gm= ail.com> wrote:
Hello - I have a couple of questions on ceph cluster stability, even=
we follow all recommendations as below:
- Having separate replication n/w and data n/w
- RACK is the failure domain
- Using SSDs for journals (1:4ratio)

Q1 - If one OSD down, cluster IO down drastically and customer Apps impacte= d.
Q2 - what is stability ratio, like with above, is ceph cluster
workable condition, if one osd down or one node down,etc.

Thanks
Swami
_______________________________________________
ceph-users mailing list
ceph-users@l= ists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-= ceph.com
--000000000000fe65f70582428b35-- --===============8817465392738494558== Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Content-Disposition: inline _______________________________________________ ceph-users mailing list ceph-users-idqoXFIVOFJgJs9I8MT0rw@public.gmane.org http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com --===============8817465392738494558==--