From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Subject: Re: [LSF/MM TOPIC][LSF/MM ATTEND] OCSSDs - SMR, Hierarchical Interface, and Vector I/Os To: Slava Dubeyko , Damien Le Moal , Viacheslav Dubeyko , "lsf-pc@lists.linux-foundation.org" , Theodore Ts'o References: <05204e9d-ed4d-f97a-88f0-41b5e008af43@bjorling.me> <1483398761.2440.4.camel@dubeyko.com> <1483464921.2440.19.camel@dubeyko.com> <9319ce16-8355-3560-95b6-45e3f07220de@bjorling.me> From: =?UTF-8?Q?Matias_Bj=c3=b8rling?= Message-ID: <6ef654fa-292e-ed9d-b8b6-fe4282fb1ef1@bjorling.me> Date: Fri, 6 Jan 2017 14:05:34 +0100 MIME-Version: 1.0 In-Reply-To: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Linux FS Devel , "linux-block@vger.kernel.org" , "linux-nvme@lists.infradead.org" Content-Type: text/plain; charset="us-ascii" Sender: "Linux-nvme" Errors-To: linux-nvme-bounces+axboe=kernel.dk@lists.infradead.org List-ID: On 01/05/2017 11:58 PM, Slava Dubeyko wrote: > Next point is read disturbance. If BER of physical page/block achieves some threshold then > we need to move data from one page/block into another one. What subsystem will be > responsible for this activity? The drive-managed case expects that device's GC will manage > read disturbance issue. But what's about host-aware or host-managed case? If the host side > hasn't information about BER then the host's software is unable to manage this issue. Finally, > it sounds that we will have GC subsystem as on file system side as on device side. As a result, > it means possible unpredictable performance degradation and decreasing device lifetime. > Let's imagine that host-aware case could be unaware about read disturbance management. > But how host-managed case can manage this issue? The OCSSD interface uses a couple of methods: 1) Piggy back soft ECC errors onto the completion entry. Tells the host that a block properly should be refreshed when appropriate. 2) Use an asynchronous interface, e.g., NVMe get log page. Report blocks through this interface that has been read disturbed. This may be coupled with the various processes running on the SSD. 3) (That Ted suggested). Expose a "reset" bit in the Report Zones command to let the host know which blocks should be reset. If the plumbing for 2) is not available, or the information has been lost on the host side, this method can be used to "resync". > > Bad block management... So, drive-managed and host-aware cases should be completely unaware > about bad blocks. But what's about host-managed case? If a device will hide bad blocks from > the host then it means mapping table presence, access to logical pages/blocks and so on. If the host > hasn't access to the bad block management then it's not host-managed model. And it sounds as > completely unmanageable situation for the host-managed model. Because if the host has access > to bad block management (but how?) then we have really simple model. Otherwise, the host > has access to logical pages/blocks only and device should have internal GC. As a result, > it means possible unpredictable performance degradation and decreasing device lifetime because > of competition of GC on device side and GC on the host side. Agree. depending on the use-case, one may expose a "perfect" interface to the host, or one may expose an interface where media errors may be reported to the host. The former case are great for consumer units, where I/O predictability isn't critical, and similarly if I/O predictability is critical, the media errors can be reported, and the host may deal with them appropriately. _______________________________________________ Linux-nvme mailing list Linux-nvme@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-nvme From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-wm0-f46.google.com ([74.125.82.46]:36992 "EHLO mail-wm0-f46.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1033882AbdAFNGF (ORCPT ); Fri, 6 Jan 2017 08:06:05 -0500 Received: by mail-wm0-f46.google.com with SMTP id t79so28507241wmt.0 for ; Fri, 06 Jan 2017 05:05:36 -0800 (PST) Subject: Re: [LSF/MM TOPIC][LSF/MM ATTEND] OCSSDs - SMR, Hierarchical Interface, and Vector I/Os To: Slava Dubeyko , Damien Le Moal , Viacheslav Dubeyko , "lsf-pc@lists.linux-foundation.org" , Theodore Ts'o References: <05204e9d-ed4d-f97a-88f0-41b5e008af43@bjorling.me> <1483398761.2440.4.camel@dubeyko.com> <1483464921.2440.19.camel@dubeyko.com> <9319ce16-8355-3560-95b6-45e3f07220de@bjorling.me> Cc: Linux FS Devel , "linux-block@vger.kernel.org" , "linux-nvme@lists.infradead.org" From: =?UTF-8?Q?Matias_Bj=c3=b8rling?= Message-ID: <6ef654fa-292e-ed9d-b8b6-fe4282fb1ef1@bjorling.me> Date: Fri, 6 Jan 2017 14:05:34 +0100 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 8bit Sender: linux-fsdevel-owner@vger.kernel.org List-ID: On 01/05/2017 11:58 PM, Slava Dubeyko wrote: > Next point is read disturbance. If BER of physical page/block achieves some threshold then > we need to move data from one page/block into another one. What subsystem will be > responsible for this activity? The drive-managed case expects that device's GC will manage > read disturbance issue. But what's about host-aware or host-managed case? If the host side > hasn't information about BER then the host's software is unable to manage this issue. Finally, > it sounds that we will have GC subsystem as on file system side as on device side. As a result, > it means possible unpredictable performance degradation and decreasing device lifetime. > Let's imagine that host-aware case could be unaware about read disturbance management. > But how host-managed case can manage this issue? The OCSSD interface uses a couple of methods: 1) Piggy back soft ECC errors onto the completion entry. Tells the host that a block properly should be refreshed when appropriate. 2) Use an asynchronous interface, e.g., NVMe get log page. Report blocks through this interface that has been read disturbed. This may be coupled with the various processes running on the SSD. 3) (That Ted suggested). Expose a "reset" bit in the Report Zones command to let the host know which blocks should be reset. If the plumbing for 2) is not available, or the information has been lost on the host side, this method can be used to "resync". > > Bad block management... So, drive-managed and host-aware cases should be completely unaware > about bad blocks. But what's about host-managed case? If a device will hide bad blocks from > the host then it means mapping table presence, access to logical pages/blocks and so on. If the host > hasn't access to the bad block management then it's not host-managed model. And it sounds as > completely unmanageable situation for the host-managed model. Because if the host has access > to bad block management (but how?) then we have really simple model. Otherwise, the host > has access to logical pages/blocks only and device should have internal GC. As a result, > it means possible unpredictable performance degradation and decreasing device lifetime because > of competition of GC on device side and GC on the host side. Agree. depending on the use-case, one may expose a "perfect" interface to the host, or one may expose an interface where media errors may be reported to the host. The former case are great for consumer units, where I/O predictability isn't critical, and similarly if I/O predictability is critical, the media errors can be reported, and the host may deal with them appropriately. From mboxrd@z Thu Jan 1 00:00:00 1970 From: m@bjorling.me (=?UTF-8?Q?Matias_Bj=c3=b8rling?=) Date: Fri, 6 Jan 2017 14:05:34 +0100 Subject: [LSF/MM TOPIC][LSF/MM ATTEND] OCSSDs - SMR, Hierarchical Interface, and Vector I/Os In-Reply-To: References: <05204e9d-ed4d-f97a-88f0-41b5e008af43@bjorling.me> <1483398761.2440.4.camel@dubeyko.com> <1483464921.2440.19.camel@dubeyko.com> <9319ce16-8355-3560-95b6-45e3f07220de@bjorling.me> Message-ID: <6ef654fa-292e-ed9d-b8b6-fe4282fb1ef1@bjorling.me> On 01/05/2017 11:58 PM, Slava Dubeyko wrote: > Next point is read disturbance. If BER of physical page/block achieves some threshold then > we need to move data from one page/block into another one. What subsystem will be > responsible for this activity? The drive-managed case expects that device's GC will manage > read disturbance issue. But what's about host-aware or host-managed case? If the host side > hasn't information about BER then the host's software is unable to manage this issue. Finally, > it sounds that we will have GC subsystem as on file system side as on device side. As a result, > it means possible unpredictable performance degradation and decreasing device lifetime. > Let's imagine that host-aware case could be unaware about read disturbance management. > But how host-managed case can manage this issue? The OCSSD interface uses a couple of methods: 1) Piggy back soft ECC errors onto the completion entry. Tells the host that a block properly should be refreshed when appropriate. 2) Use an asynchronous interface, e.g., NVMe get log page. Report blocks through this interface that has been read disturbed. This may be coupled with the various processes running on the SSD. 3) (That Ted suggested). Expose a "reset" bit in the Report Zones command to let the host know which blocks should be reset. If the plumbing for 2) is not available, or the information has been lost on the host side, this method can be used to "resync". > > Bad block management... So, drive-managed and host-aware cases should be completely unaware > about bad blocks. But what's about host-managed case? If a device will hide bad blocks from > the host then it means mapping table presence, access to logical pages/blocks and so on. If the host > hasn't access to the bad block management then it's not host-managed model. And it sounds as > completely unmanageable situation for the host-managed model. Because if the host has access > to bad block management (but how?) then we have really simple model. Otherwise, the host > has access to logical pages/blocks only and device should have internal GC. As a result, > it means possible unpredictable performance degradation and decreasing device lifetime because > of competition of GC on device side and GC on the host side. Agree. depending on the use-case, one may expose a "perfect" interface to the host, or one may expose an interface where media errors may be reported to the host. The former case are great for consumer units, where I/O predictability isn't critical, and similarly if I/O predictability is critical, the media errors can be reported, and the host may deal with them appropriately.