From: Mike Snitzer <snitzer@redhat.com> To: Keith Busch <keith.busch@intel.com> Cc: axboe@kernel.dk, hch@lst.de, emilne@redhat.com, james.smart@broadcom.com, hare@suse.de, Bart.VanAssche@wdc.com, linux-block@vger.kernel.org, linux-nvme@lists.infradead.org, dm-devel@redhat.com Subject: Re: [for-4.16 PATCH 0/5] block, nvme, dm: allow DM multipath to use NVMe's error handler Date: Tue, 26 Dec 2017 21:42:44 -0500 [thread overview] Message-ID: <20171227024244.GA24011@redhat.com> (raw) In-Reply-To: <20171226205109.GB13188@localhost.localdomain> On Tue, Dec 26 2017 at 3:51pm -0500, Keith Busch <keith.busch@intel.com> wrote: > On Tue, Dec 19, 2017 at 04:05:41PM -0500, Mike Snitzer wrote: > > These patches enable DM multipath to work well on NVMe over Fabrics > > devices. Currently that implies CONFIG_NVME_MULTIPATH is _not_ set. > > > > But follow-on work will be to make it so that native NVMe multipath > > and DM multipath can be made to co-exist (e.g. blacklisting certain > > NVMe devices from being consumed by native NVMe multipath?) > > Hi Mike, > > I've reviewed the series and I support with the goal. I'm not a big fan, > though, of having yet-another-field to set in bio and req on each IO. Yeah, I knew they'd be the primary sticking point for this patchset. I'm not loving the need to carry the function pointer around either. > Unless I'm missing something, I think we can make this simpler if you add > the new 'failover_req_fn' as an attribute of the struct request_queue > instead of threading it through bio and request. Native nvme multipath > can set the field directly in the nvme driver, and dm-mpath can set it > in each path when not using the nvme mpath. What do you think? I initially didn't like the gotchas associated [1], but I worked through them. I'll post v2 after some testing. Thanks, Mike [1]: With DM multipath, it is easy to reliably establish the function pointer. But clearing it on teardown is awkward.. because another DM multipath table may have already taken a reference on the device (as could happen when reloading the multipath table associated with the DM multipath device). You are left with a scenario where a new table load would set it, but teardown wouldn't easily know if it should be cleared. And not clearing it could easily lead to dereferencing stale memory (if/when DM multipath driver were unloaded yet NVMe request_queue outliving it).
WARNING: multiple messages have this Message-ID (diff)
From: snitzer@redhat.com (Mike Snitzer) Subject: [for-4.16 PATCH 0/5] block, nvme, dm: allow DM multipath to use NVMe's error handler Date: Tue, 26 Dec 2017 21:42:44 -0500 [thread overview] Message-ID: <20171227024244.GA24011@redhat.com> (raw) In-Reply-To: <20171226205109.GB13188@localhost.localdomain> On Tue, Dec 26 2017 at 3:51pm -0500, Keith Busch <keith.busch@intel.com> wrote: > On Tue, Dec 19, 2017@04:05:41PM -0500, Mike Snitzer wrote: > > These patches enable DM multipath to work well on NVMe over Fabrics > > devices. Currently that implies CONFIG_NVME_MULTIPATH is _not_ set. > > > > But follow-on work will be to make it so that native NVMe multipath > > and DM multipath can be made to co-exist (e.g. blacklisting certain > > NVMe devices from being consumed by native NVMe multipath?) > > Hi Mike, > > I've reviewed the series and I support with the goal. I'm not a big fan, > though, of having yet-another-field to set in bio and req on each IO. Yeah, I knew they'd be the primary sticking point for this patchset. I'm not loving the need to carry the function pointer around either. > Unless I'm missing something, I think we can make this simpler if you add > the new 'failover_req_fn' as an attribute of the struct request_queue > instead of threading it through bio and request. Native nvme multipath > can set the field directly in the nvme driver, and dm-mpath can set it > in each path when not using the nvme mpath. What do you think? I initially didn't like the gotchas associated [1], but I worked through them. I'll post v2 after some testing. Thanks, Mike [1]: With DM multipath, it is easy to reliably establish the function pointer. But clearing it on teardown is awkward.. because another DM multipath table may have already taken a reference on the device (as could happen when reloading the multipath table associated with the DM multipath device). You are left with a scenario where a new table load would set it, but teardown wouldn't easily know if it should be cleared. And not clearing it could easily lead to dereferencing stale memory (if/when DM multipath driver were unloaded yet NVMe request_queue outliving it).
next prev parent reply other threads:[~2017-12-27 2:42 UTC|newest] Thread overview: 22+ messages / expand[flat|nested] mbox.gz Atom feed top 2017-12-19 21:05 [for-4.16 PATCH 0/5] block, nvme, dm: allow DM multipath to use NVMe's error handler Mike Snitzer 2017-12-19 21:05 ` Mike Snitzer 2017-12-19 21:05 ` [for-4.16 PATCH 1/5] block: establish request failover callback infrastructure Mike Snitzer 2017-12-19 21:05 ` Mike Snitzer 2017-12-19 21:05 ` [for-4.16 PATCH 2/5] nvme: use request's failover callback for multipath failover Mike Snitzer 2017-12-19 21:05 ` Mike Snitzer 2017-12-19 21:05 ` [for-4.16 PATCH 3/5] nvme: move nvme_req_needs_failover() from multipath to core Mike Snitzer 2017-12-19 21:05 ` Mike Snitzer 2017-12-19 21:05 ` [for-4.16 PATCH 4/5] dm mpath: use NVMe error handling to know when an error is retryable Mike Snitzer 2017-12-19 21:05 ` Mike Snitzer 2017-12-20 16:58 ` Mike Snitzer 2017-12-20 16:58 ` Mike Snitzer 2017-12-20 20:33 ` [dm-devel] " Sagi Grimberg 2017-12-20 20:33 ` Sagi Grimberg 2017-12-19 21:05 ` [for-4.16 PATCH 5/5] dm mpath: skip calls to end_io_bio if using NVMe bio-based and round-robin Mike Snitzer 2017-12-19 21:05 ` Mike Snitzer 2017-12-22 18:02 ` [for-4.16 PATCH 0/5] block, nvme, dm: allow DM multipath to use NVMe's error handler Mike Snitzer 2017-12-22 18:02 ` Mike Snitzer 2017-12-26 20:51 ` Keith Busch 2017-12-26 20:51 ` Keith Busch 2017-12-27 2:42 ` Mike Snitzer [this message] 2017-12-27 2:42 ` Mike Snitzer
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20171227024244.GA24011@redhat.com \ --to=snitzer@redhat.com \ --cc=Bart.VanAssche@wdc.com \ --cc=axboe@kernel.dk \ --cc=dm-devel@redhat.com \ --cc=emilne@redhat.com \ --cc=hare@suse.de \ --cc=hch@lst.de \ --cc=james.smart@broadcom.com \ --cc=keith.busch@intel.com \ --cc=linux-block@vger.kernel.org \ --cc=linux-nvme@lists.infradead.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.