From: Petr Mladek <pmladek@suse.com> To: Shai Malin <smalin@marvell.com> Cc: netdev@vger.kernel.org, linux-nvme@lists.infradead.org, davem@davemloft.net, kuba@kernel.org, sagi@grimberg.me, hch@lst.de, axboe@fb.com, kbusch@kernel.org, aelior@marvell.com, mkalderon@marvell.com, okulkarni@marvell.com, pkushwaha@marvell.com, malin1024@gmail.com, Dean Balandin <dbalandin@marvell.com> Subject: Re: [RFC PATCH v5 01/27] nvme-tcp-offload: Add nvme-tcp-offload - NVMeTCP HW offload ULP Date: Tue, 8 Jun 2021 11:28:53 +0200 [thread overview] Message-ID: <YL84VamVh78Ds2Eg@alley> (raw) In-Reply-To: <20210519111340.20613-2-smalin@marvell.com> On Wed 2021-05-19 14:13:14, Shai Malin wrote: > This patch will present the structure for the NVMeTCP offload common > layer driver. This module is added under "drivers/nvme/host/" and future > offload drivers which will register to it will be placed under > "drivers/nvme/hw". > This new driver will be enabled by the Kconfig "NVM Express over Fabrics > TCP offload commmon layer". > In order to support the new transport type, for host mode, no change is > needed. > > Each new vendor-specific offload driver will register to this ULP during > its probe function, by filling out the nvme_tcp_ofld_dev->ops and > nvme_tcp_ofld_dev->private_data and calling nvme_tcp_ofld_register_dev > with the initialized struct. > > The internal implementation: > - tcp-offload.h: > Includes all common structs and ops to be used and shared by offload > drivers. > > - tcp-offload.c: > Includes the init function which registers as a NVMf transport just > like any other transport. > > Acked-by: Igor Russkikh <irusskikh@marvell.com> > Signed-off-by: Dean Balandin <dbalandin@marvell.com> > Signed-off-by: Prabhakar Kushwaha <pkushwaha@marvell.com> > Signed-off-by: Omkar Kulkarni <okulkarni@marvell.com> > Signed-off-by: Michal Kalderon <mkalderon@marvell.com> > Signed-off-by: Ariel Elior <aelior@marvell.com> > Signed-off-by: Shai Malin <smalin@marvell.com> > Reviewed-by: Hannes Reinecke <hare@suse.de> > --- a/drivers/nvme/host/Kconfig > +++ b/drivers/nvme/host/Kconfig > @@ -84,3 +84,19 @@ config NVME_TCP > from https://github.com/linux-nvme/nvme-cli. > > If unsure, say N. > + > +config NVME_TCP_OFFLOAD > + tristate "NVM Express over Fabrics TCP offload common layer" > + default m Is this intentional, please? > + depends on INET > + depends on BLK_DEV_NVME > + select NVME_FABRICS > + help > + This provides support for the NVMe over Fabrics protocol using > + the TCP offload transport. This allows you to use remote block devices > + exported using the NVMe protocol set. > + > + To configure a NVMe over Fabrics controller use the nvme-cli tool > + from https://github.com/linux-nvme/nvme-cli. > + > + If unsure, say N. I would expect that the default would be "n" so that people that are not sure or do not care about NWMe just take the default. IMHO, it is the usual behavior. Best Regards, Petr
WARNING: multiple messages have this Message-ID (diff)
From: Petr Mladek <pmladek@suse.com> To: Shai Malin <smalin@marvell.com> Cc: netdev@vger.kernel.org, linux-nvme@lists.infradead.org, davem@davemloft.net, kuba@kernel.org, sagi@grimberg.me, hch@lst.de, axboe@fb.com, kbusch@kernel.org, aelior@marvell.com, mkalderon@marvell.com, okulkarni@marvell.com, pkushwaha@marvell.com, malin1024@gmail.com, Dean Balandin <dbalandin@marvell.com> Subject: Re: [RFC PATCH v5 01/27] nvme-tcp-offload: Add nvme-tcp-offload - NVMeTCP HW offload ULP Date: Tue, 8 Jun 2021 11:28:53 +0200 [thread overview] Message-ID: <YL84VamVh78Ds2Eg@alley> (raw) In-Reply-To: <20210519111340.20613-2-smalin@marvell.com> On Wed 2021-05-19 14:13:14, Shai Malin wrote: > This patch will present the structure for the NVMeTCP offload common > layer driver. This module is added under "drivers/nvme/host/" and future > offload drivers which will register to it will be placed under > "drivers/nvme/hw". > This new driver will be enabled by the Kconfig "NVM Express over Fabrics > TCP offload commmon layer". > In order to support the new transport type, for host mode, no change is > needed. > > Each new vendor-specific offload driver will register to this ULP during > its probe function, by filling out the nvme_tcp_ofld_dev->ops and > nvme_tcp_ofld_dev->private_data and calling nvme_tcp_ofld_register_dev > with the initialized struct. > > The internal implementation: > - tcp-offload.h: > Includes all common structs and ops to be used and shared by offload > drivers. > > - tcp-offload.c: > Includes the init function which registers as a NVMf transport just > like any other transport. > > Acked-by: Igor Russkikh <irusskikh@marvell.com> > Signed-off-by: Dean Balandin <dbalandin@marvell.com> > Signed-off-by: Prabhakar Kushwaha <pkushwaha@marvell.com> > Signed-off-by: Omkar Kulkarni <okulkarni@marvell.com> > Signed-off-by: Michal Kalderon <mkalderon@marvell.com> > Signed-off-by: Ariel Elior <aelior@marvell.com> > Signed-off-by: Shai Malin <smalin@marvell.com> > Reviewed-by: Hannes Reinecke <hare@suse.de> > --- a/drivers/nvme/host/Kconfig > +++ b/drivers/nvme/host/Kconfig > @@ -84,3 +84,19 @@ config NVME_TCP > from https://github.com/linux-nvme/nvme-cli. > > If unsure, say N. > + > +config NVME_TCP_OFFLOAD > + tristate "NVM Express over Fabrics TCP offload common layer" > + default m Is this intentional, please? > + depends on INET > + depends on BLK_DEV_NVME > + select NVME_FABRICS > + help > + This provides support for the NVMe over Fabrics protocol using > + the TCP offload transport. This allows you to use remote block devices > + exported using the NVMe protocol set. > + > + To configure a NVMe over Fabrics controller use the nvme-cli tool > + from https://github.com/linux-nvme/nvme-cli. > + > + If unsure, say N. I would expect that the default would be "n" so that people that are not sure or do not care about NWMe just take the default. IMHO, it is the usual behavior. Best Regards, Petr _______________________________________________ Linux-nvme mailing list Linux-nvme@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-nvme
next prev parent reply other threads:[~2021-06-08 9:28 UTC|newest] Thread overview: 106+ messages / expand[flat|nested] mbox.gz Atom feed top 2021-05-19 11:13 [RFC PATCH v5 00/27] NVMeTCP Offload ULP and QEDN Device Driver Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 01/27] nvme-tcp-offload: Add nvme-tcp-offload - NVMeTCP HW offload ULP Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-21 17:06 ` Himanshu Madhani 2021-05-21 17:06 ` Himanshu Madhani 2021-05-24 20:11 ` Shai Malin 2021-05-24 20:11 ` Shai Malin 2021-05-21 22:13 ` Sagi Grimberg 2021-05-21 22:13 ` Sagi Grimberg 2021-05-24 20:08 ` Shai Malin 2021-05-24 20:08 ` Shai Malin 2021-06-08 9:28 ` Petr Mladek [this message] 2021-06-08 9:28 ` Petr Mladek 2021-05-19 11:13 ` [RFC PATCH v5 02/27] nvme-fabrics: Move NVMF_ALLOWED_OPTS and NVMF_REQUIRED_OPTS definitions Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-21 17:08 ` Himanshu Madhani 2021-05-21 17:08 ` Himanshu Madhani 2021-05-21 22:15 ` Sagi Grimberg 2021-05-21 22:15 ` Sagi Grimberg 2021-05-19 11:13 ` [RFC PATCH v5 03/27] nvme-tcp-offload: Add device scan implementation Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-21 17:22 ` Himanshu Madhani 2021-05-21 17:22 ` Himanshu Madhani 2021-05-21 22:22 ` Sagi Grimberg 2021-05-21 22:22 ` Sagi Grimberg 2021-05-24 20:14 ` Shai Malin 2021-05-24 20:14 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 04/27] nvme-tcp-offload: Add controller level implementation Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-21 17:19 ` Himanshu Madhani 2021-05-21 17:19 ` Himanshu Madhani 2021-05-21 22:31 ` Sagi Grimberg 2021-05-21 22:31 ` Sagi Grimberg 2021-05-27 20:03 ` Shai Malin 2021-05-27 20:03 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 05/27] nvme-tcp-offload: Add controller level error recovery implementation Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-21 17:42 ` Himanshu Madhani 2021-05-21 17:42 ` Himanshu Madhani 2021-05-21 22:34 ` Sagi Grimberg 2021-05-21 22:34 ` Sagi Grimberg 2021-05-19 11:13 ` [RFC PATCH v5 06/27] nvme-tcp-offload: Add queue level implementation Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-21 18:18 ` Himanshu Madhani 2021-05-21 18:18 ` Himanshu Madhani 2021-05-21 22:48 ` Sagi Grimberg 2021-05-21 22:48 ` Sagi Grimberg 2021-05-24 20:16 ` Shai Malin 2021-05-24 20:16 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 07/27] nvme-tcp-offload: Add IO " Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-21 18:26 ` Himanshu Madhani 2021-05-21 18:26 ` Himanshu Madhani 2021-05-19 11:13 ` [RFC PATCH v5 08/27] nvme-tcp-offload: Add Timeout and ASYNC Support Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-21 18:36 ` Himanshu Madhani 2021-05-21 18:36 ` Himanshu Madhani 2021-05-21 22:51 ` Sagi Grimberg 2021-05-21 22:51 ` Sagi Grimberg 2021-05-24 20:17 ` Shai Malin 2021-05-24 20:17 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 09/27] qed: Add TCP_ULP FW resource layout Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 10/27] qed: Add NVMeTCP Offload PF Level FW and HW HSI Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 11/27] qed: Add NVMeTCP Offload Connection " Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 12/27] qed: Add support of HW filter block Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 13/27] qed: Add NVMeTCP Offload IO Level FW and HW HSI Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 14/27] qed: Add NVMeTCP Offload IO Level FW Initializations Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 15/27] qed: Add IP services APIs support Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 16/27] qedn: Add qedn - Marvell's NVMeTCP HW offload vendor driver Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 17/27] qedn: Add qedn probe Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 12:31 ` Leon Romanovsky 2021-05-19 12:31 ` Leon Romanovsky 2021-05-19 14:29 ` Shai Malin 2021-05-19 14:29 ` Shai Malin 2021-05-19 15:31 ` Leon Romanovsky 2021-05-19 15:31 ` Leon Romanovsky 2021-05-19 11:13 ` [RFC PATCH v5 18/27] qedn: Add qedn_claim_dev API support Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 19/27] qedn: Add IRQ and fast-path resources initializations Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 20/27] qedn: Add connection-level slowpath functionality Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 21/27] qedn: Add support of configuring HW filter block Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 22/27] qedn: Add IO level qedn_send_req and fw_cq workqueue Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 23/27] qedn: Add support of Task and SGL Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 24/27] qedn: Add support of NVME ICReq & ICResp Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 25/27] qedn: Add IO level fastpath functionality Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 26/27] qedn: Add Connection and IO level recovery flows Shai Malin 2021-05-19 11:13 ` Shai Malin 2021-05-19 11:13 ` [RFC PATCH v5 27/27] qedn: Add support of ASYNC Shai Malin 2021-05-19 11:13 ` Shai Malin
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=YL84VamVh78Ds2Eg@alley \ --to=pmladek@suse.com \ --cc=aelior@marvell.com \ --cc=axboe@fb.com \ --cc=davem@davemloft.net \ --cc=dbalandin@marvell.com \ --cc=hch@lst.de \ --cc=kbusch@kernel.org \ --cc=kuba@kernel.org \ --cc=linux-nvme@lists.infradead.org \ --cc=malin1024@gmail.com \ --cc=mkalderon@marvell.com \ --cc=netdev@vger.kernel.org \ --cc=okulkarni@marvell.com \ --cc=pkushwaha@marvell.com \ --cc=sagi@grimberg.me \ --cc=smalin@marvell.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.