All of lore.kernel.org
 help / color / mirror / Atom feed
From: Shai Malin <smalin@marvell.com>
To: "netdev@vger.kernel.org" <netdev@vger.kernel.org>,
	"linux-nvme@lists.infradead.org" <linux-nvme@lists.infradead.org>,
	"sagi@grimberg.me" <sagi@grimberg.me>, "hch@lst.de" <hch@lst.de>,
	"axboe@fb.com" <axboe@fb.com>,
	"kbusch@kernel.org" <kbusch@kernel.org>
Cc: "davem@davemloft.net" <davem@davemloft.net>,
	"kuba@kernel.org" <kuba@kernel.org>,
	"Erik.Smith@dell.com" <Erik.Smith@dell.com>,
	"Douglas.Farley@dell.com" <Douglas.Farley@dell.com>,
	Ariel Elior <aelior@marvell.com>,
	Michal Kalderon <mkalderon@marvell.com>,
	"Prabhakar Kushwaha" <pkushwaha@marvell.com>,
	Nikolay Assa <nassa@marvell.com>,
	"malin1024@gmail.com" <malin1024@gmail.com>,
	Shai Malin <smalin@marvell.com>
Subject: RE: [RFC PATCH v3 00/11] NVMeTCP Offload ULP and QEDN Device Driver
Date: Thu, 18 Feb 2021 18:38:07 +0000	[thread overview]
Message-ID: <PH0PR18MB3845E15A62826C9B5A520628CC859@PH0PR18MB3845.namprd18.prod.outlook.com> (raw)
In-Reply-To: <20210207181324.11429-1-smalin@marvell.com>

> 
> With the goal of enabling a generic infrastructure that allows NVMe/TCP
> offload devices like NICs to seamlessly plug into the NVMe-oF stack, this
> patch series introduces the nvme-tcp-offload ULP host layer, which will be a
> new transport type called "tcp-offload" and will serve as an abstraction layer
> to work with vendor specific nvme-tcp offload drivers.
> 
> NVMeTCP offload is a full offload of the NVMeTCP protocol, this includes
> both the TCP level and the NVMeTCP level.
> 
> The nvme-tcp-offload transport can co-exist with the existing tcp and other
> transports. The tcp offload was designed so that stack changes are kept to a
> bare minimum: only registering new transports.
> All other APIs, ops etc. are identical to the regular tcp transport.
> Representing the TCP offload as a new transport allows clear and
> manageable differentiation between the connections which should use the
> offload path and those that are not offloaded (even on the same device).
> 

Sagi, Christoph, Jens, Keith,
So, as there are no more comments / questions, we understand the direction 
is acceptable and will proceed to the full series.


WARNING: multiple messages have this Message-ID (diff)
From: Shai Malin <smalin@marvell.com>
To: "netdev@vger.kernel.org" <netdev@vger.kernel.org>,
	"linux-nvme@lists.infradead.org" <linux-nvme@lists.infradead.org>,
	"sagi@grimberg.me" <sagi@grimberg.me>, "hch@lst.de" <hch@lst.de>,
	"axboe@fb.com" <axboe@fb.com>,
	"kbusch@kernel.org" <kbusch@kernel.org>
Cc: Shai Malin <smalin@marvell.com>, Ariel Elior <aelior@marvell.com>,
	Michal Kalderon <mkalderon@marvell.com>,
	Nikolay Assa <nassa@marvell.com>,
	"malin1024@gmail.com" <malin1024@gmail.com>,
	"Douglas.Farley@dell.com" <Douglas.Farley@dell.com>,
	"Erik.Smith@dell.com" <Erik.Smith@dell.com>,
	"kuba@kernel.org" <kuba@kernel.org>,
	Prabhakar Kushwaha <pkushwaha@marvell.com>,
	"davem@davemloft.net" <davem@davemloft.net>
Subject: RE: [RFC PATCH v3 00/11] NVMeTCP Offload ULP and QEDN Device Driver
Date: Thu, 18 Feb 2021 18:38:07 +0000	[thread overview]
Message-ID: <PH0PR18MB3845E15A62826C9B5A520628CC859@PH0PR18MB3845.namprd18.prod.outlook.com> (raw)
In-Reply-To: <20210207181324.11429-1-smalin@marvell.com>

> 
> With the goal of enabling a generic infrastructure that allows NVMe/TCP
> offload devices like NICs to seamlessly plug into the NVMe-oF stack, this
> patch series introduces the nvme-tcp-offload ULP host layer, which will be a
> new transport type called "tcp-offload" and will serve as an abstraction layer
> to work with vendor specific nvme-tcp offload drivers.
> 
> NVMeTCP offload is a full offload of the NVMeTCP protocol, this includes
> both the TCP level and the NVMeTCP level.
> 
> The nvme-tcp-offload transport can co-exist with the existing tcp and other
> transports. The tcp offload was designed so that stack changes are kept to a
> bare minimum: only registering new transports.
> All other APIs, ops etc. are identical to the regular tcp transport.
> Representing the TCP offload as a new transport allows clear and
> manageable differentiation between the connections which should use the
> offload path and those that are not offloaded (even on the same device).
> 

Sagi, Christoph, Jens, Keith,
So, as there are no more comments / questions, we understand the direction 
is acceptable and will proceed to the full series.


_______________________________________________
Linux-nvme mailing list
Linux-nvme@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-nvme

  parent reply	other threads:[~2021-02-18 19:17 UTC|newest]

Thread overview: 38+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-02-07 18:13 [RFC PATCH v3 00/11] NVMeTCP Offload ULP and QEDN Device Driver Shai Malin
2021-02-07 18:13 ` Shai Malin
2021-02-07 18:13 ` [RFC PATCH v3 01/11] nvme-tcp-offload: Add nvme-tcp-offload - NVMeTCP HW offload ULP Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-07 22:57   ` kernel test robot
2021-02-07 18:13 ` [RFC PATCH v3 02/11] nvme-fabrics: Move NVMF_ALLOWED_OPTS and NVMF_REQUIRED_OPTS definitions Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-07 18:13 ` [RFC PATCH v3 03/11] nvme-tcp-offload: Add device scan implementation Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-08  0:53   ` kernel test robot
2021-02-07 18:13 ` [RFC PATCH v3 04/11] nvme-tcp-offload: Add controller level implementation Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-07 18:13 ` [RFC PATCH v3 05/11] nvme-tcp-offload: Add controller level error recovery implementation Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-07 18:13 ` [RFC PATCH v3 06/11] nvme-tcp-offload: Add queue level implementation Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-07 18:13 ` [RFC PATCH v3 07/11] nvme-tcp-offload: Add IO " Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-07 18:13 ` [RFC PATCH v3 08/11] nvme-qedn: Add qedn - Marvell's NVMeTCP HW offload vendor driver Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-07 18:13 ` [RFC PATCH v3 09/11] net-qed: Add NVMeTCP Offload PF Level FW and HW HSI Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-07 18:13 ` [RFC PATCH v3 10/11] nvme-qedn: Add qedn probe Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-08  2:37   ` kernel test robot
2021-02-08  2:52   ` kernel test robot
2021-02-07 18:13 ` [RFC PATCH v3 11/11] nvme-qedn: Add IRQ and fast-path resources initializations Shai Malin
2021-02-07 18:13   ` Shai Malin
2021-02-12 18:06 ` [RFC PATCH v3 00/11] NVMeTCP Offload ULP and QEDN Device Driver Chris Leech
2021-02-12 18:06   ` Chris Leech
2021-02-13 16:47   ` Shai Malin
2021-02-13 16:47     ` Shai Malin
2021-02-18 18:38 ` Shai Malin [this message]
2021-02-18 18:38   ` Shai Malin
2021-02-19  9:12   ` hch
2021-02-19  9:12     ` hch
2021-02-19 21:28     ` [EXT] " Ariel Elior
2021-02-19 21:28       ` Ariel Elior

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=PH0PR18MB3845E15A62826C9B5A520628CC859@PH0PR18MB3845.namprd18.prod.outlook.com \
    --to=smalin@marvell.com \
    --cc=Douglas.Farley@dell.com \
    --cc=Erik.Smith@dell.com \
    --cc=aelior@marvell.com \
    --cc=axboe@fb.com \
    --cc=davem@davemloft.net \
    --cc=hch@lst.de \
    --cc=kbusch@kernel.org \
    --cc=kuba@kernel.org \
    --cc=linux-nvme@lists.infradead.org \
    --cc=malin1024@gmail.com \
    --cc=mkalderon@marvell.com \
    --cc=nassa@marvell.com \
    --cc=netdev@vger.kernel.org \
    --cc=pkushwaha@marvell.com \
    --cc=sagi@grimberg.me \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.