From: Shai Malin <malin1024@gmail.com>
To: Sagi Grimberg <sagi@grimberg.me>
Cc: netdev@vger.kernel.org, linux-nvme@lists.infradead.org,
davem@davemloft.net, kuba@kernel.org, hch@lst.de, axboe@fb.com,
kbusch@kernel.org, Ariel Elior <aelior@marvell.com>,
Michal Kalderon <mkalderon@marvell.com>,
okulkarni@marvell.com, pkushwaha@marvell.com,
Dean Balandin <dbalandin@marvell.com>,
Shai Malin <smalin@marvell.com>
Subject: Re: [RFC PATCH v5 03/27] nvme-tcp-offload: Add device scan implementation
Date: Mon, 24 May 2021 23:14:41 +0300 [thread overview]
Message-ID: <CAKKgK4y77NNN7N81GOdTm=btirDCv0uLGqESjuhccZs1CB5opg@mail.gmail.com> (raw)
In-Reply-To: <e986c505-b81a-36ab-6366-35d5fbc193e1@grimberg.me>
On 5/22/21 1:22 AM, Sagi Grimberg wrote:
> On 5/19/21 4:13 AM, Shai Malin wrote:
> > From: Dean Balandin <dbalandin@marvell.com>
> >
> > As part of create_ctrl(), it scans the registered devices and calls
> > the claim_dev op on each of them, to find the first devices that matches
> > the connection params. Once the correct devices is found (claim_dev
> > returns true), we raise the refcnt of that device and return that device
> > as the device to be used for ctrl currently being created.
> >
> > Acked-by: Igor Russkikh <irusskikh@marvell.com>
> > Signed-off-by: Dean Balandin <dbalandin@marvell.com>
> > Signed-off-by: Prabhakar Kushwaha <pkushwaha@marvell.com>
> > Signed-off-by: Omkar Kulkarni <okulkarni@marvell.com>
> > Signed-off-by: Michal Kalderon <mkalderon@marvell.com>
> > Signed-off-by: Ariel Elior <aelior@marvell.com>
> > Signed-off-by: Shai Malin <smalin@marvell.com>
> > ---
> > drivers/nvme/host/tcp-offload.c | 94 +++++++++++++++++++++++++++++++++
> > 1 file changed, 94 insertions(+)
> >
> > diff --git a/drivers/nvme/host/tcp-offload.c b/drivers/nvme/host/tcp-offload.c
> > index 711232eba339..aa7cc239abf2 100644
> > --- a/drivers/nvme/host/tcp-offload.c
> > +++ b/drivers/nvme/host/tcp-offload.c
> > @@ -13,6 +13,11 @@
> > static LIST_HEAD(nvme_tcp_ofld_devices);
> > static DECLARE_RWSEM(nvme_tcp_ofld_devices_rwsem);
> >
> > +static inline struct nvme_tcp_ofld_ctrl *to_tcp_ofld_ctrl(struct nvme_ctrl *nctrl)
> > +{
> > + return container_of(nctrl, struct nvme_tcp_ofld_ctrl, nctrl);
> > +}
> > +
> > /**
> > * nvme_tcp_ofld_register_dev() - NVMeTCP Offload Library registration
> > * function.
> > @@ -98,6 +103,94 @@ void nvme_tcp_ofld_req_done(struct nvme_tcp_ofld_req *req,
> > /* Placeholder - complete request with/without error */
> > }
> >
> > +struct nvme_tcp_ofld_dev *
> > +nvme_tcp_ofld_lookup_dev(struct nvme_tcp_ofld_ctrl *ctrl)
> > +{
> > + struct nvme_tcp_ofld_dev *dev;
> > +
> > + down_read(&nvme_tcp_ofld_devices_rwsem);
> > + list_for_each_entry(dev, &nvme_tcp_ofld_devices, entry) {
> > + if (dev->ops->claim_dev(dev, &ctrl->conn_params)) {
> > + /* Increase driver refcnt */
> > + if (!try_module_get(dev->ops->module)) {
>
> This means that every controller will take a long-lived reference on the
> module? Why?
This is in order to create a per controller dependency between the
nvme-tcp-offload and the vendor driver which is currently used.
We believe that the vendor driver which offloads a controller should exist
as long as the controller exists.
>
> > + pr_err("try_module_get failed\n");
> > + dev = NULL;
> > + }
> > +
> > + goto out;
> > + }
> > + }
> > +
> > + dev = NULL;
> > +out:
> > + up_read(&nvme_tcp_ofld_devices_rwsem);
> > +
> > + return dev;
> > +}
> > +
> > +static int nvme_tcp_ofld_setup_ctrl(struct nvme_ctrl *nctrl, bool new)
> > +{
> > + /* Placeholder - validates inputs and creates admin and IO queues */
> > +
> > + return 0;
> > +}
> > +
> > +static struct nvme_ctrl *
> > +nvme_tcp_ofld_create_ctrl(struct device *ndev, struct nvmf_ctrl_options *opts)
> > +{
> > + struct nvme_tcp_ofld_ctrl *ctrl;
> > + struct nvme_tcp_ofld_dev *dev;
> > + struct nvme_ctrl *nctrl;
> > + int rc = 0;
> > +
> > + ctrl = kzalloc(sizeof(*ctrl), GFP_KERNEL);
> > + if (!ctrl)
> > + return ERR_PTR(-ENOMEM);
> > +
> > + nctrl = &ctrl->nctrl;
> > +
> > + /* Init nvme_tcp_ofld_ctrl and nvme_ctrl params based on received opts */
> > +
> > + /* Find device that can reach the dest addr */
> > + dev = nvme_tcp_ofld_lookup_dev(ctrl);
> > + if (!dev) {
> > + pr_info("no device found for addr %s:%s.\n",
> > + opts->traddr, opts->trsvcid);
> > + rc = -EINVAL;
> > + goto out_free_ctrl;
> > + }
> > +
> > + ctrl->dev = dev;
> > +
> > + if (ctrl->dev->ops->max_hw_sectors)
> > + nctrl->max_hw_sectors = ctrl->dev->ops->max_hw_sectors;
> > + if (ctrl->dev->ops->max_segments)
> > + nctrl->max_segments = ctrl->dev->ops->max_segments;
> > +
> > + /* Init queues */
> > +
> > + /* Call nvme_init_ctrl */
> > +
> > + rc = ctrl->dev->ops->setup_ctrl(ctrl, true);
> > + if (rc)
> > + goto out_module_put;
>
> goto module_put without an explicit module_get is confusing.
We will modify the function to use explicit module_get so it will be clear.
>
> > +
> > + rc = nvme_tcp_ofld_setup_ctrl(nctrl, true);
> > + if (rc)
> > + goto out_uninit_ctrl;
>
> ops->setup_ctrl and then call to nvme_tcp_ofld_setup_ctrl?
> Looks weird, why are these separated?
We will move the vendor specific setup_ctrl() call into
nvme_tcp_ofld_setup_ctrl().
>
> > +
> > + return nctrl;
> > +
> > +out_uninit_ctrl:
> > + ctrl->dev->ops->release_ctrl(ctrl);
> > +out_module_put:
> > + module_put(dev->ops->module);
> > +out_free_ctrl:
> > + kfree(ctrl);
> > +
> > + return ERR_PTR(rc);
> > +}
> > +
> > static struct nvmf_transport_ops nvme_tcp_ofld_transport = {
> > .name = "tcp_offload",
> > .module = THIS_MODULE,
> > @@ -107,6 +200,7 @@ static struct nvmf_transport_ops nvme_tcp_ofld_transport = {
> > NVMF_OPT_RECONNECT_DELAY | NVMF_OPT_HDR_DIGEST |
> > NVMF_OPT_DATA_DIGEST | NVMF_OPT_NR_POLL_QUEUES |
> > NVMF_OPT_TOS,
> > + .create_ctrl = nvme_tcp_ofld_create_ctrl,
> > };
> >
> > static int __init nvme_tcp_ofld_init_module(void)
> >
next prev parent reply other threads:[~2021-05-24 20:14 UTC|newest]
Thread overview: 53+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-05-19 11:13 [RFC PATCH v5 00/27] NVMeTCP Offload ULP and QEDN Device Driver Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 01/27] nvme-tcp-offload: Add nvme-tcp-offload - NVMeTCP HW offload ULP Shai Malin
2021-05-21 17:06 ` Himanshu Madhani
2021-05-24 20:11 ` Shai Malin
2021-05-21 22:13 ` Sagi Grimberg
2021-05-24 20:08 ` Shai Malin
2021-06-08 9:28 ` Petr Mladek
2021-05-19 11:13 ` [RFC PATCH v5 02/27] nvme-fabrics: Move NVMF_ALLOWED_OPTS and NVMF_REQUIRED_OPTS definitions Shai Malin
2021-05-21 17:08 ` Himanshu Madhani
2021-05-21 22:15 ` Sagi Grimberg
2021-05-19 11:13 ` [RFC PATCH v5 03/27] nvme-tcp-offload: Add device scan implementation Shai Malin
2021-05-21 17:22 ` Himanshu Madhani
2021-05-21 22:22 ` Sagi Grimberg
2021-05-24 20:14 ` Shai Malin [this message]
2021-05-19 11:13 ` [RFC PATCH v5 04/27] nvme-tcp-offload: Add controller level implementation Shai Malin
2021-05-21 17:19 ` Himanshu Madhani
2021-05-21 22:31 ` Sagi Grimberg
2021-05-27 20:03 ` Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 05/27] nvme-tcp-offload: Add controller level error recovery implementation Shai Malin
2021-05-21 17:42 ` Himanshu Madhani
2021-05-21 22:34 ` Sagi Grimberg
2021-05-19 11:13 ` [RFC PATCH v5 06/27] nvme-tcp-offload: Add queue level implementation Shai Malin
2021-05-21 18:18 ` Himanshu Madhani
2021-05-21 22:48 ` Sagi Grimberg
2021-05-24 20:16 ` Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 07/27] nvme-tcp-offload: Add IO " Shai Malin
2021-05-21 18:26 ` Himanshu Madhani
2021-05-19 11:13 ` [RFC PATCH v5 08/27] nvme-tcp-offload: Add Timeout and ASYNC Support Shai Malin
2021-05-21 18:36 ` Himanshu Madhani
2021-05-21 22:51 ` Sagi Grimberg
2021-05-24 20:17 ` Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 09/27] qed: Add TCP_ULP FW resource layout Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 10/27] qed: Add NVMeTCP Offload PF Level FW and HW HSI Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 11/27] qed: Add NVMeTCP Offload Connection " Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 12/27] qed: Add support of HW filter block Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 13/27] qed: Add NVMeTCP Offload IO Level FW and HW HSI Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 14/27] qed: Add NVMeTCP Offload IO Level FW Initializations Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 15/27] qed: Add IP services APIs support Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 16/27] qedn: Add qedn - Marvell's NVMeTCP HW offload vendor driver Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 17/27] qedn: Add qedn probe Shai Malin
2021-05-19 12:31 ` Leon Romanovsky
2021-05-19 14:29 ` Shai Malin
2021-05-19 15:31 ` Leon Romanovsky
2021-05-19 11:13 ` [RFC PATCH v5 18/27] qedn: Add qedn_claim_dev API support Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 19/27] qedn: Add IRQ and fast-path resources initializations Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 20/27] qedn: Add connection-level slowpath functionality Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 21/27] qedn: Add support of configuring HW filter block Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 22/27] qedn: Add IO level qedn_send_req and fw_cq workqueue Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 23/27] qedn: Add support of Task and SGL Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 24/27] qedn: Add support of NVME ICReq & ICResp Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 25/27] qedn: Add IO level fastpath functionality Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 26/27] qedn: Add Connection and IO level recovery flows Shai Malin
2021-05-19 11:13 ` [RFC PATCH v5 27/27] qedn: Add support of ASYNC Shai Malin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='CAKKgK4y77NNN7N81GOdTm=btirDCv0uLGqESjuhccZs1CB5opg@mail.gmail.com' \
--to=malin1024@gmail.com \
--cc=aelior@marvell.com \
--cc=axboe@fb.com \
--cc=davem@davemloft.net \
--cc=dbalandin@marvell.com \
--cc=hch@lst.de \
--cc=kbusch@kernel.org \
--cc=kuba@kernel.org \
--cc=linux-nvme@lists.infradead.org \
--cc=mkalderon@marvell.com \
--cc=netdev@vger.kernel.org \
--cc=okulkarni@marvell.com \
--cc=pkushwaha@marvell.com \
--cc=sagi@grimberg.me \
--cc=smalin@marvell.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).