From: Max Gurtovoy <maxg@mellanox.com> To: maxg@mellanox.com, linux-nvme@lists.infradead.org, kbusch@kernel.org, hch@lst.de, sagi@grimberg.me, martin.petersen@oracle.com, jsmart2021@gmail.com, axboe@kernel.dk Cc: vladimirk@mellanox.com, shlomin@mellanox.com, israelr@mellanox.com, idanb@mellanox.com, oren@mellanox.com, nitzanc@mellanox.com Subject: [PATCH 14/15] nvmet: add metadata support for block devices Date: Tue, 28 Apr 2020 16:11:34 +0300 Message-ID: <20200428131135.211521-15-maxg@mellanox.com> (raw) In-Reply-To: <20200428131135.211521-1-maxg@mellanox.com> From: Israel Rukshin <israelr@mellanox.com> Allocate the metadata SGL buffers and set metadata fields for the request. Then create a block IO request for the metadata from the protection SG list. Signed-off-by: Israel Rukshin <israelr@mellanox.com> Signed-off-by: Max Gurtovoy <maxg@mellanox.com> --- drivers/nvme/target/core.c | 90 +++++++++++++++++++++++++++++++-------- drivers/nvme/target/io-cmd-bdev.c | 85 +++++++++++++++++++++++++++++++++++- drivers/nvme/target/nvmet.h | 3 ++ 3 files changed, 159 insertions(+), 19 deletions(-) diff --git a/drivers/nvme/target/core.c b/drivers/nvme/target/core.c index a57ab0c..c1456ce 100644 --- a/drivers/nvme/target/core.c +++ b/drivers/nvme/target/core.c @@ -884,8 +884,11 @@ bool nvmet_req_init(struct nvmet_req *req, struct nvmet_cq *cq, req->sq = sq; req->ops = ops; req->sg = NULL; + req->md_sg = NULL; req->sg_cnt = 0; + req->md_sg_cnt = 0; req->transfer_len = 0; + req->md_len = 0; req->cqe->status = 0; req->cqe->sq_head = 0; req->ns = NULL; @@ -970,9 +973,67 @@ bool nvmet_check_data_len_lte(struct nvmet_req *req, size_t data_len) return true; } +void nvmet_req_free_p2pmem_sgls(struct nvmet_req *req) +{ + pci_p2pmem_free_sgl(req->p2p_dev, req->sg); + if (req->md_sg) + pci_p2pmem_free_sgl(req->p2p_dev, req->md_sg); +} + +static int nvmet_req_alloc_p2pmem_sgls(struct nvmet_req *req, int data_len, + struct pci_dev *p2p_dev) +{ + req->sg = pci_p2pmem_alloc_sgl(p2p_dev, &req->sg_cnt, data_len); + if (!req->sg) + goto out_err; + + if (req->md_len) { + req->md_sg = pci_p2pmem_alloc_sgl(p2p_dev, &req->md_sg_cnt, + req->md_len); + if (!req->md_sg) + goto out_free_sg; + } + req->p2p_dev = p2p_dev; + return 0; + +out_free_sg: + pci_p2pmem_free_sgl(p2p_dev, req->sg); +out_err: + return -ENOMEM; +} + +void nvmet_req_free_mem_sgls(struct nvmet_req *req) +{ + sgl_free(req->sg); + if (req->md_sg) + sgl_free(req->md_sg); +} + +static int nvmet_req_alloc_mem_sgls(struct nvmet_req *req, int data_len) +{ + req->sg = sgl_alloc(data_len, GFP_KERNEL, &req->sg_cnt); + if (unlikely(!req->sg)) + goto out; + + if (req->md_len) { + req->md_sg = sgl_alloc(req->md_len, GFP_KERNEL, + &req->md_sg_cnt); + if (unlikely(!req->md_sg)) + goto out_free; + } + + return 0; + +out_free: + sgl_free(req->sg); +out: + return -ENOMEM; +} + int nvmet_req_alloc_sgl(struct nvmet_req *req) { struct pci_dev *p2p_dev = NULL; + int data_len = req->transfer_len - req->md_len; if (IS_ENABLED(CONFIG_PCI_P2PDMA)) { if (req->sq->ctrl && req->ns) @@ -981,37 +1042,32 @@ int nvmet_req_alloc_sgl(struct nvmet_req *req) req->p2p_dev = NULL; if (req->sq->qid && p2p_dev) { - req->sg = pci_p2pmem_alloc_sgl(p2p_dev, &req->sg_cnt, - req->transfer_len); - if (req->sg) { - req->p2p_dev = p2p_dev; + int ret = nvmet_req_alloc_p2pmem_sgls(req, data_len, + p2p_dev); + if (!ret) return 0; - } } - - /* - * If no P2P memory was available we fallback to using - * regular memory - */ } - req->sg = sgl_alloc(req->transfer_len, GFP_KERNEL, &req->sg_cnt); - if (unlikely(!req->sg)) - return -ENOMEM; - - return 0; + /* + * If no P2P memory was available/enabled we fallback to using regular + * memory. + */ + return nvmet_req_alloc_mem_sgls(req, data_len); } EXPORT_SYMBOL_GPL(nvmet_req_alloc_sgl); void nvmet_req_free_sgl(struct nvmet_req *req) { if (req->p2p_dev) - pci_p2pmem_free_sgl(req->p2p_dev, req->sg); + nvmet_req_free_p2pmem_sgls(req); else - sgl_free(req->sg); + nvmet_req_free_mem_sgls(req); req->sg = NULL; + req->md_sg = NULL; req->sg_cnt = 0; + req->md_sg_cnt = 0; } EXPORT_SYMBOL_GPL(nvmet_req_free_sgl); diff --git a/drivers/nvme/target/io-cmd-bdev.c b/drivers/nvme/target/io-cmd-bdev.c index d265cf5..4243156 100644 --- a/drivers/nvme/target/io-cmd-bdev.c +++ b/drivers/nvme/target/io-cmd-bdev.c @@ -174,6 +174,61 @@ static void nvmet_bio_done(struct bio *bio) bio_put(bio); } +#ifdef CONFIG_BLK_DEV_INTEGRITY +static int nvmet_bdev_alloc_bip(struct nvmet_req *req, struct bio *bio, + struct sg_mapping_iter *miter) +{ + struct blk_integrity *bi; + struct bio_integrity_payload *bip; + struct block_device *bdev = req->ns->bdev; + int rc; + size_t resid, len; + + bi = bdev_get_integrity(bdev); + if (unlikely(!bi)) { + pr_err("Unable to locate bio_integrity\n"); + return -ENODEV; + } + + bip = bio_integrity_alloc(bio, GFP_NOIO, + min_t(unsigned int, req->md_sg_cnt, BIO_MAX_PAGES)); + if (IS_ERR(bip)) { + pr_err("Unable to allocate bio_integrity_payload\n"); + return PTR_ERR(bip); + } + + bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio)); + /* virtual start sector must be in integrity interval units */ + bip_set_seed(bip, bio->bi_iter.bi_sector >> + (bi->interval_exp - SECTOR_SHIFT)); + + resid = bip->bip_iter.bi_size; + while (resid > 0 && sg_miter_next(miter)) { + len = min_t(size_t, miter->length, resid); + rc = bio_integrity_add_page(bio, miter->page, len, + offset_in_page(miter->addr)); + if (unlikely(rc != len)) { + pr_err("bio_integrity_add_page() failed; %d\n", rc); + sg_miter_stop(miter); + return -ENOMEM; + } + + resid -= len; + if (len < miter->length) + miter->consumed -= miter->length - len; + } + sg_miter_stop(miter); + + return 0; +} +#else +static int nvmet_bdev_alloc_bip(struct nvmet_req *req, struct bio *bio, + struct sg_mapping_iter *miter) +{ + return -EINVAL; +} +#endif /* CONFIG_BLK_DEV_INTEGRITY */ + static void nvmet_bdev_execute_rw(struct nvmet_req *req) { int sg_cnt = req->sg_cnt; @@ -181,9 +236,11 @@ static void nvmet_bdev_execute_rw(struct nvmet_req *req) struct scatterlist *sg; struct blk_plug plug; sector_t sector; - int op, i; + int op, i, rc; + struct sg_mapping_iter prot_miter; - if (!nvmet_check_transfer_len(req, nvmet_rw_data_len(req))) + if (!nvmet_check_transfer_len(req, + nvmet_rw_data_len(req) + req->md_len)) return; if (!req->sg_cnt) { @@ -218,11 +275,25 @@ static void nvmet_bdev_execute_rw(struct nvmet_req *req) bio->bi_opf = op; blk_start_plug(&plug); + if (req->md_len) + sg_miter_start(&prot_miter, req->md_sg, req->md_sg_cnt, + op == REQ_OP_READ ? SG_MITER_FROM_SG : + SG_MITER_TO_SG); + for_each_sg(req->sg, sg, req->sg_cnt, i) { while (bio_add_page(bio, sg_page(sg), sg->length, sg->offset) != sg->length) { struct bio *prev = bio; + if (req->md_len) { + rc = nvmet_bdev_alloc_bip(req, bio, + &prot_miter); + if (unlikely(rc)) { + bio_io_error(bio); + return; + } + } + bio = bio_alloc(GFP_KERNEL, min(sg_cnt, BIO_MAX_PAGES)); bio_set_dev(bio, req->ns->bdev); bio->bi_iter.bi_sector = sector; @@ -236,6 +307,14 @@ static void nvmet_bdev_execute_rw(struct nvmet_req *req) sg_cnt--; } + if (req->md_len) { + rc = nvmet_bdev_alloc_bip(req, bio, &prot_miter); + if (unlikely(rc)) { + bio_io_error(bio); + return; + } + } + submit_bio(bio); blk_finish_plug(&plug); } @@ -363,6 +442,8 @@ u16 nvmet_bdev_parse_io_cmd(struct nvmet_req *req) case nvme_cmd_read: case nvme_cmd_write: req->execute = nvmet_bdev_execute_rw; + if (req->sq->ctrl->pi_support && nvmet_ns_has_pi(req->ns)) + req->md_len = nvmet_rw_md_len(req); return 0; case nvme_cmd_flush: req->execute = nvmet_bdev_execute_flush; diff --git a/drivers/nvme/target/nvmet.h b/drivers/nvme/target/nvmet.h index ccdf820..7069520 100644 --- a/drivers/nvme/target/nvmet.h +++ b/drivers/nvme/target/nvmet.h @@ -309,6 +309,7 @@ struct nvmet_req { struct nvmet_cq *cq; struct nvmet_ns *ns; struct scatterlist *sg; + struct scatterlist *md_sg; struct bio_vec inline_bvec[NVMET_MAX_INLINE_BIOVEC]; union { struct { @@ -322,8 +323,10 @@ struct nvmet_req { } f; }; int sg_cnt; + int md_sg_cnt; /* data length as parsed from the SGL descriptor: */ size_t transfer_len; + size_t md_len; struct nvmet_port *port; -- 1.8.3.1 _______________________________________________ linux-nvme mailing list linux-nvme@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-nvme
next prev parent reply index Thread overview: 37+ messages / expand[flat|nested] mbox.gz Atom feed top 2020-04-28 13:11 [PATCH 00/15 V6] nvme-rdma/nvmet-rdma: Add metadata/T10-PI support Max Gurtovoy 2020-04-28 13:11 ` [PATCH 01/15] nvme: introduce namespace features flag Max Gurtovoy 2020-05-01 13:20 ` Christoph Hellwig 2020-05-01 14:24 ` Christoph Hellwig 2020-05-01 14:33 ` Max Gurtovoy 2020-04-28 13:11 ` [PATCH 02/15] nvme: introduce NVME_NS_METADATA_SUPPORTED flag Max Gurtovoy 2020-05-01 13:20 ` Christoph Hellwig 2020-04-28 13:11 ` [PATCH 03/15] nvme: make nvme_ns_has_pi accessible to transports Max Gurtovoy 2020-05-01 13:20 ` Christoph Hellwig 2020-04-28 13:11 ` [PATCH 04/15] nvme: enforce extended LBA format for fabrics metadata Max Gurtovoy 2020-05-01 13:21 ` Christoph Hellwig 2020-05-01 13:41 ` Christoph Hellwig 2020-04-28 13:11 ` [PATCH 05/15] nvme: introduce max_integrity_segments ctrl attribute Max Gurtovoy 2020-04-28 13:11 ` [PATCH 06/15] nvme: introduce NVME_INLINE_MD_SG_CNT Max Gurtovoy 2020-04-28 13:11 ` [PATCH 07/15] nvme-rdma: introduce nvme_rdma_sgl structure Max Gurtovoy 2020-04-28 13:11 ` [PATCH 08/15] nvme-rdma: add metadata/T10-PI support Max Gurtovoy 2020-05-01 14:26 ` Christoph Hellwig 2020-05-01 15:00 ` Max Gurtovoy 2020-04-28 13:11 ` [PATCH 09/15] nvmet: add metadata characteristics for a namespace Max Gurtovoy 2020-05-01 15:50 ` Christoph Hellwig 2020-04-28 13:11 ` [PATCH 10/15] nvmet: rename nvmet_rw_len to nvmet_rw_data_len Max Gurtovoy 2020-04-28 13:11 ` [PATCH 11/15] nvmet: rename nvmet_check_data_len to nvmet_check_transfer_len Max Gurtovoy 2020-04-28 13:11 ` [PATCH 12/15] nvme: add Metadata Capabilities enumerations Max Gurtovoy 2020-05-01 15:53 ` Christoph Hellwig 2020-05-03 12:43 ` Max Gurtovoy 2020-04-28 13:11 ` [PATCH 13/15] nvmet: add metadata/T10-PI support Max Gurtovoy 2020-05-01 15:58 ` Christoph Hellwig 2020-05-01 16:19 ` Christoph Hellwig 2020-04-28 13:11 ` Max Gurtovoy [this message] 2020-05-01 16:25 ` [PATCH 14/15] nvmet: add metadata support for block devices Christoph Hellwig 2020-04-28 13:11 ` [PATCH 15/15] nvmet-rdma: add metadata/T10-PI support Max Gurtovoy 2020-05-01 16:48 ` Christoph Hellwig 2020-05-03 16:29 ` Max Gurtovoy 2020-05-04 14:08 ` Christoph Hellwig 2020-05-04 14:19 ` Max Gurtovoy -- strict thread matches above, loose matches on Subject: below -- 2020-01-06 13:37 [PATCH 00/15 V3] nvme-rdma/nvmet-rdma: Add " Max Gurtovoy 2020-01-06 13:37 ` [PATCH 14/15] nvmet: Add metadata support for block devices Max Gurtovoy 2019-11-05 16:20 [PATCH 00/15] nvme-rdma/nvmet-rdma: Add metadata/T10-PI support Max Gurtovoy 2019-11-05 16:20 ` [PATCH 14/15] nvmet: Add metadata support for block devices Max Gurtovoy
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20200428131135.211521-15-maxg@mellanox.com \ --to=maxg@mellanox.com \ --cc=axboe@kernel.dk \ --cc=hch@lst.de \ --cc=idanb@mellanox.com \ --cc=israelr@mellanox.com \ --cc=jsmart2021@gmail.com \ --cc=kbusch@kernel.org \ --cc=linux-nvme@lists.infradead.org \ --cc=martin.petersen@oracle.com \ --cc=nitzanc@mellanox.com \ --cc=oren@mellanox.com \ --cc=sagi@grimberg.me \ --cc=shlomin@mellanox.com \ --cc=vladimirk@mellanox.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: link
Linux-NVME Archive on lore.kernel.org Archives are clonable: git clone --mirror https://lore.kernel.org/linux-nvme/0 linux-nvme/git/0.git # If you have public-inbox 1.1+ installed, you may # initialize and index your mirror using the following commands: public-inbox-init -V2 linux-nvme linux-nvme/ https://lore.kernel.org/linux-nvme \ linux-nvme@lists.infradead.org public-inbox-index linux-nvme Example config snippet for mirrors Newsgroup available over NNTP: nntp://nntp.lore.kernel.org/org.infradead.lists.linux-nvme AGPL code for this site: git clone https://public-inbox.org/public-inbox.git