From: Daniel Wagner <dwagner@suse.de> To: linux-nvme@lists.infradead.org Cc: linux-kernel@vger.kernel.org, Hannes Reinecke <hare@suse.de>, Sagi Grimberg <sagi@grimberg.me>, yi.he@emc.com, Daniel Wagner <dwagner@suse.de> Subject: [PATCH v2] nvme-tcp: Do not reset transport on data digest errors Date: Wed, 25 Aug 2021 14:42:59 +0200 [thread overview] Message-ID: <20210825124259.28707-1-dwagner@suse.de> (raw) The spec says 7.4.6.1 Digest Error handling When a host detects a data digest error in a C2HData PDU, that host shall continue processing C2HData PDUs associated with the command and when the command processing has completed, if a successful status was returned by the controller, the host shall fail the command with a non-fatal transport error. Currently the transport is reseted when a data digest error is detected. To fix this, keep track of the final status in the queue object and use it when completing the request. Signed-off-by: Daniel Wagner <dwagner@suse.de> --- The status member placed so that it fills up a hole in struct nvme_tcp_request: struct nvme_tcp_request { struct nvme_request req; /* 0 32 */ void * pdu; /* 32 8 */ struct nvme_tcp_queue * queue; /* 40 8 */ u32 data_len; /* 48 4 */ u32 pdu_len; /* 52 4 */ u32 pdu_sent; /* 56 4 */ u16 ttag; /* 60 2 */ u16 status; /* 62 2 */ /* --- cacheline 1 boundary (64 bytes) --- */ struct list_head entry; /* 64 16 */ struct llist_node lentry; /* 80 8 */ __le32 ddgst; /* 88 4 */ /* XXX 4 bytes hole, try to pack */ struct bio * curr_bio; /* 96 8 */ struct iov_iter iter; /* 104 40 */ /* --- cacheline 2 boundary (128 bytes) was 16 bytes ago --- */ size_t offset; /* 144 8 */ size_t data_sent; /* 152 8 */ enum nvme_tcp_send_state state; /* 160 4 */ /* size: 168, cachelines: 3, members: 16 */ /* sum members: 160, holes: 1, sum holes: 4 */ /* padding: 4 */ /* last cacheline: 40 bytes */ }; v1: - https://lore.kernel.org/linux-nvme/20210805121541.77613-1-dwagner@suse.de/ - moved 'status' from nvme_tcp_queue to nvme_tcp_request. drivers/nvme/host/tcp.c | 25 +++++++++++++++++++++---- 1 file changed, 21 insertions(+), 4 deletions(-) diff --git a/drivers/nvme/host/tcp.c b/drivers/nvme/host/tcp.c index 645025620154..23a8f7e11cfa 100644 --- a/drivers/nvme/host/tcp.c +++ b/drivers/nvme/host/tcp.c @@ -45,6 +45,7 @@ struct nvme_tcp_request { u32 pdu_len; u32 pdu_sent; u16 ttag; + u16 status; struct list_head entry; struct llist_node lentry; __le32 ddgst; @@ -485,7 +486,9 @@ static void nvme_tcp_error_recovery(struct nvme_ctrl *ctrl) static int nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue, struct nvme_completion *cqe) { + struct nvme_tcp_request *req; struct request *rq; + u16 status; rq = nvme_find_rq(nvme_tcp_tagset(queue), cqe->command_id); if (!rq) { @@ -496,7 +499,12 @@ static int nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue, return -EINVAL; } - if (!nvme_try_complete_req(rq, cqe->status, cqe->result)) + req = blk_mq_rq_to_pdu(rq); + status = req->status; + if (status == NVME_SC_SUCCESS) + status = cqe->status; + + if (!nvme_try_complete_req(rq, status, cqe->result)) nvme_complete_rq(rq); queue->nr_cqe++; @@ -506,6 +514,7 @@ static int nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue, static int nvme_tcp_handle_c2h_data(struct nvme_tcp_queue *queue, struct nvme_tcp_data_pdu *pdu) { + struct nvme_tcp_request *req; struct request *rq; rq = nvme_find_rq(nvme_tcp_tagset(queue), pdu->command_id); @@ -534,6 +543,8 @@ static int nvme_tcp_handle_c2h_data(struct nvme_tcp_queue *queue, return -EPROTO; } + req = blk_mq_rq_to_pdu(rq); + req->status = NVME_SC_SUCCESS; return 0; } @@ -758,7 +769,7 @@ static int nvme_tcp_recv_data(struct nvme_tcp_queue *queue, struct sk_buff *skb, queue->ddgst_remaining = NVME_TCP_DIGEST_LENGTH; } else { if (pdu->hdr.flags & NVME_TCP_F_DATA_SUCCESS) { - nvme_tcp_end_request(rq, NVME_SC_SUCCESS); + nvme_tcp_end_request(rq, req->status); queue->nr_cqe++; } nvme_tcp_init_recv_ctx(queue); @@ -788,18 +799,24 @@ static int nvme_tcp_recv_ddgst(struct nvme_tcp_queue *queue, return 0; if (queue->recv_ddgst != queue->exp_ddgst) { + struct request *rq = nvme_cid_to_rq(nvme_tcp_tagset(queue), + pdu->command_id); + struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); + + req->status = NVME_SC_DATA_XFER_ERROR; + dev_err(queue->ctrl->ctrl.device, "data digest error: recv %#x expected %#x\n", le32_to_cpu(queue->recv_ddgst), le32_to_cpu(queue->exp_ddgst)); - return -EIO; } if (pdu->hdr.flags & NVME_TCP_F_DATA_SUCCESS) { struct request *rq = nvme_cid_to_rq(nvme_tcp_tagset(queue), pdu->command_id); + struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); - nvme_tcp_end_request(rq, NVME_SC_SUCCESS); + nvme_tcp_end_request(rq, req->status); queue->nr_cqe++; } -- 2.29.2
WARNING: multiple messages have this Message-ID (diff)
From: Daniel Wagner <dwagner@suse.de> To: linux-nvme@lists.infradead.org Cc: linux-kernel@vger.kernel.org, Hannes Reinecke <hare@suse.de>, Sagi Grimberg <sagi@grimberg.me>, yi.he@emc.com, Daniel Wagner <dwagner@suse.de> Subject: [PATCH v2] nvme-tcp: Do not reset transport on data digest errors Date: Wed, 25 Aug 2021 14:42:59 +0200 [thread overview] Message-ID: <20210825124259.28707-1-dwagner@suse.de> (raw) The spec says 7.4.6.1 Digest Error handling When a host detects a data digest error in a C2HData PDU, that host shall continue processing C2HData PDUs associated with the command and when the command processing has completed, if a successful status was returned by the controller, the host shall fail the command with a non-fatal transport error. Currently the transport is reseted when a data digest error is detected. To fix this, keep track of the final status in the queue object and use it when completing the request. Signed-off-by: Daniel Wagner <dwagner@suse.de> --- The status member placed so that it fills up a hole in struct nvme_tcp_request: struct nvme_tcp_request { struct nvme_request req; /* 0 32 */ void * pdu; /* 32 8 */ struct nvme_tcp_queue * queue; /* 40 8 */ u32 data_len; /* 48 4 */ u32 pdu_len; /* 52 4 */ u32 pdu_sent; /* 56 4 */ u16 ttag; /* 60 2 */ u16 status; /* 62 2 */ /* --- cacheline 1 boundary (64 bytes) --- */ struct list_head entry; /* 64 16 */ struct llist_node lentry; /* 80 8 */ __le32 ddgst; /* 88 4 */ /* XXX 4 bytes hole, try to pack */ struct bio * curr_bio; /* 96 8 */ struct iov_iter iter; /* 104 40 */ /* --- cacheline 2 boundary (128 bytes) was 16 bytes ago --- */ size_t offset; /* 144 8 */ size_t data_sent; /* 152 8 */ enum nvme_tcp_send_state state; /* 160 4 */ /* size: 168, cachelines: 3, members: 16 */ /* sum members: 160, holes: 1, sum holes: 4 */ /* padding: 4 */ /* last cacheline: 40 bytes */ }; v1: - https://lore.kernel.org/linux-nvme/20210805121541.77613-1-dwagner@suse.de/ - moved 'status' from nvme_tcp_queue to nvme_tcp_request. drivers/nvme/host/tcp.c | 25 +++++++++++++++++++++---- 1 file changed, 21 insertions(+), 4 deletions(-) diff --git a/drivers/nvme/host/tcp.c b/drivers/nvme/host/tcp.c index 645025620154..23a8f7e11cfa 100644 --- a/drivers/nvme/host/tcp.c +++ b/drivers/nvme/host/tcp.c @@ -45,6 +45,7 @@ struct nvme_tcp_request { u32 pdu_len; u32 pdu_sent; u16 ttag; + u16 status; struct list_head entry; struct llist_node lentry; __le32 ddgst; @@ -485,7 +486,9 @@ static void nvme_tcp_error_recovery(struct nvme_ctrl *ctrl) static int nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue, struct nvme_completion *cqe) { + struct nvme_tcp_request *req; struct request *rq; + u16 status; rq = nvme_find_rq(nvme_tcp_tagset(queue), cqe->command_id); if (!rq) { @@ -496,7 +499,12 @@ static int nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue, return -EINVAL; } - if (!nvme_try_complete_req(rq, cqe->status, cqe->result)) + req = blk_mq_rq_to_pdu(rq); + status = req->status; + if (status == NVME_SC_SUCCESS) + status = cqe->status; + + if (!nvme_try_complete_req(rq, status, cqe->result)) nvme_complete_rq(rq); queue->nr_cqe++; @@ -506,6 +514,7 @@ static int nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue, static int nvme_tcp_handle_c2h_data(struct nvme_tcp_queue *queue, struct nvme_tcp_data_pdu *pdu) { + struct nvme_tcp_request *req; struct request *rq; rq = nvme_find_rq(nvme_tcp_tagset(queue), pdu->command_id); @@ -534,6 +543,8 @@ static int nvme_tcp_handle_c2h_data(struct nvme_tcp_queue *queue, return -EPROTO; } + req = blk_mq_rq_to_pdu(rq); + req->status = NVME_SC_SUCCESS; return 0; } @@ -758,7 +769,7 @@ static int nvme_tcp_recv_data(struct nvme_tcp_queue *queue, struct sk_buff *skb, queue->ddgst_remaining = NVME_TCP_DIGEST_LENGTH; } else { if (pdu->hdr.flags & NVME_TCP_F_DATA_SUCCESS) { - nvme_tcp_end_request(rq, NVME_SC_SUCCESS); + nvme_tcp_end_request(rq, req->status); queue->nr_cqe++; } nvme_tcp_init_recv_ctx(queue); @@ -788,18 +799,24 @@ static int nvme_tcp_recv_ddgst(struct nvme_tcp_queue *queue, return 0; if (queue->recv_ddgst != queue->exp_ddgst) { + struct request *rq = nvme_cid_to_rq(nvme_tcp_tagset(queue), + pdu->command_id); + struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); + + req->status = NVME_SC_DATA_XFER_ERROR; + dev_err(queue->ctrl->ctrl.device, "data digest error: recv %#x expected %#x\n", le32_to_cpu(queue->recv_ddgst), le32_to_cpu(queue->exp_ddgst)); - return -EIO; } if (pdu->hdr.flags & NVME_TCP_F_DATA_SUCCESS) { struct request *rq = nvme_cid_to_rq(nvme_tcp_tagset(queue), pdu->command_id); + struct nvme_tcp_request *req = blk_mq_rq_to_pdu(rq); - nvme_tcp_end_request(rq, NVME_SC_SUCCESS); + nvme_tcp_end_request(rq, req->status); queue->nr_cqe++; } -- 2.29.2 _______________________________________________ Linux-nvme mailing list Linux-nvme@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-nvme
next reply other threads:[~2021-08-25 12:43 UTC|newest] Thread overview: 6+ messages / expand[flat|nested] mbox.gz Atom feed top 2021-08-25 12:42 Daniel Wagner [this message] 2021-08-25 12:42 ` [PATCH v2] nvme-tcp: Do not reset transport on data digest errors Daniel Wagner 2021-08-25 15:27 ` Hannes Reinecke 2021-08-25 15:27 ` Hannes Reinecke 2021-08-26 8:04 ` Daniel Wagner 2021-08-26 8:04 ` Daniel Wagner
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20210825124259.28707-1-dwagner@suse.de \ --to=dwagner@suse.de \ --cc=hare@suse.de \ --cc=linux-kernel@vger.kernel.org \ --cc=linux-nvme@lists.infradead.org \ --cc=sagi@grimberg.me \ --cc=yi.he@emc.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.