netdev.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Hannes Reinecke <hare@suse.de>
To: Shai Malin <smalin@marvell.com>,
	netdev@vger.kernel.org, linux-nvme@lists.infradead.org,
	sagi@grimberg.me, hch@lst.de, axboe@fb.com, kbusch@kernel.org
Cc: "David S . Miller davem @ davemloft . net --cc=Jakub Kicinski" 
	<kuba@kernel.org>,
	aelior@marvell.com, mkalderon@marvell.com, okulkarni@marvell.com,
	pkushwaha@marvell.com, malin1024@gmail.com
Subject: Re: [RFC PATCH v4 02/27] qed: Add NVMeTCP Offload Connection Level FW and HW HSI
Date: Sat, 1 May 2021 19:28:07 +0200	[thread overview]
Message-ID: <c35e89df-e9b7-fd45-9ad6-bf0f6be1dc5d@suse.de> (raw)
In-Reply-To: <20210429190926.5086-3-smalin@marvell.com>

On 4/29/21 9:09 PM, Shai Malin wrote:
> This patch introduces the NVMeTCP HSI and HSI functionality in order to
> initialize and interact with the HW device as part of the connection level
> HSI.
> 
> This includes:
> - Connection offload: offload a TCP connection to the FW.
> - Connection update: update the ICReq-ICResp params
> - Connection clear SQ: outstanding IOs FW flush.
> - Connection termination: terminate the TCP connection and flush the FW.
> 
> Acked-by: Igor Russkikh <irusskikh@marvell.com>
> Signed-off-by: Prabhakar Kushwaha <pkushwaha@marvell.com>
> Signed-off-by: Omkar Kulkarni <okulkarni@marvell.com>
> Signed-off-by: Shai Malin <smalin@marvell.com>
> Signed-off-by: Michal Kalderon <mkalderon@marvell.com>
> Signed-off-by: Ariel Elior <aelior@marvell.com>
> ---
>   drivers/net/ethernet/qlogic/qed/qed_nvmetcp.c | 580 +++++++++++++++++-
>   drivers/net/ethernet/qlogic/qed/qed_nvmetcp.h |  63 ++
>   drivers/net/ethernet/qlogic/qed/qed_sp.h      |   3 +
>   include/linux/qed/nvmetcp_common.h            | 143 +++++
>   include/linux/qed/qed_nvmetcp_if.h            |  94 +++
>   5 files changed, 881 insertions(+), 2 deletions(-)
> 
> diff --git a/drivers/net/ethernet/qlogic/qed/qed_nvmetcp.c b/drivers/net/ethernet/qlogic/qed/qed_nvmetcp.c
> index da3b5002d216..79bd1cc6677f 100644
> --- a/drivers/net/ethernet/qlogic/qed/qed_nvmetcp.c
> +++ b/drivers/net/ethernet/qlogic/qed/qed_nvmetcp.c
> @@ -259,6 +259,578 @@ static int qed_nvmetcp_start(struct qed_dev *cdev,
>   	return 0;
>   }
>   
> +static struct qed_hash_nvmetcp_con *qed_nvmetcp_get_hash(struct qed_dev *cdev,
> +							 u32 handle)
> +{
> +	struct qed_hash_nvmetcp_con *hash_con = NULL;
> +
> +	if (!(cdev->flags & QED_FLAG_STORAGE_STARTED))
> +		return NULL;
> +
> +	hash_for_each_possible(cdev->connections, hash_con, node, handle) {
> +		if (hash_con->con->icid == handle)
> +			break;
> +	}
> +
> +	if (!hash_con || hash_con->con->icid != handle)
> +		return NULL;
> +
> +	return hash_con;
> +}
> +
> +static int qed_sp_nvmetcp_conn_offload(struct qed_hwfn *p_hwfn,
> +				       struct qed_nvmetcp_conn *p_conn,
> +				       enum spq_mode comp_mode,
> +				       struct qed_spq_comp_cb *p_comp_addr)
> +{
> +	struct nvmetcp_spe_conn_offload *p_ramrod = NULL;
> +	struct tcp_offload_params_opt2 *p_tcp2 = NULL;
> +	struct qed_sp_init_data init_data = { 0 };
> +	struct qed_spq_entry *p_ent = NULL;
> +	dma_addr_t r2tq_pbl_addr;
> +	dma_addr_t xhq_pbl_addr;
> +	dma_addr_t uhq_pbl_addr;
> +	u16 physical_q;
> +	int rc = 0;
> +	u32 dval;
> +	u8 i;
> +
> +	/* Get SPQ entry */
> +	init_data.cid = p_conn->icid;
> +	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
> +	init_data.comp_mode = comp_mode;
> +	init_data.p_comp_data = p_comp_addr;
> +
> +	rc = qed_sp_init_request(p_hwfn, &p_ent,
> +				 NVMETCP_RAMROD_CMD_ID_OFFLOAD_CONN,
> +				 PROTOCOLID_NVMETCP, &init_data);
> +	if (rc)
> +		return rc;
> +
> +	p_ramrod = &p_ent->ramrod.nvmetcp_conn_offload;
> +
> +	/* Transmission PQ is the first of the PF */
> +	physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_OFLD);
> +	p_conn->physical_q0 = cpu_to_le16(physical_q);
> +	p_ramrod->nvmetcp.physical_q0 = cpu_to_le16(physical_q);
> +
> +	/* nvmetcp Pure-ACK PQ */
> +	physical_q = qed_get_cm_pq_idx(p_hwfn, PQ_FLAGS_ACK);
> +	p_conn->physical_q1 = cpu_to_le16(physical_q);
> +	p_ramrod->nvmetcp.physical_q1 = cpu_to_le16(physical_q);
> +
> +	p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id);
> +
> +	DMA_REGPAIR_LE(p_ramrod->nvmetcp.sq_pbl_addr, p_conn->sq_pbl_addr);
> +
> +	r2tq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->r2tq);
> +	DMA_REGPAIR_LE(p_ramrod->nvmetcp.r2tq_pbl_addr, r2tq_pbl_addr);
> +
> +	xhq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->xhq);
> +	DMA_REGPAIR_LE(p_ramrod->nvmetcp.xhq_pbl_addr, xhq_pbl_addr);
> +
> +	uhq_pbl_addr = qed_chain_get_pbl_phys(&p_conn->uhq);
> +	DMA_REGPAIR_LE(p_ramrod->nvmetcp.uhq_pbl_addr, uhq_pbl_addr);
> +
> +	p_ramrod->nvmetcp.flags = p_conn->offl_flags;
> +	p_ramrod->nvmetcp.default_cq = p_conn->default_cq;
> +	p_ramrod->nvmetcp.initial_ack = 0;
> +
> +	DMA_REGPAIR_LE(p_ramrod->nvmetcp.nvmetcp.cccid_itid_table_addr,
> +		       p_conn->nvmetcp_cccid_itid_table_addr);
> +	p_ramrod->nvmetcp.nvmetcp.cccid_max_range =
> +		 cpu_to_le16(p_conn->nvmetcp_cccid_max_range);
> +
> +	p_tcp2 = &p_ramrod->tcp;
> +
> +	qed_set_fw_mac_addr(&p_tcp2->remote_mac_addr_hi,
> +			    &p_tcp2->remote_mac_addr_mid,
> +			    &p_tcp2->remote_mac_addr_lo, p_conn->remote_mac);
> +	qed_set_fw_mac_addr(&p_tcp2->local_mac_addr_hi,
> +			    &p_tcp2->local_mac_addr_mid,
> +			    &p_tcp2->local_mac_addr_lo, p_conn->local_mac);
> +
> +	p_tcp2->vlan_id = cpu_to_le16(p_conn->vlan_id);
> +	p_tcp2->flags = cpu_to_le16(p_conn->tcp_flags);
> +
> +	p_tcp2->ip_version = p_conn->ip_version;
> +	for (i = 0; i < 4; i++) {
> +		dval = p_conn->remote_ip[i];
> +		p_tcp2->remote_ip[i] = cpu_to_le32(dval);
> +		dval = p_conn->local_ip[i];
> +		p_tcp2->local_ip[i] = cpu_to_le32(dval);
> +	}
> +

What is this?
Some convoluted way of assigning the IP address in little endian?
Pointless if it's IPv4, as then each bit is just one byte.
And if it's for IPv6, what do you do for IPv4?
And isn't there a helper for it?

> +	p_tcp2->flow_label = cpu_to_le32(p_conn->flow_label);
> +	p_tcp2->ttl = p_conn->ttl;
> +	p_tcp2->tos_or_tc = p_conn->tos_or_tc;
> +	p_tcp2->remote_port = cpu_to_le16(p_conn->remote_port);
> +	p_tcp2->local_port = cpu_to_le16(p_conn->local_port);
> +	p_tcp2->mss = cpu_to_le16(p_conn->mss);
> +	p_tcp2->rcv_wnd_scale = p_conn->rcv_wnd_scale;
> +	p_tcp2->connect_mode = p_conn->connect_mode;
> +	p_tcp2->cwnd = cpu_to_le32(p_conn->cwnd);
> +	p_tcp2->ka_max_probe_cnt = p_conn->ka_max_probe_cnt;
> +	p_tcp2->ka_timeout = cpu_to_le32(p_conn->ka_timeout);
> +	p_tcp2->max_rt_time = cpu_to_le32(p_conn->max_rt_time);
> +	p_tcp2->ka_interval = cpu_to_le32(p_conn->ka_interval);
> +
> +	return qed_spq_post(p_hwfn, p_ent, NULL);
> +}
> +
> +static int qed_sp_nvmetcp_conn_update(struct qed_hwfn *p_hwfn,
> +				      struct qed_nvmetcp_conn *p_conn,
> +				      enum spq_mode comp_mode,
> +				      struct qed_spq_comp_cb *p_comp_addr)
> +{
> +	struct nvmetcp_conn_update_ramrod_params *p_ramrod = NULL;
> +	struct qed_spq_entry *p_ent = NULL;
> +	struct qed_sp_init_data init_data;
> +	int rc = -EINVAL;
> +	u32 dval;
> +
> +	/* Get SPQ entry */
> +	memset(&init_data, 0, sizeof(init_data));
> +	init_data.cid = p_conn->icid;
> +	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
> +	init_data.comp_mode = comp_mode;
> +	init_data.p_comp_data = p_comp_addr;
> +
> +	rc = qed_sp_init_request(p_hwfn, &p_ent,
> +				 NVMETCP_RAMROD_CMD_ID_UPDATE_CONN,
> +				 PROTOCOLID_NVMETCP, &init_data);
> +	if (rc)
> +		return rc;
> +
> +	p_ramrod = &p_ent->ramrod.nvmetcp_conn_update;
> +	p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id);
> +	p_ramrod->flags = p_conn->update_flag;
> +	p_ramrod->max_seq_size = cpu_to_le32(p_conn->max_seq_size);
> +	dval = p_conn->max_recv_pdu_length;
> +	p_ramrod->max_recv_pdu_length = cpu_to_le32(dval);
> +	dval = p_conn->max_send_pdu_length;
> +	p_ramrod->max_send_pdu_length = cpu_to_le32(dval);
> +	dval = p_conn->first_seq_length;
> +	p_ramrod->first_seq_length = cpu_to_le32(dval);
> +
> +	return qed_spq_post(p_hwfn, p_ent, NULL);
> +}
> +
> +static int qed_sp_nvmetcp_conn_terminate(struct qed_hwfn *p_hwfn,
> +					 struct qed_nvmetcp_conn *p_conn,
> +					 enum spq_mode comp_mode,
> +					 struct qed_spq_comp_cb *p_comp_addr)
> +{
> +	struct nvmetcp_spe_conn_termination *p_ramrod = NULL;
> +	struct qed_spq_entry *p_ent = NULL;
> +	struct qed_sp_init_data init_data;
> +	int rc = -EINVAL;
> +
> +	/* Get SPQ entry */
> +	memset(&init_data, 0, sizeof(init_data));
> +	init_data.cid = p_conn->icid;
> +	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
> +	init_data.comp_mode = comp_mode;
> +	init_data.p_comp_data = p_comp_addr;
> +
> +	rc = qed_sp_init_request(p_hwfn, &p_ent,
> +				 NVMETCP_RAMROD_CMD_ID_TERMINATION_CONN,
> +				 PROTOCOLID_NVMETCP, &init_data);
> +	if (rc)
> +		return rc;
> +
> +	p_ramrod = &p_ent->ramrod.nvmetcp_conn_terminate;
> +	p_ramrod->conn_id = cpu_to_le16(p_conn->conn_id);
> +	p_ramrod->abortive = p_conn->abortive_dsconnect;
> +
> +	return qed_spq_post(p_hwfn, p_ent, NULL);
> +}
> +
> +static int qed_sp_nvmetcp_conn_clear_sq(struct qed_hwfn *p_hwfn,
> +					struct qed_nvmetcp_conn *p_conn,
> +					enum spq_mode comp_mode,
> +					struct qed_spq_comp_cb *p_comp_addr)
> +{
> +	struct qed_spq_entry *p_ent = NULL;
> +	struct qed_sp_init_data init_data;
> +	int rc = -EINVAL;
> +
> +	/* Get SPQ entry */
> +	memset(&init_data, 0, sizeof(init_data));
> +	init_data.cid = p_conn->icid;
> +	init_data.opaque_fid = p_hwfn->hw_info.opaque_fid;
> +	init_data.comp_mode = comp_mode;
> +	init_data.p_comp_data = p_comp_addr;
> +
> +	rc = qed_sp_init_request(p_hwfn, &p_ent,
> +				 NVMETCP_RAMROD_CMD_ID_CLEAR_SQ,
> +				 PROTOCOLID_NVMETCP, &init_data);
> +	if (rc)
> +		return rc;
> +
> +	return qed_spq_post(p_hwfn, p_ent, NULL);
> +}
> +
> +static void __iomem *qed_nvmetcp_get_db_addr(struct qed_hwfn *p_hwfn, u32 cid)
> +{
> +	return (u8 __iomem *)p_hwfn->doorbells +
> +			     qed_db_addr(cid, DQ_DEMS_LEGACY);
> +}
> +
> +static int qed_nvmetcp_allocate_connection(struct qed_hwfn *p_hwfn,
> +					   struct qed_nvmetcp_conn **p_out_conn)
> +{
> +	struct qed_chain_init_params params = {
> +		.mode		= QED_CHAIN_MODE_PBL,
> +		.intended_use	= QED_CHAIN_USE_TO_CONSUME_PRODUCE,
> +		.cnt_type	= QED_CHAIN_CNT_TYPE_U16,
> +	};
> +	struct qed_nvmetcp_pf_params *p_params = NULL;
> +	struct qed_nvmetcp_conn *p_conn = NULL;
> +	int rc = 0;
> +
> +	/* Try finding a free connection that can be used */
> +	spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +	if (!list_empty(&p_hwfn->p_nvmetcp_info->free_list))
> +		p_conn = list_first_entry(&p_hwfn->p_nvmetcp_info->free_list,
> +					  struct qed_nvmetcp_conn, list_entry);
> +	if (p_conn) {
> +		list_del(&p_conn->list_entry);
> +		spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +		*p_out_conn = p_conn;
> +
> +		return 0;
> +	}
> +	spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +
> +	/* Need to allocate a new connection */
> +	p_params = &p_hwfn->pf_params.nvmetcp_pf_params;
> +
> +	p_conn = kzalloc(sizeof(*p_conn), GFP_KERNEL);
> +	if (!p_conn)
> +		return -ENOMEM;
> +
> +	params.num_elems = p_params->num_r2tq_pages_in_ring *
> +			   QED_CHAIN_PAGE_SIZE / sizeof(struct nvmetcp_wqe);
> +	params.elem_size = sizeof(struct nvmetcp_wqe);
> +
> +	rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->r2tq, &params);
> +	if (rc)
> +		goto nomem_r2tq;
> +
> +	params.num_elems = p_params->num_uhq_pages_in_ring *
> +			   QED_CHAIN_PAGE_SIZE / sizeof(struct iscsi_uhqe);
> +	params.elem_size = sizeof(struct iscsi_uhqe);
> +
> +	rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->uhq, &params);
> +	if (rc)
> +		goto nomem_uhq;
> +
> +	params.elem_size = sizeof(struct iscsi_xhqe);
> +
> +	rc = qed_chain_alloc(p_hwfn->cdev, &p_conn->xhq, &params);
> +	if (rc)
> +		goto nomem;
> +
> +	p_conn->free_on_delete = true;
> +	*p_out_conn = p_conn;
> +
> +	return 0;
> +
> +nomem:
> +	qed_chain_free(p_hwfn->cdev, &p_conn->uhq);
> +nomem_uhq:
> +	qed_chain_free(p_hwfn->cdev, &p_conn->r2tq);
> +nomem_r2tq:
> +	kfree(p_conn);
> +
> +	return -ENOMEM;
> +}
> +
> +static int qed_nvmetcp_acquire_connection(struct qed_hwfn *p_hwfn,
> +					  struct qed_nvmetcp_conn **p_out_conn)
> +{
> +	struct qed_nvmetcp_conn *p_conn = NULL;
> +	int rc = 0;
> +	u32 icid;
> +
> +	spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +	rc = qed_cxt_acquire_cid(p_hwfn, PROTOCOLID_NVMETCP, &icid);
> +	spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +
> +	if (rc)
> +		return rc;
> +
> +	rc = qed_nvmetcp_allocate_connection(p_hwfn, &p_conn);
> +	if (rc) {
> +		spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +		qed_cxt_release_cid(p_hwfn, icid);
> +		spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +
> +		return rc;
> +	}
> +
> +	p_conn->icid = icid;
> +	p_conn->conn_id = (u16)icid;
> +	p_conn->fw_cid = (p_hwfn->hw_info.opaque_fid << 16) | icid;
> +	*p_out_conn = p_conn;
> +
> +	return rc;
> +}
> +
> +static void qed_nvmetcp_release_connection(struct qed_hwfn *p_hwfn,
> +					   struct qed_nvmetcp_conn *p_conn)
> +{
> +	spin_lock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +	list_add_tail(&p_conn->list_entry, &p_hwfn->p_nvmetcp_info->free_list);
> +	qed_cxt_release_cid(p_hwfn, p_conn->icid);
> +	spin_unlock_bh(&p_hwfn->p_nvmetcp_info->lock);
> +}
> +
> +static void qed_nvmetcp_free_connection(struct qed_hwfn *p_hwfn,
> +					struct qed_nvmetcp_conn *p_conn)
> +{
> +	qed_chain_free(p_hwfn->cdev, &p_conn->xhq);
> +	qed_chain_free(p_hwfn->cdev, &p_conn->uhq);
> +	qed_chain_free(p_hwfn->cdev, &p_conn->r2tq);
> +
> +	kfree(p_conn);
> +}
> +
> +int qed_nvmetcp_alloc(struct qed_hwfn *p_hwfn)
> +{
> +	struct qed_nvmetcp_info *p_nvmetcp_info;
> +
> +	p_nvmetcp_info = kzalloc(sizeof(*p_nvmetcp_info), GFP_KERNEL);
> +	if (!p_nvmetcp_info)
> +		return -ENOMEM;
> +
> +	INIT_LIST_HEAD(&p_nvmetcp_info->free_list);
> +
> +	p_hwfn->p_nvmetcp_info = p_nvmetcp_info;
> +
> +	return 0;
> +}
> +
> +void qed_nvmetcp_setup(struct qed_hwfn *p_hwfn)
> +{
> +	spin_lock_init(&p_hwfn->p_nvmetcp_info->lock);
> +}
> +
> +void qed_nvmetcp_free(struct qed_hwfn *p_hwfn)
> +{
> +	struct qed_nvmetcp_conn *p_conn = NULL;
> +
> +	if (!p_hwfn->p_nvmetcp_info)
> +		return;
> +
> +	while (!list_empty(&p_hwfn->p_nvmetcp_info->free_list)) {
> +		p_conn = list_first_entry(&p_hwfn->p_nvmetcp_info->free_list,
> +					  struct qed_nvmetcp_conn, list_entry);
> +		if (p_conn) {
> +			list_del(&p_conn->list_entry);
> +			qed_nvmetcp_free_connection(p_hwfn, p_conn);
> +		}
> +	}
> +
> +	kfree(p_hwfn->p_nvmetcp_info);
> +	p_hwfn->p_nvmetcp_info = NULL;
> +}
> +
> +static int qed_nvmetcp_acquire_conn(struct qed_dev *cdev,
> +				    u32 *handle,
> +				    u32 *fw_cid, void __iomem **p_doorbell)
> +{
> +	struct qed_hash_nvmetcp_con *hash_con;
> +	int rc;
> +
> +	/* Allocate a hashed connection */
> +	hash_con = kzalloc(sizeof(*hash_con), GFP_ATOMIC);
> +	if (!hash_con)
> +		return -ENOMEM;
> +
> +	/* Acquire the connection */
> +	rc = qed_nvmetcp_acquire_connection(QED_AFFIN_HWFN(cdev),
> +					    &hash_con->con);
> +	if (rc) {
> +		DP_NOTICE(cdev, "Failed to acquire Connection\n");
> +		kfree(hash_con);
> +
> +		return rc;
> +	}
> +
> +	/* Added the connection to hash table */
> +	*handle = hash_con->con->icid;
> +	*fw_cid = hash_con->con->fw_cid;
> +	hash_add(cdev->connections, &hash_con->node, *handle);
> +
> +	if (p_doorbell)
> +		*p_doorbell = qed_nvmetcp_get_db_addr(QED_AFFIN_HWFN(cdev),
> +						      *handle);
> +
> +	return 0;
> +}
> +
> +static int qed_nvmetcp_release_conn(struct qed_dev *cdev, u32 handle)
> +{
> +	struct qed_hash_nvmetcp_con *hash_con;
> +
> +	hash_con = qed_nvmetcp_get_hash(cdev, handle);
> +	if (!hash_con) {
> +		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
> +			  handle);
> +
> +		return -EINVAL;
> +	}
> +
> +	hlist_del(&hash_con->node);
> +	qed_nvmetcp_release_connection(QED_AFFIN_HWFN(cdev), hash_con->con);
> +	kfree(hash_con);
> +
> +	return 0;
> +}
> +
> +static int qed_nvmetcp_offload_conn(struct qed_dev *cdev, u32 handle,
> +				    struct qed_nvmetcp_params_offload *conn_info)
> +{
> +	struct qed_hash_nvmetcp_con *hash_con;
> +	struct qed_nvmetcp_conn *con;
> +
> +	hash_con = qed_nvmetcp_get_hash(cdev, handle);
> +	if (!hash_con) {
> +		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
> +			  handle);
> +
> +		return -EINVAL;
> +	}
> +
> +	/* Update the connection with information from the params */
> +	con = hash_con->con;
> +
> +	/* FW initializations */
> +	con->layer_code = NVMETCP_SLOW_PATH_LAYER_CODE;
> +	con->sq_pbl_addr = conn_info->sq_pbl_addr;
> +	con->nvmetcp_cccid_max_range = conn_info->nvmetcp_cccid_max_range;
> +	con->nvmetcp_cccid_itid_table_addr = conn_info->nvmetcp_cccid_itid_table_addr;
> +	con->default_cq = conn_info->default_cq;
> +
> +	SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_TARGET_MODE, 0);
> +	SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_NVMETCP_MODE, 1);
> +	SET_FIELD(con->offl_flags, NVMETCP_CONN_OFFLOAD_PARAMS_TCP_ON_CHIP_1B, 1);
> +
> +	/* Networking and TCP stack initializations */
> +	ether_addr_copy(con->local_mac, conn_info->src.mac);
> +	ether_addr_copy(con->remote_mac, conn_info->dst.mac);
> +	memcpy(con->local_ip, conn_info->src.ip, sizeof(con->local_ip));
> +	memcpy(con->remote_ip, conn_info->dst.ip, sizeof(con->remote_ip));
> +	con->local_port = conn_info->src.port;
> +	con->remote_port = conn_info->dst.port;
> +	con->vlan_id = conn_info->vlan_id;
> +
> +	if (conn_info->timestamp_en)
> +		SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_TS_EN, 1);
> +
> +	if (conn_info->delayed_ack_en)
> +		SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_DA_EN, 1);
> +
> +	if (conn_info->tcp_keep_alive_en)
> +		SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_KA_EN, 1);
> +
> +	if (conn_info->ecn_en)
> +		SET_FIELD(con->tcp_flags, TCP_OFFLOAD_PARAMS_OPT2_ECN_EN, 1);
> +
> +	con->ip_version = conn_info->ip_version;
> +	con->flow_label = QED_TCP_FLOW_LABEL;
> +	con->ka_max_probe_cnt = conn_info->ka_max_probe_cnt;
> +	con->ka_timeout = conn_info->ka_timeout;
> +	con->ka_interval = conn_info->ka_interval;
> +	con->max_rt_time = conn_info->max_rt_time;
> +	con->ttl = conn_info->ttl;
> +	con->tos_or_tc = conn_info->tos_or_tc;
> +	con->mss = conn_info->mss;
> +	con->cwnd = conn_info->cwnd;
> +	con->rcv_wnd_scale = conn_info->rcv_wnd_scale;
> +	con->connect_mode = 0; /* TCP_CONNECT_ACTIVE */
> +
> +	return qed_sp_nvmetcp_conn_offload(QED_AFFIN_HWFN(cdev), con,
> +					 QED_SPQ_MODE_EBLOCK, NULL);
> +}
> +
> +static int qed_nvmetcp_update_conn(struct qed_dev *cdev,
> +				   u32 handle,
> +				   struct qed_nvmetcp_params_update *conn_info)
> +{
> +	struct qed_hash_nvmetcp_con *hash_con;
> +	struct qed_nvmetcp_conn *con;
> +
> +	hash_con = qed_nvmetcp_get_hash(cdev, handle);
> +	if (!hash_con) {
> +		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
> +			  handle);
> +
> +		return -EINVAL;
> +	}
> +
> +	/* Update the connection with information from the params */
> +	con = hash_con->con;
> +
> +	SET_FIELD(con->update_flag,
> +		  ISCSI_CONN_UPDATE_RAMROD_PARAMS_INITIAL_R2T, 0);
> +	SET_FIELD(con->update_flag,
> +		  ISCSI_CONN_UPDATE_RAMROD_PARAMS_IMMEDIATE_DATA, 1);
> +
> +	if (conn_info->hdr_digest_en)
> +		SET_FIELD(con->update_flag, ISCSI_CONN_UPDATE_RAMROD_PARAMS_HD_EN, 1);
> +
> +	if (conn_info->data_digest_en)
> +		SET_FIELD(con->update_flag, ISCSI_CONN_UPDATE_RAMROD_PARAMS_DD_EN, 1);
> +
> +	/* Placeholder - initialize pfv, cpda, hpda */
> +
> +	con->max_seq_size = conn_info->max_io_size;
> +	con->max_recv_pdu_length = conn_info->max_recv_pdu_length;
> +	con->max_send_pdu_length = conn_info->max_send_pdu_length;
> +	con->first_seq_length = conn_info->max_io_size;
> +
> +	return qed_sp_nvmetcp_conn_update(QED_AFFIN_HWFN(cdev), con,
> +					QED_SPQ_MODE_EBLOCK, NULL);
> +}
> +
> +static int qed_nvmetcp_clear_conn_sq(struct qed_dev *cdev, u32 handle)
> +{
> +	struct qed_hash_nvmetcp_con *hash_con;
> +
> +	hash_con = qed_nvmetcp_get_hash(cdev, handle);
> +	if (!hash_con) {
> +		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
> +			  handle);
> +
> +		return -EINVAL;
> +	}
> +
> +	return qed_sp_nvmetcp_conn_clear_sq(QED_AFFIN_HWFN(cdev), hash_con->con,
> +					    QED_SPQ_MODE_EBLOCK, NULL);
> +}
> +
> +static int qed_nvmetcp_destroy_conn(struct qed_dev *cdev,
> +				    u32 handle, u8 abrt_conn)
> +{
> +	struct qed_hash_nvmetcp_con *hash_con;
> +
> +	hash_con = qed_nvmetcp_get_hash(cdev, handle);
> +	if (!hash_con) {
> +		DP_NOTICE(cdev, "Failed to find connection for handle %d\n",
> +			  handle);
> +
> +		return -EINVAL;
> +	}
> +
> +	hash_con->con->abortive_dsconnect = abrt_conn;
> +
> +	return qed_sp_nvmetcp_conn_terminate(QED_AFFIN_HWFN(cdev), hash_con->con,
> +					   QED_SPQ_MODE_EBLOCK, NULL);
> +}
> +
>   static const struct qed_nvmetcp_ops qed_nvmetcp_ops_pass = {
>   	.common = &qed_common_ops_pass,
>   	.ll2 = &qed_ll2_ops_pass,
> @@ -266,8 +838,12 @@ static const struct qed_nvmetcp_ops qed_nvmetcp_ops_pass = {
>   	.register_ops = &qed_register_nvmetcp_ops,
>   	.start = &qed_nvmetcp_start,
>   	.stop = &qed_nvmetcp_stop,
> -
> -	/* Placeholder - Connection level ops */
> +	.acquire_conn = &qed_nvmetcp_acquire_conn,
> +	.release_conn = &qed_nvmetcp_release_conn,
> +	.offload_conn = &qed_nvmetcp_offload_conn,
> +	.update_conn = &qed_nvmetcp_update_conn,
> +	.destroy_conn = &qed_nvmetcp_destroy_conn,
> +	.clear_sq = &qed_nvmetcp_clear_conn_sq,
>   };
>   
>   const struct qed_nvmetcp_ops *qed_get_nvmetcp_ops(void)
> diff --git a/drivers/net/ethernet/qlogic/qed/qed_nvmetcp.h b/drivers/net/ethernet/qlogic/qed/qed_nvmetcp.h
> index 774b46ade408..749169f0bdb1 100644
> --- a/drivers/net/ethernet/qlogic/qed/qed_nvmetcp.h
> +++ b/drivers/net/ethernet/qlogic/qed/qed_nvmetcp.h
> @@ -19,6 +19,7 @@
>   #define QED_NVMETCP_FW_CQ_SIZE (4 * 1024)
>   
>   /* tcp parameters */
> +#define QED_TCP_FLOW_LABEL 0
>   #define QED_TCP_TWO_MSL_TIMER 4000
>   #define QED_TCP_HALF_WAY_CLOSE_TIMEOUT 10
>   #define QED_TCP_MAX_FIN_RT 2
> @@ -32,6 +33,68 @@ struct qed_nvmetcp_info {
>   	nvmetcp_event_cb_t event_cb;
>   };
>   
> +struct qed_hash_nvmetcp_con {
> +	struct hlist_node node;
> +	struct qed_nvmetcp_conn *con;
> +};
> +
> +struct qed_nvmetcp_conn {
> +	struct list_head list_entry;
> +	bool free_on_delete;
> +
> +	u16 conn_id;
> +	u32 icid;
> +	u32 fw_cid;
> +
> +	u8 layer_code;
> +	u8 offl_flags;
> +	u8 connect_mode;
> +
> +	dma_addr_t sq_pbl_addr;
> +	struct qed_chain r2tq;
> +	struct qed_chain xhq;
> +	struct qed_chain uhq;
> +
> +	u8 local_mac[6];
> +	u8 remote_mac[6];
> +	u8 ip_version;
> +	u8 ka_max_probe_cnt;
> +
> +	u16 vlan_id;
> +	u16 tcp_flags;
> +	u32 remote_ip[4];
> +	u32 local_ip[4];
> +
> +	u32 flow_label;
> +	u32 ka_timeout;
> +	u32 ka_interval;
> +	u32 max_rt_time;
> +
> +	u8 ttl;
> +	u8 tos_or_tc;
> +	u16 remote_port;
> +	u16 local_port;
> +	u16 mss;
> +	u8 rcv_wnd_scale;
> +	u32 rcv_wnd;
> +	u32 cwnd;
> +
> +	u8 update_flag;
> +	u8 default_cq;
> +	u8 abortive_dsconnect;
> +
> +	u32 max_seq_size;
> +	u32 max_recv_pdu_length;
> +	u32 max_send_pdu_length;
> +	u32 first_seq_length;
> +
> +	u16 physical_q0;
> +	u16 physical_q1;
> +
> +	u16 nvmetcp_cccid_max_range;
> +	dma_addr_t nvmetcp_cccid_itid_table_addr;
> +};
> +
>   #if IS_ENABLED(CONFIG_QED_NVMETCP)
>   int qed_nvmetcp_alloc(struct qed_hwfn *p_hwfn);
>   void qed_nvmetcp_setup(struct qed_hwfn *p_hwfn);
> diff --git a/drivers/net/ethernet/qlogic/qed/qed_sp.h b/drivers/net/ethernet/qlogic/qed/qed_sp.h
> index 525159e747a5..60ff3222bf55 100644
> --- a/drivers/net/ethernet/qlogic/qed/qed_sp.h
> +++ b/drivers/net/ethernet/qlogic/qed/qed_sp.h
> @@ -101,6 +101,9 @@ union ramrod_data {
>   	struct iscsi_spe_conn_termination iscsi_conn_terminate;
>   
>   	struct nvmetcp_init_ramrod_params nvmetcp_init;
> +	struct nvmetcp_spe_conn_offload nvmetcp_conn_offload;
> +	struct nvmetcp_conn_update_ramrod_params nvmetcp_conn_update;
> +	struct nvmetcp_spe_conn_termination nvmetcp_conn_terminate;
>   
>   	struct vf_start_ramrod_data vf_start;
>   	struct vf_stop_ramrod_data vf_stop;
> diff --git a/include/linux/qed/nvmetcp_common.h b/include/linux/qed/nvmetcp_common.h
> index e9ccfc07041d..c8836b71b866 100644
> --- a/include/linux/qed/nvmetcp_common.h
> +++ b/include/linux/qed/nvmetcp_common.h
> @@ -6,6 +6,8 @@
>   
>   #include "tcp_common.h"
>   
> +#define NVMETCP_SLOW_PATH_LAYER_CODE (6)
> +
>   /* NVMeTCP firmware function init parameters */
>   struct nvmetcp_spe_func_init {
>   	__le16 half_way_close_timeout;
> @@ -43,6 +45,10 @@ enum nvmetcp_ramrod_cmd_id {
>   	NVMETCP_RAMROD_CMD_ID_UNUSED = 0,
>   	NVMETCP_RAMROD_CMD_ID_INIT_FUNC = 1,
>   	NVMETCP_RAMROD_CMD_ID_DESTROY_FUNC = 2,
> +	NVMETCP_RAMROD_CMD_ID_OFFLOAD_CONN = 3,
> +	NVMETCP_RAMROD_CMD_ID_UPDATE_CONN = 4,
> +	NVMETCP_RAMROD_CMD_ID_TERMINATION_CONN = 5,
> +	NVMETCP_RAMROD_CMD_ID_CLEAR_SQ = 6,
>   	MAX_NVMETCP_RAMROD_CMD_ID
>   };
>   
> @@ -51,4 +57,141 @@ struct nvmetcp_glbl_queue_entry {
>   	struct regpair reserved;
>   };
>   
> +/* NVMeTCP conn level EQEs */
> +enum nvmetcp_eqe_opcode {
> +	NVMETCP_EVENT_TYPE_INIT_FUNC = 0, /* Response after init Ramrod */
> +	NVMETCP_EVENT_TYPE_DESTROY_FUNC, /* Response after destroy Ramrod */
> +	NVMETCP_EVENT_TYPE_OFFLOAD_CONN,/* Response after option 2 offload Ramrod */
> +	NVMETCP_EVENT_TYPE_UPDATE_CONN, /* Response after update Ramrod */
> +	NVMETCP_EVENT_TYPE_CLEAR_SQ, /* Response after clear sq Ramrod */
> +	NVMETCP_EVENT_TYPE_TERMINATE_CONN, /* Response after termination Ramrod */
> +	NVMETCP_EVENT_TYPE_RESERVED0,
> +	NVMETCP_EVENT_TYPE_RESERVED1,
> +	NVMETCP_EVENT_TYPE_ASYN_CONNECT_COMPLETE, /* Connect completed (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_ASYN_TERMINATE_DONE, /* Termination completed (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_START_OF_ERROR_TYPES = 10, /* Separate EQs from err EQs */
> +	NVMETCP_EVENT_TYPE_ASYN_ABORT_RCVD, /* TCP RST packet receive (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_ASYN_CLOSE_RCVD, /* TCP FIN packet receive (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_ASYN_SYN_RCVD, /* TCP SYN+ACK packet receive (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_ASYN_MAX_RT_TIME, /* TCP max retransmit time (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_ASYN_MAX_RT_CNT, /* TCP max retransmit count (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_ASYN_MAX_KA_PROBES_CNT, /* TCP ka probes count (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_ASYN_FIN_WAIT2, /* TCP fin wait 2 (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_NVMETCP_CONN_ERROR, /* NVMeTCP error response (A-syn EQE) */
> +	NVMETCP_EVENT_TYPE_TCP_CONN_ERROR, /* NVMeTCP error - tcp error (A-syn EQE) */
> +	MAX_NVMETCP_EQE_OPCODE
> +};
> +
> +struct nvmetcp_conn_offload_section {
> +	struct regpair cccid_itid_table_addr; /* CCCID to iTID table address */
> +	__le16 cccid_max_range; /* CCCID max value - used for validation */
> +	__le16 reserved[3];
> +};
> +
> +/* NVMe TCP connection offload params passed by driver to FW in NVMeTCP offload ramrod */
> +struct nvmetcp_conn_offload_params {
> +	struct regpair sq_pbl_addr;
> +	struct regpair r2tq_pbl_addr;
> +	struct regpair xhq_pbl_addr;
> +	struct regpair uhq_pbl_addr;
> +	__le16 physical_q0;
> +	__le16 physical_q1;
> +	u8 flags;
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_TCP_ON_CHIP_1B_MASK 0x1
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_TCP_ON_CHIP_1B_SHIFT 0
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_TARGET_MODE_MASK 0x1
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_TARGET_MODE_SHIFT 1
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_RESTRICTED_MODE_MASK 0x1
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_RESTRICTED_MODE_SHIFT 2
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_NVMETCP_MODE_MASK 0x1
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_NVMETCP_MODE_SHIFT 3
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_RESERVED1_MASK 0xF
> +#define NVMETCP_CONN_OFFLOAD_PARAMS_RESERVED1_SHIFT 4
> +	u8 default_cq;
> +	__le16 reserved0;
> +	__le32 reserved1;
> +	__le32 initial_ack;
> +
> +	struct nvmetcp_conn_offload_section nvmetcp; /* NVMe/TCP section */
> +};
> +
> +/* NVMe TCP and TCP connection offload params passed by driver to FW in NVMeTCP offload ramrod. */
> +struct nvmetcp_spe_conn_offload {
> +	__le16 reserved;
> +	__le16 conn_id;
> +	__le32 fw_cid;
> +	struct nvmetcp_conn_offload_params nvmetcp;
> +	struct tcp_offload_params_opt2 tcp;
> +};
> +
> +/* NVMeTCP connection update params passed by driver to FW in NVMETCP update ramrod. */
> +struct nvmetcp_conn_update_ramrod_params {
> +	__le16 reserved0;
> +	__le16 conn_id;
> +	__le32 reserved1;
> +	u8 flags;
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_HD_EN_MASK 0x1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_HD_EN_SHIFT 0
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_DD_EN_MASK 0x1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_DD_EN_SHIFT 1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED0_MASK 0x1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED0_SHIFT 2
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED1_MASK 0x1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED1_DATA_SHIFT 3
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED2_MASK 0x1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED2_SHIFT 4
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED3_MASK 0x1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED3_SHIFT 5
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED4_MASK 0x1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED4_SHIFT 6
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED5_MASK 0x1
> +#define NVMETCP_CONN_UPDATE_RAMROD_PARAMS_RESERVED5_SHIFT 7
> +	u8 reserved3[3];
> +	__le32 max_seq_size;
> +	__le32 max_send_pdu_length;
> +	__le32 max_recv_pdu_length;
> +	__le32 first_seq_length;
> +	__le32 reserved4[5];
> +};
> +
> +/* NVMeTCP connection termination request */
> +struct nvmetcp_spe_conn_termination {
> +	__le16 reserved0;
> +	__le16 conn_id;
> +	__le32 reserved1;
> +	u8 abortive;
> +	u8 reserved2[7];
> +	struct regpair reserved3;
> +	struct regpair reserved4;
> +};
> +
> +struct nvmetcp_dif_flags {
> +	u8 flags;
> +};
> +
> +enum nvmetcp_wqe_type {
> +	NVMETCP_WQE_TYPE_NORMAL,
> +	NVMETCP_WQE_TYPE_TASK_CLEANUP,
> +	NVMETCP_WQE_TYPE_MIDDLE_PATH,
> +	NVMETCP_WQE_TYPE_IC,
> +	MAX_NVMETCP_WQE_TYPE
> +};
> +
> +struct nvmetcp_wqe {
> +	__le16 task_id;
> +	u8 flags;
> +#define NVMETCP_WQE_WQE_TYPE_MASK 0x7 /* [use nvmetcp_wqe_type] */
> +#define NVMETCP_WQE_WQE_TYPE_SHIFT 0
> +#define NVMETCP_WQE_NUM_SGES_MASK 0xF
> +#define NVMETCP_WQE_NUM_SGES_SHIFT 3
> +#define NVMETCP_WQE_RESPONSE_MASK 0x1
> +#define NVMETCP_WQE_RESPONSE_SHIFT 7
> +	struct nvmetcp_dif_flags prot_flags;
> +	__le32 contlen_cdbsize;
> +#define NVMETCP_WQE_CONT_LEN_MASK 0xFFFFFF
> +#define NVMETCP_WQE_CONT_LEN_SHIFT 0
> +#define NVMETCP_WQE_CDB_SIZE_OR_NVMETCP_CMD_MASK 0xFF
> +#define NVMETCP_WQE_CDB_SIZE_OR_NVMETCP_CMD_SHIFT 24
> +};
> +
>   #endif /* __NVMETCP_COMMON__ */
> diff --git a/include/linux/qed/qed_nvmetcp_if.h b/include/linux/qed/qed_nvmetcp_if.h
> index abc1f41862e3..96263e3cfa1e 100644
> --- a/include/linux/qed/qed_nvmetcp_if.h
> +++ b/include/linux/qed/qed_nvmetcp_if.h
> @@ -25,6 +25,50 @@ struct qed_nvmetcp_tid {
>   	u8 *blocks[MAX_TID_BLOCKS_NVMETCP];
>   };
>   
> +struct qed_nvmetcp_id_params {
> +	u8 mac[ETH_ALEN];
> +	u32 ip[4];
> +	u16 port;
> +};
> +
> +struct qed_nvmetcp_params_offload {
> +	/* FW initializations */
> +	dma_addr_t sq_pbl_addr;
> +	dma_addr_t nvmetcp_cccid_itid_table_addr;
> +	u16 nvmetcp_cccid_max_range;
> +	u8 default_cq;
> +
> +	/* Networking and TCP stack initializations */
> +	struct qed_nvmetcp_id_params src;
> +	struct qed_nvmetcp_id_params dst;
> +	u32 ka_timeout;
> +	u32 ka_interval;
> +	u32 max_rt_time;
> +	u32 cwnd;
> +	u16 mss;
> +	u16 vlan_id;
> +	bool timestamp_en;
> +	bool delayed_ack_en;
> +	bool tcp_keep_alive_en;
> +	bool ecn_en;
> +	u8 ip_version;
> +	u8 ka_max_probe_cnt;
> +	u8 ttl;
> +	u8 tos_or_tc;
> +	u8 rcv_wnd_scale;
> +};
> +
> +struct qed_nvmetcp_params_update {
> +	u32 max_io_size;
> +	u32 max_recv_pdu_length;
> +	u32 max_send_pdu_length;
> +
> +	/* Placeholder: pfv, cpda, hpda */
> +
> +	bool hdr_digest_en;
> +	bool data_digest_en;
> +};
> +
>   struct qed_nvmetcp_cb_ops {
>   	struct qed_common_cb_ops common;
>   };
> @@ -48,6 +92,38 @@ struct qed_nvmetcp_cb_ops {
>    * @stop:		nvmetcp in FW
>    *			@param cdev
>    *			return 0 on success, otherwise error value.
> + * @acquire_conn:	acquire a new nvmetcp connection
> + *			@param cdev
> + *			@param handle - qed will fill handle that should be
> + *				used henceforth as identifier of the
> + *				connection.
> + *			@param p_doorbell - qed will fill the address of the
> + *				doorbell.
> + *			@return 0 on sucesss, otherwise error value.
> + * @release_conn:	release a previously acquired nvmetcp connection
> + *			@param cdev
> + *			@param handle - the connection handle.
> + *			@return 0 on success, otherwise error value.
> + * @offload_conn:	configures an offloaded connection
> + *			@param cdev
> + *			@param handle - the connection handle.
> + *			@param conn_info - the configuration to use for the
> + *				offload.
> + *			@return 0 on success, otherwise error value.
> + * @update_conn:	updates an offloaded connection
> + *			@param cdev
> + *			@param handle - the connection handle.
> + *			@param conn_info - the configuration to use for the
> + *				offload.
> + *			@return 0 on success, otherwise error value.
> + * @destroy_conn:	stops an offloaded connection
> + *			@param cdev
> + *			@param handle - the connection handle.
> + *			@return 0 on success, otherwise error value.
> + * @clear_sq:		clear all task in sq
> + *			@param cdev
> + *			@param handle - the connection handle.
> + *			@return 0 on success, otherwise error value.
>    */
>   struct qed_nvmetcp_ops {
>   	const struct qed_common_ops *common;
> @@ -65,6 +141,24 @@ struct qed_nvmetcp_ops {
>   		     void *event_context, nvmetcp_event_cb_t async_event_cb);
>   
>   	int (*stop)(struct qed_dev *cdev);
> +
> +	int (*acquire_conn)(struct qed_dev *cdev,
> +			    u32 *handle,
> +			    u32 *fw_cid, void __iomem **p_doorbell);
> +
> +	int (*release_conn)(struct qed_dev *cdev, u32 handle);
> +
> +	int (*offload_conn)(struct qed_dev *cdev,
> +			    u32 handle,
> +			    struct qed_nvmetcp_params_offload *conn_info);
> +
> +	int (*update_conn)(struct qed_dev *cdev,
> +			   u32 handle,
> +			   struct qed_nvmetcp_params_update *conn_info);
> +
> +	int (*destroy_conn)(struct qed_dev *cdev, u32 handle, u8 abrt_conn);
> +
> +	int (*clear_sq)(struct qed_dev *cdev, u32 handle);
>   };
>   
>   const struct qed_nvmetcp_ops *qed_get_nvmetcp_ops(void);
> 
Cheers,

Hannes
-- 
Dr. Hannes Reinecke                Kernel Storage Architect
hare@suse.de                              +49 911 74053 688
SUSE Software Solutions GmbH, Maxfeldstr. 5, 90409 Nürnberg
HRB 36809 (AG Nürnberg), Geschäftsführer: Felix Imendörffer

  reply	other threads:[~2021-05-01 17:28 UTC|newest]

Thread overview: 82+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-04-29 19:08 [RFC PATCH v4 00/27] NVMeTCP Offload ULP and QEDN Device Driver Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 01/27] qed: Add NVMeTCP Offload PF Level FW and HW HSI Shai Malin
2021-05-01 16:50   ` Hannes Reinecke
2021-05-03 15:23     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 02/27] qed: Add NVMeTCP Offload Connection " Shai Malin
2021-05-01 17:28   ` Hannes Reinecke [this message]
2021-05-03 15:25     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 03/27] qed: Add qed-NVMeTCP personality Shai Malin
2021-05-02 11:11   ` Hannes Reinecke
2021-05-03 15:26     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 04/27] qed: Add support of HW filter block Shai Malin
2021-05-02 11:13   ` Hannes Reinecke
2021-05-03 15:27     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 05/27] qed: Add NVMeTCP Offload IO Level FW and HW HSI Shai Malin
2021-05-02 11:22   ` Hannes Reinecke
2021-05-04 16:25     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 06/27] qed: Add NVMeTCP Offload IO Level FW Initializations Shai Malin
2021-05-02 11:24   ` Hannes Reinecke
2021-05-04 16:28     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 07/27] qed: Add IP services APIs support Shai Malin
2021-05-02 11:26   ` Hannes Reinecke
2021-05-03 15:44     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 08/27] nvme-tcp-offload: Add nvme-tcp-offload - NVMeTCP HW offload ULP Shai Malin
2021-05-01 12:18   ` Hannes Reinecke
2021-05-03 15:46     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 09/27] nvme-fabrics: Move NVMF_ALLOWED_OPTS and NVMF_REQUIRED_OPTS definitions Shai Malin
2021-05-01 12:19   ` Hannes Reinecke
2021-05-03 15:50     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 10/27] nvme-tcp-offload: Add device scan implementation Shai Malin
2021-05-01 12:25   ` Hannes Reinecke
2021-05-05 17:52     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 11/27] nvme-tcp-offload: Add controller level implementation Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 12/27] nvme-tcp-offload: Add controller level error recovery implementation Shai Malin
2021-05-01 16:29   ` Hannes Reinecke
2021-05-03 15:52     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 13/27] nvme-tcp-offload: Add queue level implementation Shai Malin
2021-05-01 16:36   ` Hannes Reinecke
2021-05-03 15:56     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 14/27] nvme-tcp-offload: Add IO " Shai Malin
2021-05-01 16:38   ` Hannes Reinecke
2021-05-04 16:34     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 15/27] nvme-tcp-offload: Add Timeout and ASYNC Support Shai Malin
2021-05-01 16:45   ` Hannes Reinecke
2021-05-04 16:49     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 16/27] qedn: Add qedn - Marvell's NVMeTCP HW offload vendor driver Shai Malin
2021-05-02 11:27   ` Hannes Reinecke
2021-05-04 16:52     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 17/27] qedn: Add qedn probe Shai Malin
2021-05-02 11:28   ` Hannes Reinecke
2021-05-08 11:01     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 18/27] qedn: Add qedn_claim_dev API support Shai Malin
2021-05-02 11:29   ` Hannes Reinecke
2021-05-07 13:57     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 19/27] qedn: Add IRQ and fast-path resources initializations Shai Malin
2021-05-02 11:32   ` Hannes Reinecke
2021-05-05 17:54     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 20/27] qedn: Add connection-level slowpath functionality Shai Malin
2021-05-02 11:37   ` Hannes Reinecke
2021-05-05 17:56     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 21/27] qedn: Add support of configuring HW filter block Shai Malin
2021-05-02 11:38   ` Hannes Reinecke
2021-05-05 17:57     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 22/27] qedn: Add IO level nvme_req and fw_cq workqueues Shai Malin
2021-05-02 11:42   ` Hannes Reinecke
2021-05-07 13:56     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 23/27] qedn: Add support of Task and SGL Shai Malin
2021-05-02 11:48   ` Hannes Reinecke
2021-05-07 14:00     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 24/27] qedn: Add support of NVME ICReq & ICResp Shai Malin
2021-05-02 11:53   ` Hannes Reinecke
2021-05-05 18:01     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 25/27] qedn: Add IO level fastpath functionality Shai Malin
2021-05-02 11:54   ` Hannes Reinecke
2021-05-05 18:04     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 26/27] qedn: Add Connection and IO level recovery flows Shai Malin
2021-05-02 11:57   ` Hannes Reinecke
2021-05-05 18:06     ` Shai Malin
2021-04-29 19:09 ` [RFC PATCH v4 27/27] qedn: Add support of ASYNC Shai Malin
2021-05-02 11:59   ` Hannes Reinecke
2021-05-05 18:08     ` Shai Malin
2021-05-01 16:47 ` [RFC PATCH v4 00/27] NVMeTCP Offload ULP and QEDN Device Driver Hannes Reinecke
2021-05-03 15:13   ` Shai Malin

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=c35e89df-e9b7-fd45-9ad6-bf0f6be1dc5d@suse.de \
    --to=hare@suse.de \
    --cc=aelior@marvell.com \
    --cc=axboe@fb.com \
    --cc=hch@lst.de \
    --cc=kbusch@kernel.org \
    --cc=kuba@kernel.org \
    --cc=linux-nvme@lists.infradead.org \
    --cc=malin1024@gmail.com \
    --cc=mkalderon@marvell.com \
    --cc=netdev@vger.kernel.org \
    --cc=okulkarni@marvell.com \
    --cc=pkushwaha@marvell.com \
    --cc=sagi@grimberg.me \
    --cc=smalin@marvell.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).