From: James Smart <jsmart2021@gmail.com>
To: linux-scsi@vger.kernel.org
Cc: maier@linux.ibm.com, dwagner@suse.de, bvanassche@acm.org,
James Smart <jsmart2021@gmail.com>,
Ram Vegesna <ram.vegesna@broadcom.com>
Subject: [PATCH v2 16/32] elx: efct: Driver initialization routines
Date: Fri, 20 Dec 2019 14:37:07 -0800 [thread overview]
Message-ID: <20191220223723.26563-17-jsmart2021@gmail.com> (raw)
In-Reply-To: <20191220223723.26563-1-jsmart2021@gmail.com>
This patch continues the efct driver population.
This patch adds driver definitions for:
Emulex FC Target driver init, attach and hardware setup routines.
Signed-off-by: Ram Vegesna <ram.vegesna@broadcom.com>
Signed-off-by: James Smart <jsmart2021@gmail.com>
---
drivers/scsi/elx/efct/efct_driver.c | 1031 +++++++++++++++++++++++++++++
drivers/scsi/elx/efct/efct_driver.h | 150 +++++
drivers/scsi/elx/efct/efct_hw.c | 1222 +++++++++++++++++++++++++++++++++++
drivers/scsi/elx/efct/efct_hw.h | 16 +-
drivers/scsi/elx/efct/efct_xport.c | 587 +++++++++++++++++
drivers/scsi/elx/efct/efct_xport.h | 205 ++++++
6 files changed, 3210 insertions(+), 1 deletion(-)
create mode 100644 drivers/scsi/elx/efct/efct_driver.c
create mode 100644 drivers/scsi/elx/efct/efct_driver.h
create mode 100644 drivers/scsi/elx/efct/efct_hw.c
create mode 100644 drivers/scsi/elx/efct/efct_xport.c
create mode 100644 drivers/scsi/elx/efct/efct_xport.h
diff --git a/drivers/scsi/elx/efct/efct_driver.c b/drivers/scsi/elx/efct/efct_driver.c
new file mode 100644
index 000000000000..f0ec132bdd0e
--- /dev/null
+++ b/drivers/scsi/elx/efct/efct_driver.c
@@ -0,0 +1,1031 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 Broadcom. All Rights Reserved. The term
+ * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
+ */
+
+#include "efct_driver.h"
+#include "efct_utils.h"
+
+#include "efct_els.h"
+#include "efct_hw.h"
+#include "efct_unsol.h"
+#include "efct_scsi.h"
+
+struct efct *efct_devices[MAX_EFCT_DEVICES];
+
+static int logmask;
+module_param(logmask, int, 0444);
+MODULE_PARM_DESC(logmask, "logging bitmask (default 0)");
+
+static struct libefc_function_template efct_libefc_templ = {
+ .hw_domain_alloc = efct_hw_domain_alloc,
+ .hw_domain_attach = efct_hw_domain_attach,
+ .hw_domain_free = efct_hw_domain_free,
+ .hw_domain_force_free = efct_hw_domain_force_free,
+ .domain_hold_frames = efct_domain_hold_frames,
+ .domain_accept_frames = efct_domain_accept_frames,
+
+ .hw_port_alloc = efct_hw_port_alloc,
+ .hw_port_attach = efct_hw_port_attach,
+ .hw_port_free = efct_hw_port_free,
+
+ .hw_node_alloc = efct_hw_node_alloc,
+ .hw_node_attach = efct_hw_node_attach,
+ .hw_node_detach = efct_hw_node_detach,
+ .hw_node_free_resources = efct_hw_node_free_resources,
+ .node_purge_pending = efct_node_purge_pending,
+
+ .scsi_io_alloc_disable = efct_scsi_io_alloc_disable,
+ .scsi_io_alloc_enable = efct_scsi_io_alloc_enable,
+ .scsi_validate_node = efct_scsi_validate_initiator,
+ .new_domain = efct_scsi_tgt_new_domain,
+ .del_domain = efct_scsi_tgt_del_domain,
+ .new_sport = efct_scsi_tgt_new_sport,
+ .del_sport = efct_scsi_tgt_del_sport,
+ .scsi_new_node = efct_scsi_new_initiator,
+ .scsi_del_node = efct_scsi_del_initiator,
+
+ .els_send = efct_els_req_send,
+ .els_send_ct = efct_els_send_ct,
+ .els_send_resp = efct_els_resp_send,
+ .bls_send_acc_hdr = efct_bls_send_acc_hdr,
+ .send_flogi_p2p_acc = efct_send_flogi_p2p_acc,
+ .send_ct_rsp = efct_send_ct_rsp,
+ .send_ls_rjt = efct_send_ls_rjt,
+
+ .node_io_cleanup = efct_node_io_cleanup,
+ .node_els_cleanup = efct_node_els_cleanup,
+ .node_abort_all_els = efct_node_abort_all_els,
+
+ .dispatch_fcp_cmd = efct_dispatch_fcp_cmd,
+ .recv_abts_frame = efct_node_recv_abts_frame,
+};
+
+static char *queue_topology =
+ "eq cq rq cq mq $nulp($nwq(cq wq:ulp=$rpt1)) cq wq:len=256:class=1";
+
+static int
+efct_device_init(void)
+{
+ int rc;
+
+ hw_global.queue_topology_string = queue_topology;
+
+ /* driver-wide init for target-server */
+ rc = efct_scsi_tgt_driver_init();
+ if (rc) {
+ pr_err("efct_scsi_tgt_init failed rc=%d\n",
+ rc);
+ return -1;
+ }
+
+ rc = efct_scsi_reg_fc_transport();
+ if (rc) {
+ pr_err("failed to register to FC host\n");
+ return -1;
+ }
+
+ return 0;
+}
+
+static void
+efct_device_shutdown(void)
+{
+ efct_scsi_release_fc_transport();
+
+ efct_scsi_tgt_driver_exit();
+}
+
+static void *
+efct_device_alloc(u32 nid)
+{
+ struct efct *efct = NULL;
+ u32 i;
+
+ efct = kmalloc_node(sizeof(*efct), GFP_ATOMIC, nid);
+
+ if (efct) {
+ memset(efct, 0, sizeof(*efct));
+ for (i = 0; i < ARRAY_SIZE(efct_devices); i++) {
+ if (!efct_devices[i]) {
+ efct->instance_index = i;
+ efct_devices[i] = efct;
+ break;
+ }
+ }
+
+ if (i == ARRAY_SIZE(efct_devices)) {
+ pr_err("Exceeded max supported devices.\n");
+ kfree(efct);
+ efct = NULL;
+ } else {
+ efct->attached = false;
+ }
+ }
+ return efct;
+}
+
+static struct efct *
+efct_get_instance(u32 index)
+{
+ if (index < ARRAY_SIZE(efct_devices))
+ return efct_devices[index];
+
+ return NULL;
+}
+
+static void
+efct_device_interrupt_handler(struct efct *efct, u32 vector)
+{
+ efct_hw_process(&efct->hw, vector, efct->max_isr_time_msec);
+}
+
+static int
+efct_intr_thread(struct efct_intr_context *intr_context)
+{
+ struct efct *efct = intr_context->efct;
+ int rc;
+ u32 tstart, tnow;
+
+ tstart = jiffies_to_msecs(jiffies);
+
+ while (!kthread_should_stop()) {
+ rc = wait_for_completion_timeout(&intr_context->done,
+ usecs_to_jiffies(100000));
+ if (!rc)
+ continue;
+
+ efct_device_interrupt_handler(efct, intr_context->index);
+
+ /* If we've been running for too long, then yield */
+ tnow = jiffies_to_msecs(jiffies);
+ if ((tnow - tstart) > 5000) {
+ cond_resched();
+ tstart = tnow;
+ }
+ }
+
+ return 0;
+}
+
+static int
+efct_start_event_processing(struct efct *efct)
+{
+ u32 i;
+
+ for (i = 0; i < efct->n_msix_vec; i++) {
+ char label[32];
+ struct efct_intr_context *intr_ctx = NULL;
+
+ intr_ctx = &efct->intr_context[i];
+
+ intr_ctx->efct = efct;
+ intr_ctx->index = i;
+
+ init_completion(&intr_ctx->done);
+
+ snprintf(label, sizeof(label),
+ "efct:%d:%d", efct->instance_index, i);
+
+ intr_ctx->thread =
+ kthread_create((int(*)(void *)) efct_intr_thread,
+ intr_ctx, label);
+
+ if (IS_ERR(intr_ctx->thread)) {
+ efc_log_err(efct, "kthread_create failed: %ld\n",
+ PTR_ERR(intr_ctx->thread));
+ intr_ctx->thread = NULL;
+
+ return -1;
+ }
+
+ wake_up_process(intr_ctx->thread);
+ }
+
+ return 0;
+}
+
+static void
+efct_teardown_msix(struct efct *efct)
+{
+ u32 i;
+
+ for (i = 0; i < efct->n_msix_vec; i++) {
+ synchronize_irq(efct->msix_vec[i].vector);
+ free_irq(efct->msix_vec[i].vector,
+ &efct->intr_context[i]);
+ }
+ pci_disable_msix(efct->pcidev);
+}
+
+static int
+efct_efclib_config(struct efct *efct, struct libefc_function_template *tt)
+{
+ struct efc *efc;
+ struct sli4 *sli;
+
+ efc = kmalloc(sizeof(*efc), GFP_KERNEL);
+ if (!efc)
+ return -1;
+
+ memset(efc, 0, sizeof(struct efc));
+ efct->efcport = efc;
+
+ memcpy(&efc->tt, tt, sizeof(*tt));
+ efc->base = efct;
+ efc->pcidev = efct->pcidev;
+
+ efc->def_wwnn = efct_get_wwn(&efct->hw, EFCT_HW_WWN_NODE);
+ efc->def_wwpn = efct_get_wwn(&efct->hw, EFCT_HW_WWN_PORT);
+ efc->enable_tgt = 1;
+ efc->log_level = EFC_LOG_LIB;
+
+ sli = &efct->hw.sli;
+ efc->max_xfer_size = sli->sge_supported_length *
+ sli_get_max_sgl(&efct->hw.sli);
+
+ efcport_init(efc);
+
+ return 0;
+}
+
+static int efct_request_firmware_update(struct efct *efct);
+
+static int
+efct_device_attach(struct efct *efct)
+{
+ u32 rc = 0, i = 0;
+
+ if (efct->attached) {
+ efc_log_warn(efct, "Device is already attached\n");
+ rc = -1;
+ } else {
+ snprintf(efct->display_name, sizeof(efct->display_name),
+ "[%s%d] ", "fc", efct->instance_index);
+
+ efct->logmask = logmask;
+ efct->enable_numa_support = 1;
+ efct->filter_def = "0,0,0,0";
+ efct->max_isr_time_msec = EFCT_OS_MAX_ISR_TIME_MSEC;
+ efct->model =
+ (efct->pcidev->device == EFCT_DEVICE_ID_LPE31004) ?
+ "LPE31004" : "unknown";
+ efct->fw_version = (const char *)efct_hw_get_ptr(&efct->hw,
+ EFCT_HW_FW_REV);
+ efct->driver_version = EFCT_DRIVER_VERSION;
+
+ efct->efct_req_fw_upgrade = true;
+
+ /* Allocate transport object and bring online */
+ efct->xport = efct_xport_alloc(efct);
+ if (!efct->xport) {
+ efc_log_err(efct, "failed to allocate transport object\n");
+ rc = -1;
+ } else if (efct_xport_attach(efct->xport) != 0) {
+ efc_log_err(efct, "failed to attach transport object\n");
+ rc = -1;
+ } else if (efct_xport_initialize(efct->xport) != 0) {
+ efc_log_err(efct, "failed to initialize transport object\n");
+ rc = -1;
+ } else if (efct_efclib_config(efct, &efct_libefc_templ)) {
+ efc_log_err(efct, "failed to init efclib\n");
+ rc = -1;
+ } else if (efct_start_event_processing(efct)) {
+ efc_log_err(efct, "failed to start event processing\n");
+ rc = -1;
+ } else {
+ for (i = 0; i < efct->n_msix_vec; i++) {
+ efc_log_debug(efct, "irq %d enabled\n",
+ efct->msix_vec[i].vector);
+ enable_irq(efct->msix_vec[i].vector);
+ }
+ }
+
+ efct->desc = efct->hw.sli.modeldesc;
+ efc_log_info(efct, "adapter model description: %s\n",
+ efct->desc);
+
+ if (rc == 0) {
+ efct->attached = true;
+ } else {
+ efct_teardown_msix(efct);
+ if (efct->xport) {
+ efct_xport_free(efct->xport);
+ efct->xport = NULL;
+ }
+ }
+
+ if (efct->efct_req_fw_upgrade) {
+ efc_log_debug(efct, "firmware update is in progress\n");
+ efct_request_firmware_update(efct);
+ }
+ }
+
+ return rc;
+}
+
+static void
+efct_stop_event_processing(struct efct *efct)
+{
+ u32 i;
+ struct task_struct *thread = NULL;
+
+ for (i = 0; i < efct->n_msix_vec; i++) {
+ disable_irq(efct->msix_vec[i].vector);
+
+ thread = efct->intr_context[i].thread;
+ if (!thread)
+ continue;
+
+ /* Call stop */
+ kthread_stop(thread);
+ }
+}
+
+static int
+efct_device_detach(struct efct *efct)
+{
+ int rc = 0;
+
+ if (efct) {
+ if (!efct->attached) {
+ efc_log_warn(efct, "Device is not attached\n");
+ return -1;
+ }
+
+ rc = efct_xport_control(efct->xport, EFCT_XPORT_SHUTDOWN);
+ if (rc)
+ efc_log_err(efct, "Transport Shutdown timed out\n");
+
+ efct_stop_event_processing(efct);
+
+ if (efct_xport_detach(efct->xport) != 0)
+ efc_log_err(efct, "Transport detach failed\n");
+
+ efct_xport_free(efct->xport);
+ efct->xport = NULL;
+
+ efcport_destroy(efct->efcport);
+ kfree(efct->efcport);
+
+ efct->attached = false;
+ }
+
+ return 0;
+}
+
+static int
+efct_fw_reset(struct efct *efct)
+{
+ int rc = 0;
+ int index = 0;
+ u8 bus, dev;
+ struct efct *other_efct;
+
+ bus = efct->pcidev->bus->number;
+ dev = PCI_SLOT(efct->pcidev->devfn);
+
+ while ((other_efct = efct_get_instance(index++)) != NULL) {
+ u8 other_bus, other_dev;
+
+ other_bus = other_efct->pcidev->bus->number;
+ other_dev = PCI_SLOT(other_efct->pcidev->devfn);
+
+ if (bus == other_bus && dev == other_dev &&
+ timer_pending(&other_efct->xport->stats_timer)) {
+ efc_log_debug(other_efct,
+ "removing link stats timer\n");
+ del_timer(&other_efct->xport->stats_timer);
+ }
+ }
+
+ if (efct_hw_reset(&efct->hw, EFCT_HW_RESET_FIRMWARE)) {
+ efc_log_test(efct, "failed to reset firmware\n");
+ rc = -1;
+ } else {
+ efc_log_debug(efct,
+ "successfully reset firmware.Now resetting port\n");
+ /* now flag all functions on the same device
+ * as this port as uninitialized
+ */
+ index = 0;
+
+ while ((other_efct = efct_get_instance(index++)) != NULL) {
+ u8 other_bus, other_dev;
+
+ other_bus = other_efct->pcidev->bus->number;
+ other_dev = PCI_SLOT(other_efct->pcidev->devfn);
+
+ if (bus == other_bus && dev == other_dev) {
+ if (other_efct->hw.state !=
+ EFCT_HW_STATE_UNINITIALIZED) {
+ other_efct->hw.state =
+ EFCT_HW_STATE_QUEUES_ALLOCATED;
+ }
+ efct_device_detach(efct);
+ rc = efct_device_attach(efct);
+
+ efc_log_debug(other_efct,
+ "re-start driver with new firmware\n");
+ }
+ }
+ }
+ return rc;
+}
+
+static void
+efct_fw_write_cb(int status, u32 actual_write_length,
+ u32 change_status, void *arg)
+{
+ struct efct_fw_write_result *result = arg;
+
+ result->status = status;
+ result->actual_xfer = actual_write_length;
+ result->change_status = change_status;
+
+ complete(&result->done);
+}
+
+static int
+efct_firmware_write(struct efct *efct, const u8 *buf, size_t buf_len,
+ u8 *change_status)
+{
+ int rc = 0;
+ u32 bytes_left;
+ u32 xfer_size;
+ u32 offset;
+ struct efc_dma dma;
+ int last = 0;
+ struct efct_fw_write_result result;
+
+ init_completion(&result.done);
+
+ bytes_left = buf_len;
+ offset = 0;
+
+ dma.size = FW_WRITE_BUFSIZE;
+ dma.virt = dma_alloc_coherent(&efct->pcidev->dev,
+ dma.size, &dma.phys, GFP_DMA);
+ if (!dma.virt)
+ return -ENOMEM;
+
+ while (bytes_left > 0) {
+ if (bytes_left > FW_WRITE_BUFSIZE)
+ xfer_size = FW_WRITE_BUFSIZE;
+ else
+ xfer_size = bytes_left;
+
+ memcpy(dma.virt, buf + offset, xfer_size);
+
+ if (bytes_left == xfer_size)
+ last = 1;
+
+ efct_hw_firmware_write(&efct->hw, &dma, xfer_size, offset,
+ last, efct_fw_write_cb, &result);
+
+ if (wait_for_completion_interruptible(&result.done) != 0) {
+ rc = -ENXIO;
+ break;
+ }
+
+ if (result.actual_xfer == 0 || result.status != 0) {
+ rc = -EFAULT;
+ break;
+ }
+
+ if (last)
+ *change_status = result.change_status;
+
+ bytes_left -= result.actual_xfer;
+ offset += result.actual_xfer;
+ }
+
+ dma_free_coherent(&efct->pcidev->dev, dma.size, dma.virt, dma.phys);
+ return rc;
+}
+
+static int
+efct_request_firmware_update(struct efct *efct)
+{
+ int rc = 0;
+ u8 file_name[256], fw_change_status = 0;
+ const struct firmware *fw;
+ struct efct_hw_grp_hdr *fw_image;
+
+ snprintf(file_name, 256, "%s.grp", efct->model);
+ rc = request_firmware(&fw, file_name, &efct->pcidev->dev);
+ if (rc) {
+ efc_log_err(efct, "Firmware file(%s) not found.\n", file_name);
+ return rc;
+ }
+ fw_image = (struct efct_hw_grp_hdr *)fw->data;
+
+ /* Check if firmware provided is compatible with this particular
+ * Adapter of not
+ */
+ if ((be32_to_cpu(fw_image->magic_number) != EFCT_HW_OBJECT_G5) &&
+ (be32_to_cpu(fw_image->magic_number) != EFCT_HW_OBJECT_G6)) {
+ efc_log_warn(efct,
+ "Invalid FW image found Magic: 0x%x Size: %ld\n",
+ be32_to_cpu(fw_image->magic_number), fw->size);
+ rc = -1;
+ goto exit;
+ }
+
+ if (!strncmp(efct->fw_version, fw_image->revision,
+ strnlen(fw_image->revision, 16))) {
+ efc_log_debug(efct,
+ "No update req. Firmware is already up to date.\n");
+ rc = 0;
+ goto exit;
+ }
+ rc = efct_firmware_write(efct, fw->data, fw->size, &fw_change_status);
+ if (rc) {
+ efc_log_err(efct,
+ "Firmware update failed. Return code = %d\n", rc);
+ } else {
+ efc_log_info(efct, "Firmware updated successfully\n");
+ switch (fw_change_status) {
+ case 0x00:
+ efc_log_debug(efct,
+ "No reset needed, new firmware is active.\n");
+ break;
+ case 0x01:
+ efc_log_warn(efct,
+ "A physical device reset (host reboot) is needed to activate the new firmware\n");
+ break;
+ case 0x02:
+ case 0x03:
+ efc_log_debug(efct,
+ "firmware is resetting to activate the new firmware\n");
+ efct_fw_reset(efct);
+ break;
+ default:
+ efc_log_debug(efct,
+ "Unexected value change_status: %d\n",
+ fw_change_status);
+ break;
+ }
+ }
+
+exit:
+ release_firmware(fw);
+
+ return rc;
+}
+
+static void
+efct_device_free(struct efct *efct)
+{
+ if (efct) {
+ efct_devices[efct->instance_index] = NULL;
+
+ kfree(efct);
+ }
+}
+
+static int
+efct_device_interrupts_required(struct efct *efct)
+{
+ if (efct_hw_setup(&efct->hw, efct, efct->pcidev)
+ != EFCT_HW_RTN_SUCCESS) {
+ return -1;
+ }
+ return efct_hw_qtop_eq_count(&efct->hw);
+}
+
+static irqreturn_t
+efct_intr_msix(int irq, void *handle)
+{
+ struct efct_intr_context *intr_context = handle;
+
+ complete(&intr_context->done);
+ return IRQ_HANDLED;
+}
+
+static int
+efct_setup_msix(struct efct *efct, u32 num_interrupts)
+{
+ int rc = 0;
+ u32 i;
+
+ if (!pci_find_capability(efct->pcidev, PCI_CAP_ID_MSIX)) {
+ dev_err(&efct->pcidev->dev,
+ "%s : MSI-X not available\n", __func__);
+ return -EINVAL;
+ }
+
+ if (num_interrupts > ARRAY_SIZE(efct->msix_vec)) {
+ dev_err(&efct->pcidev->dev,
+ "%s : num_interrupts: %d greater than vectors\n",
+ __func__, num_interrupts);
+ return -1;
+ }
+
+ efct->n_msix_vec = num_interrupts;
+ for (i = 0; i < num_interrupts; i++)
+ efct->msix_vec[i].entry = i;
+
+ rc = pci_enable_msix_exact(efct->pcidev,
+ efct->msix_vec, efct->n_msix_vec);
+ if (!rc) {
+ for (i = 0; i < num_interrupts; i++) {
+ rc = request_irq(efct->msix_vec[i].vector,
+ efct_intr_msix,
+ 0, EFCT_DRIVER_NAME,
+ &efct->intr_context[i]);
+ if (rc)
+ break;
+ }
+ } else {
+ dev_err(&efct->pcidev->dev,
+ "%s : rc % d returned, IRQ allocation failed\n",
+ __func__, rc);
+ }
+
+ return rc;
+}
+
+static struct pci_device_id efct_pci_table[] = {
+ {PCI_DEVICE(EFCT_VENDOR_ID, EFCT_DEVICE_ID_LPE31004), 0},
+ {PCI_DEVICE(EFCT_VENDOR_ID, EFCT_DEVICE_ID_G7), 0},
+ {} /* terminate list */
+};
+
+static int
+efct_pci_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
+{
+ struct efct *efct = NULL;
+ int rc;
+ u32 i, r;
+ int num_interrupts = 0;
+ int nid;
+ struct task_struct *thread = NULL;
+
+ dev_info(&pdev->dev, "%s\n", EFCT_DRIVER_NAME);
+
+ rc = pci_enable_device_mem(pdev);
+ if (rc)
+ goto efct_pci_probe_err_enable;
+
+ pci_set_master(pdev);
+
+ rc = pci_set_mwi(pdev);
+ if (rc) {
+ dev_info(&pdev->dev,
+ "pci_set_mwi returned %d\n", rc);
+ goto efct_pci_probe_err_set_mwi;
+ }
+
+ rc = pci_request_regions(pdev, EFCT_DRIVER_NAME);
+ if (rc) {
+ dev_err(&pdev->dev, "pci_request_regions failed\n");
+ goto efct_pci_probe_err_request_regions;
+ }
+
+ /* Fetch the Numa node id for this device */
+ nid = dev_to_node(&pdev->dev);
+ if (nid < 0) {
+ dev_err(&pdev->dev,
+ "Warning Numa node ID is %d\n", nid);
+ nid = 0;
+ }
+
+ /* Allocate efct */
+ efct = efct_device_alloc(nid);
+ if (!efct) {
+ dev_err(&pdev->dev, "Failed to allocate efct_t\n");
+ rc = -ENOMEM;
+ goto efct_pci_probe_err_efct_device_alloc;
+ }
+
+ efct->pcidev = pdev;
+
+ if (efct->enable_numa_support)
+ efct->numa_node = nid;
+
+ /* Map all memory BARs */
+ for (i = 0, r = 0; i < EFCT_PCI_MAX_REGS; i++) {
+ if (pci_resource_flags(pdev, i) & IORESOURCE_MEM) {
+ efct->reg[r] = ioremap(pci_resource_start(pdev, i),
+ pci_resource_len(pdev, i));
+ r++;
+ }
+
+ /*
+ * If the 64-bit attribute is set, both this BAR and the
+ * next form the complete address. Skip processing the
+ * next BAR.
+ */
+ if (pci_resource_flags(pdev, i) & IORESOURCE_MEM_64)
+ i++;
+ }
+
+ pci_set_drvdata(pdev, efct);
+
+ if (pci_set_dma_mask(pdev, DMA_BIT_MASK(64)) != 0 ||
+ pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64)) != 0) {
+ dev_warn(&pdev->dev,
+ "trying DMA_BIT_MASK(32)\n");
+ if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32)) != 0 ||
+ pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)) != 0) {
+ dev_err(&pdev->dev,
+ "setting DMA_BIT_MASK failed\n");
+ rc = -1;
+ goto efct_pci_probe_err_setup_thread;
+ }
+ }
+
+ num_interrupts = efct_device_interrupts_required(efct);
+ if (num_interrupts < 0) {
+ efc_log_err(efct, "efct_device_interrupts_required failed\n");
+ rc = -1;
+ goto efct_pci_probe_err_setup_thread;
+ }
+
+ /*
+ * Initialize MSIX interrupts, note,
+ * efct_setup_msix() enables the interrupt
+ */
+ rc = efct_setup_msix(efct, num_interrupts);
+ if (rc) {
+ dev_err(&pdev->dev, "Can't setup msix\n");
+ goto efct_pci_probe_err_setup_msix;
+ }
+ /* Disable interrupt for now */
+ for (i = 0; i < efct->n_msix_vec; i++) {
+ efc_log_debug(efct, "irq %d disabled\n",
+ efct->msix_vec[i].vector);
+ disable_irq(efct->msix_vec[i].vector);
+ }
+
+ rc = efct_device_attach((struct efct *)efct);
+ if (rc)
+ goto efct_pci_probe_err_setup_msix;
+
+ return 0;
+
+efct_pci_probe_err_setup_msix:
+ for (i = 0; i < (u32)num_interrupts; i++) {
+ thread = efct->intr_context[i].thread;
+ if (!thread)
+ continue;
+
+ /* Call stop */
+ kthread_stop(thread);
+ }
+
+efct_pci_probe_err_setup_thread:
+ pci_set_drvdata(pdev, NULL);
+
+ for (i = 0; i < EFCT_PCI_MAX_REGS; i++) {
+ if (efct->reg[i])
+ iounmap(efct->reg[i]);
+ }
+ efct_device_free(efct);
+efct_pci_probe_err_efct_device_alloc:
+ pci_release_regions(pdev);
+efct_pci_probe_err_request_regions:
+ pci_clear_mwi(pdev);
+efct_pci_probe_err_set_mwi:
+ pci_disable_device(pdev);
+efct_pci_probe_err_enable:
+ return rc;
+}
+
+static void
+efct_pci_remove(struct pci_dev *pdev)
+{
+ struct efct *efct = pci_get_drvdata(pdev);
+ u32 i;
+
+ if (!efct)
+ return;
+
+ efct_device_detach(efct);
+
+ efct_teardown_msix(efct);
+
+ for (i = 0; i < EFCT_PCI_MAX_REGS; i++) {
+ if (efct->reg[i])
+ iounmap(efct->reg[i]);
+ }
+
+ pci_set_drvdata(pdev, NULL);
+
+ efct_devices[efct->instance_index] = NULL;
+
+ efct_device_free(efct);
+
+ pci_release_regions(pdev);
+
+ pci_disable_device(pdev);
+}
+
+static void
+efct_device_prep_for_reset(struct efct *efct, struct pci_dev *pdev)
+{
+ if (efct) {
+ efc_log_debug(efct,
+ "PCI channel disable preparing for reset\n");
+ efct_device_detach(efct);
+ /* Disable interrupt and pci device */
+ efct_teardown_msix(efct);
+ }
+ pci_disable_device(pdev);
+}
+
+static void
+efct_device_prep_for_recover(struct efct *efct)
+{
+ if (efct) {
+ efc_log_debug(efct, "PCI channel preparing for recovery\n");
+ efct_hw_io_abort_all(&efct->hw);
+ }
+}
+
+/**
+ * efct_pci_io_error_detected - method for handling PCI I/O error
+ * @pdev: pointer to PCI device.
+ * @state: the current PCI connection state.
+ *
+ * This routine is registered to the PCI subsystem for error handling. This
+ * function is called by the PCI subsystem after a PCI bus error affecting
+ * this device has been detected. When this routine is invoked, it dispatches
+ * device error detected handling routine, which will perform the proper
+ * error detected operation.
+ *
+ * Return codes
+ * PCI_ERS_RESULT_NEED_RESET - need to reset before recovery
+ * PCI_ERS_RESULT_DISCONNECT - device could not be recovered
+ */
+static pci_ers_result_t
+efct_pci_io_error_detected(struct pci_dev *pdev, pci_channel_state_t state)
+{
+ struct efct *efct = pci_get_drvdata(pdev);
+ pci_ers_result_t rc;
+
+ switch (state) {
+ case pci_channel_io_normal:
+ efct_device_prep_for_recover(efct);
+ rc = PCI_ERS_RESULT_CAN_RECOVER;
+ break;
+ case pci_channel_io_frozen:
+ efct_device_prep_for_reset(efct, pdev);
+ rc = PCI_ERS_RESULT_NEED_RESET;
+ break;
+ case pci_channel_io_perm_failure:
+ efct_device_detach(efct);
+ rc = PCI_ERS_RESULT_DISCONNECT;
+ break;
+ default:
+ efc_log_debug(efct, "Unknown PCI error state:0x%x\n",
+ state);
+ efct_device_prep_for_reset(efct, pdev);
+ rc = PCI_ERS_RESULT_NEED_RESET;
+ break;
+ }
+
+ return rc;
+}
+
+static pci_ers_result_t
+efct_pci_io_slot_reset(struct pci_dev *pdev)
+{
+ int rc;
+ struct efct *efct = pci_get_drvdata(pdev);
+
+ rc = pci_enable_device_mem(pdev);
+ if (rc) {
+ efc_log_err(efct,
+ "failed to re-enable PCI device after reset.\n");
+ return PCI_ERS_RESULT_DISCONNECT;
+ }
+
+ /*
+ * As the new kernel behavior of pci_restore_state() API call clears
+ * device saved_state flag, need to save the restored state again.
+ */
+
+ pci_save_state(pdev);
+
+ pci_set_master(pdev);
+
+ rc = efct_setup_msix(efct, efct->n_msix_vec);
+ if (rc)
+ efc_log_err(efct, "rc %d returned, IRQ allocation failed\n",
+ rc);
+
+ /* Perform device reset */
+ efct_device_detach(efct);
+ /* Bring device to online*/
+ efct_device_attach(efct);
+
+ return PCI_ERS_RESULT_RECOVERED;
+}
+
+static void
+efct_pci_io_resume(struct pci_dev *pdev)
+{
+ struct efct *efct = pci_get_drvdata(pdev);
+
+ /* Perform device reset */
+ efct_device_detach(efct);
+ /* Bring device to online*/
+ efct_device_attach(efct);
+}
+
+MODULE_DEVICE_TABLE(pci, efct_pci_table);
+
+static struct pci_error_handlers efct_pci_err_handler = {
+ .error_detected = efct_pci_io_error_detected,
+ .slot_reset = efct_pci_io_slot_reset,
+ .resume = efct_pci_io_resume,
+};
+
+static struct pci_driver efct_pci_driver = {
+ .name = EFCT_DRIVER_NAME,
+ .id_table = efct_pci_table,
+ .probe = efct_pci_probe,
+ .remove = efct_pci_remove,
+ .err_handler = &efct_pci_err_handler,
+};
+
+static int efct_proc_get(struct seq_file *m, void *v)
+{
+ u32 i;
+ u32 j;
+ u32 device_count = 0;
+
+ for (i = 0; i < ARRAY_SIZE(efct_devices); i++) {
+ if (efct_devices[i])
+ device_count++;
+ }
+
+ seq_printf(m, "%d\n", device_count);
+
+ for (i = 0; i < ARRAY_SIZE(efct_devices); i++) {
+ if (efct_devices[i]) {
+ struct efct *efct = efct_devices[i];
+
+ for (j = 0; j < efct->n_msix_vec; j++) {
+ seq_printf(m, "%d,%d,%d\n", i,
+ efct->msix_vec[j].vector,
+ -1);
+ }
+ }
+ }
+
+ return 0;
+}
+
+static int efct_proc_open(struct inode *indoe, struct file *file)
+{
+ return single_open(file, efct_proc_get, NULL);
+}
+
+static const struct file_operations efct_proc_fops = {
+ .owner = THIS_MODULE,
+ .open = efct_proc_open,
+ .read = seq_read,
+ .llseek = seq_lseek,
+ .release = single_release,
+};
+
+
+static
+int __init efct_init(void)
+{
+ int rc = -ENODEV;
+
+ rc = efct_device_init();
+ if (rc) {
+ pr_err("efct_device_init failed rc=%d\n", rc);
+ return -ENOMEM;
+ }
+
+ rc = pci_register_driver(&efct_pci_driver);
+ if (rc)
+ goto l1;
+
+ proc_create(EFCT_DRIVER_NAME, 0444, NULL, &efct_proc_fops);
+ return rc;
+
+l1:
+ efct_device_shutdown();
+ return rc;
+}
+
+static void __exit efct_exit(void)
+{
+ pci_unregister_driver(&efct_pci_driver);
+ remove_proc_entry(EFCT_DRIVER_NAME, NULL);
+ efct_device_shutdown();
+}
+
+module_init(efct_init);
+module_exit(efct_exit);
+MODULE_VERSION(EFCT_DRIVER_VERSION);
+MODULE_LICENSE("GPL");
+MODULE_AUTHOR("Broadcom");
diff --git a/drivers/scsi/elx/efct/efct_driver.h b/drivers/scsi/elx/efct/efct_driver.h
new file mode 100644
index 000000000000..70c0dfaa4c7c
--- /dev/null
+++ b/drivers/scsi/elx/efct/efct_driver.h
@@ -0,0 +1,150 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+/*
+ * Copyright (C) 2019 Broadcom. All Rights Reserved. The term
+ * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
+ */
+
+#if !defined(__EFCT_DRIVER_H__)
+#define __EFCT_DRIVER_H__
+
+/***************************************************************************
+ * OS specific includes
+ */
+#include <stdarg.h>
+#include <linux/version.h>
+#include <linux/init.h>
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/list.h>
+#include <linux/interrupt.h>
+#include <asm-generic/ioctl.h>
+#include <linux/module.h>
+#include <linux/kernel.h>
+#include <linux/pci.h>
+#include <linux/dma-mapping.h>
+#include <linux/bitmap.h>
+#include <linux/slab.h>
+#include <linux/spinlock.h>
+#include <asm/byteorder.h>
+#include <linux/timer.h>
+#include <linux/delay.h>
+#include <linux/fs.h>
+#include <linux/uaccess.h>
+#include <linux/sched.h>
+#include <asm/current.h>
+#include <asm/cacheflush.h>
+#include <linux/pagemap.h>
+#include <linux/kthread.h>
+#include <linux/proc_fs.h>
+#include <linux/seq_file.h>
+#include <linux/random.h>
+#include <linux/sched.h>
+#include <linux/jiffies.h>
+#include <linux/ctype.h>
+#include <linux/debugfs.h>
+#include <linux/firmware.h>
+#include <linux/sched/signal.h>
+#include "../include/efc_common.h"
+
+#define EFCT_DRIVER_NAME "efct"
+#define EFCT_DRIVER_VERSION "1.0.0.0"
+
+/* EFCT_OS_MAX_ISR_TIME_MSEC -
+ * maximum time driver code should spend in an interrupt
+ * or kernel thread context without yielding
+ */
+#define EFCT_OS_MAX_ISR_TIME_MSEC 1000
+
+#define EFCT_FC_RQ_SIZE_DEFAULT 1024
+#define EFCT_FC_MAX_SGL 64
+#define EFCT_FC_DIF_SEED 0
+
+/* Timeouts */
+#define EFCT_FC_ELS_SEND_DEFAULT_TIMEOUT 0
+#define EFCT_FC_ELS_DEFAULT_RETRIES 3
+#define EFCT_FC_FLOGI_TIMEOUT_SEC 5
+#define EFCT_FC_DOMAIN_SHUTDOWN_TIMEOUT_USEC 30000000 /* 30 seconds */
+
+/* Watermark */
+#define EFCT_WATERMARK_HIGH_PCT 90
+#define EFCT_WATERMARK_LOW_PCT 80
+#define EFCT_IO_WATERMARK_PER_INITIATOR 8
+
+#include "efct_utils.h"
+#include "../libefc/efclib.h"
+#include "efct_hw.h"
+#include "efct_io.h"
+#include "efct_xport.h"
+
+#define EFCT_PCI_MAX_REGS 6
+#define MAX_PCI_INTERRUPTS 16
+
+struct efct_intr_context {
+ struct efct *efct;
+ u32 index;
+ struct completion done;
+ struct task_struct *thread;
+};
+
+struct efct {
+ struct pci_dev *pcidev;
+ void __iomem *reg[EFCT_PCI_MAX_REGS];
+
+ struct msix_entry msix_vec[MAX_PCI_INTERRUPTS];
+ u32 n_msix_vec;
+ struct efct_intr_context intr_context[MAX_PCI_INTERRUPTS];
+ u32 numa_node;
+
+ char display_name[EFC_DISPLAY_NAME_LENGTH];
+ bool attached;
+ struct efct_scsi_tgt tgt_efct;
+ struct efct_xport *xport;
+ struct efc *efcport;
+ struct Scsi_Host *shost;
+ int logmask;
+ u32 max_isr_time_msec;
+
+ const char *desc;
+ u32 instance_index;
+
+ const char *model;
+ const char *driver_version;
+ const char *fw_version;
+
+ struct efct_hw hw;
+
+ u32 num_vports;
+ u32 rq_selection_policy;
+ char *filter_def;
+
+ bool soft_wwn_enable;
+
+ /*
+ * Target IO timer value:
+ * Zero: target command timeout disabled.
+ * Non-zero: Timeout value, in seconds, for target commands
+ */
+ u32 target_io_timer_sec;
+
+ int speed;
+ int topology;
+
+ bool enable_numa_support;
+ u8 efct_req_fw_upgrade;
+ u16 sw_feature_cap;
+ struct dentry *sess_debugfs_dir;
+};
+
+#define FW_WRITE_BUFSIZE (64 * 1024)
+
+struct efct_fw_write_result {
+ struct completion done;
+ int status;
+ u32 actual_xfer;
+ u32 change_status;
+};
+
+#define MAX_EFCT_DEVICES 64
+extern struct efct *efct_devices[MAX_EFCT_DEVICES];
+
+#endif /* __EFCT_DRIVER_H__ */
diff --git a/drivers/scsi/elx/efct/efct_hw.c b/drivers/scsi/elx/efct/efct_hw.c
new file mode 100644
index 000000000000..41e400f9d401
--- /dev/null
+++ b/drivers/scsi/elx/efct/efct_hw.c
@@ -0,0 +1,1222 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 Broadcom. All Rights Reserved. The term
+ * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
+ */
+
+#include "efct_driver.h"
+#include "efct_hw.h"
+
+#define EFCT_HW_MQ_DEPTH 128
+#define EFCT_HW_WQ_TIMER_PERIOD_MS 500
+
+#define EFCT_HW_REQUE_XRI_REGTAG 65534
+
+/* HW global data */
+struct efct_hw_global hw_global;
+
+static enum efct_hw_rtn
+efct_hw_link_event_init(struct efct_hw *hw)
+{
+ hw->link.status = SLI_LINK_STATUS_MAX;
+ hw->link.topology = SLI_LINK_TOPO_NONE;
+ hw->link.medium = SLI_LINK_MEDIUM_MAX;
+ hw->link.speed = 0;
+ hw->link.loop_map = NULL;
+ hw->link.fc_id = U32_MAX;
+
+ return EFCT_HW_RTN_SUCCESS;
+}
+
+/**
+ * Adjust the number of WQs and CQs within the HW.
+ *
+ * Calculates the number of WQs and associated CQs needed in the HW based on
+ * the number of IOs. Calculates the starting CQ index for each WQ, RQ and
+ * MQ.
+ */
+static void
+efct_hw_adjust_wqs(struct efct_hw *hw)
+{
+ u32 max_wq_num = hw->sli.qinfo.max_qcount[SLI_QTYPE_WQ];
+ u32 max_wq_entries = hw->num_qentries[SLI_QTYPE_WQ];
+ u32 max_cq_entries = hw->num_qentries[SLI_QTYPE_CQ];
+
+ /*
+ * possibly adjust the the size of the WQs so that the CQ is twice as
+ * big as the WQ to allow for 2 completions per IO. This allows us to
+ * handle multi-phase as well as aborts.
+ */
+ if (max_cq_entries < max_wq_entries * 2) {
+ hw->num_qentries[SLI_QTYPE_WQ] = max_cq_entries / 2;
+ max_wq_entries = hw->num_qentries[SLI_QTYPE_WQ];
+ }
+
+ /*
+ * Calculate the number of WQs to use base on the number of IOs.
+ *
+ * Note: We need to reserve room for aborts which must be sent down
+ * the same WQ as the IO. So we allocate enough WQ space to
+ * handle 2 times the number of IOs. Half of the space will be
+ * used for normal IOs and the other hwf is reserved for aborts.
+ */
+ hw->config.n_wq = ((hw->config.n_io * 2) + (max_wq_entries - 1))
+ / max_wq_entries;
+
+ /* make sure we haven't exceeded the max supported in the HW */
+ if (hw->config.n_wq > EFCT_HW_MAX_NUM_WQ)
+ hw->config.n_wq = EFCT_HW_MAX_NUM_WQ;
+
+ /* make sure we haven't exceeded the chip maximum */
+ if (hw->config.n_wq > max_wq_num)
+ hw->config.n_wq = max_wq_num;
+
+}
+
+static inline void
+efct_hw_add_io_timed_wqe(struct efct_hw *hw, struct efct_hw_io *io)
+{
+ unsigned long flags = 0;
+
+ if (hw->config.emulate_tgt_wqe_timeout && io->tgt_wqe_timeout) {
+ /*
+ * Active WQE list currently only used for
+ * target WQE timeouts.
+ */
+ spin_lock_irqsave(&hw->io_lock, flags);
+ INIT_LIST_HEAD(&io->wqe_link);
+ list_add_tail(&io->wqe_link, &hw->io_timed_wqe);
+ io->submit_ticks = jiffies_64;
+ spin_unlock_irqrestore(&hw->io_lock, flags);
+ }
+}
+
+static inline void
+efct_hw_remove_io_timed_wqe(struct efct_hw *hw, struct efct_hw_io *io)
+{
+ unsigned long flags = 0;
+
+ if (hw->config.emulate_tgt_wqe_timeout) {
+ /*
+ * If target wqe timeouts are enabled,
+ * remove from active wqe list.
+ */
+ spin_lock_irqsave(&hw->io_lock, flags);
+ if (io->wqe_link.next)
+ list_del(&io->wqe_link);
+ spin_unlock_irqrestore(&hw->io_lock, flags);
+ }
+}
+
+static enum efct_hw_rtn
+efct_hw_read_max_dump_size(struct efct_hw *hw)
+{
+ u8 buf[SLI4_BMBX_SIZE];
+ u8 func;
+ struct efct *efct = hw->os;
+ int rc = 0;
+
+ /* attempt to detemine the dump size for function 0 only. */
+ func = PCI_FUNC(efct->pcidev->devfn);
+ if (func == 0) {
+ if (!sli_cmd_common_set_dump_location(&hw->sli, buf,
+ SLI4_BMBX_SIZE, 1, 0,
+ NULL, 0)) {
+ struct sli4_rsp_cmn_set_dump_location *rsp =
+ (struct sli4_rsp_cmn_set_dump_location *)
+ (buf + offsetof(struct sli4_cmd_sli_config,
+ payload.embed));
+
+ rc = efct_hw_command(hw, buf, EFCT_CMD_POLL, NULL,
+ NULL);
+ if (rc != EFCT_HW_RTN_SUCCESS) {
+ efc_log_test(hw->os,
+ "set dump location cmd failed\n");
+ return rc;
+ }
+ hw->dump_size =
+ (le32_to_cpu(rsp->buffer_length_dword) &
+ RSP_SET_DUMP_BUFFER_LEN);
+ efc_log_debug(hw->os, "Dump size %x\n",
+ hw->dump_size);
+ }
+ }
+ return EFCT_HW_RTN_SUCCESS;
+}
+
+enum efct_hw_rtn
+efct_hw_setup(struct efct_hw *hw, void *os, struct pci_dev *pdev)
+{
+ u32 i;
+ struct sli4 *sli = &hw->sli;
+
+ if (!hw) {
+ pr_err("bad parameter(s) hw=%p\n", hw);
+ return EFCT_HW_RTN_ERROR;
+ }
+
+ if (hw->hw_setup_called)
+ return EFCT_HW_RTN_SUCCESS;
+
+ /*
+ * efct_hw_init() relies on NULL pointers indicating that a structure
+ * needs allocation. If a structure is non-NULL, efct_hw_init() won't
+ * free/realloc that memory
+ */
+ memset(hw, 0, sizeof(struct efct_hw));
+
+ hw->hw_setup_called = true;
+
+ hw->os = os;
+
+ spin_lock_init(&hw->cmd_lock);
+ INIT_LIST_HEAD(&hw->cmd_head);
+ INIT_LIST_HEAD(&hw->cmd_pending);
+ hw->cmd_head_count = 0;
+
+ spin_lock_init(&hw->io_lock);
+ spin_lock_init(&hw->io_abort_lock);
+
+ atomic_set(&hw->io_alloc_failed_count, 0);
+
+ hw->config.speed = FC_LINK_SPEED_AUTO_16_8_4;
+ hw->config.dif_seed = 0;
+ if (sli_setup(&hw->sli, hw->os, pdev, ((struct efct *)os)->reg)) {
+ efc_log_err(hw->os, "SLI setup failed\n");
+ return EFCT_HW_RTN_ERROR;
+ }
+
+ efct_hw_link_event_init(hw);
+
+ sli_callback(&hw->sli, SLI4_CB_LINK, efct_hw_cb_link, hw);
+
+ /*
+ * Set all the queue sizes to the maximum allowed.
+ */
+ for (i = 0; i < ARRAY_SIZE(hw->num_qentries); i++)
+ hw->num_qentries[i] = hw->sli.qinfo.max_qentries[i];
+
+ /*
+ * The RQ assignment for RQ pair mode.
+ */
+ hw->config.rq_default_buffer_size = EFCT_HW_RQ_SIZE_PAYLOAD;
+ hw->config.n_io = hw->sli.extent[SLI_RSRC_XRI].size;
+
+ (void)efct_hw_read_max_dump_size(hw);
+
+ /* calculate the number of WQs required. */
+ efct_hw_adjust_wqs(hw);
+
+ /* Set the default dif mode */
+ if (!(sli->features & SLI4_REQFEAT_DIF &&
+ sli->t10_dif_inline_capable)) {
+ efc_log_test(hw->os,
+ "not inline capable, setting mode to separate\n");
+ hw->config.dif_mode = EFCT_HW_DIF_MODE_SEPARATE;
+ }
+
+ hw->config.queue_topology = hw_global.queue_topology_string;
+
+ hw->qtop = efct_hw_qtop_parse(hw, hw->config.queue_topology);
+ if (!hw->qtop) {
+ efc_log_crit(hw->os, "Queue topology string is invalid\n");
+ return EFCT_HW_RTN_ERROR;
+ }
+
+ hw->config.n_eq = hw->qtop->entry_counts[QTOP_EQ];
+ hw->config.n_cq = hw->qtop->entry_counts[QTOP_CQ];
+ hw->config.n_rq = hw->qtop->entry_counts[QTOP_RQ];
+ hw->config.n_wq = hw->qtop->entry_counts[QTOP_WQ];
+ hw->config.n_mq = hw->qtop->entry_counts[QTOP_MQ];
+
+ return EFCT_HW_RTN_SUCCESS;
+}
+
+static void
+efct_logfcfi(struct efct_hw *hw, u32 j, u32 i, u32 id)
+{
+ efc_log_info(hw->os,
+ "REG_FCFI: filter[%d] %08X -> RQ[%d] id=%d\n",
+ j, hw->config.filter_def[j], i, id);
+}
+
+enum efct_hw_rtn
+efct_hw_init(struct efct_hw *hw)
+{
+ enum efct_hw_rtn rc;
+ u32 i = 0;
+ u8 buf[SLI4_BMBX_SIZE];
+ u32 max_rpi;
+ int rem_count;
+ u32 count;
+ unsigned long flags = 0;
+ struct efct_hw_io *temp;
+ struct sli4_cmd_rq_cfg rq_cfg[SLI4_CMD_REG_FCFI_NUM_RQ_CFG];
+ struct sli4 *sli = &hw->sli;
+ struct efct *efct = hw->os;
+
+ /*
+ * Make sure the command lists are empty. If this is start-of-day,
+ * they'll be empty since they were just initialized in efct_hw_setup.
+ * If we've just gone through a reset, the command and command pending
+ * lists should have been cleaned up as part of the reset
+ * (efct_hw_reset()).
+ */
+ spin_lock_irqsave(&hw->cmd_lock, flags);
+ if (!list_empty(&hw->cmd_head)) {
+ efc_log_test(hw->os, "command found on cmd list\n");
+ spin_unlock_irqrestore(&hw->cmd_lock, flags);
+ return EFCT_HW_RTN_ERROR;
+ }
+ if (!list_empty(&hw->cmd_pending)) {
+ efc_log_test(hw->os,
+ "command found on pending list\n");
+ spin_unlock_irqrestore(&hw->cmd_lock, flags);
+ return EFCT_HW_RTN_ERROR;
+ }
+ spin_unlock_irqrestore(&hw->cmd_lock, flags);
+
+ /* Free RQ buffers if prevously allocated */
+ efct_hw_rx_free(hw);
+
+ /*
+ * The IO queues must be initialized here for the reset case. The
+ * efct_hw_init_io() function will re-add the IOs to the free list.
+ * The cmd_head list should be OK since we free all entries in
+ * efct_hw_command_cancel() that is called in the efct_hw_reset().
+ */
+
+ /* If we are in this function due to a reset, there may be stale items
+ * on lists that need to be removed. Clean them up.
+ */
+ rem_count = 0;
+ if (hw->io_wait_free.next) {
+ while ((!list_empty(&hw->io_wait_free))) {
+ rem_count++;
+ temp = list_first_entry(&hw->io_wait_free,
+ struct efct_hw_io,
+ list_entry);
+ list_del(&temp->list_entry);
+ }
+ if (rem_count > 0) {
+ efc_log_debug(hw->os,
+ "rmvd %d items from io_wait_free list\n",
+ rem_count);
+ }
+ }
+ rem_count = 0;
+ if (hw->io_inuse.next) {
+ while ((!list_empty(&hw->io_inuse))) {
+ rem_count++;
+ temp = list_first_entry(&hw->io_inuse,
+ struct efct_hw_io,
+ list_entry);
+ list_del(&temp->list_entry);
+ }
+ if (rem_count > 0)
+ efc_log_debug(hw->os,
+ "rmvd %d items from io_inuse list\n",
+ rem_count);
+ }
+ rem_count = 0;
+ if (hw->io_free.next) {
+ while ((!list_empty(&hw->io_free))) {
+ rem_count++;
+ temp = list_first_entry(&hw->io_free,
+ struct efct_hw_io,
+ list_entry);
+ list_del(&temp->list_entry);
+ }
+ if (rem_count > 0)
+ efc_log_debug(hw->os,
+ "rmvd %d items from io_free list\n",
+ rem_count);
+ }
+
+ INIT_LIST_HEAD(&hw->io_inuse);
+ INIT_LIST_HEAD(&hw->io_free);
+ INIT_LIST_HEAD(&hw->io_wait_free);
+ INIT_LIST_HEAD(&hw->io_timed_wqe);
+
+ /* If MRQ not required, Make sure we dont request feature. */
+ if (hw->config.n_rq == 1)
+ hw->sli.features &= (~SLI4_REQFEAT_MRQP);
+
+ if (sli_init(&hw->sli)) {
+ efc_log_err(hw->os, "SLI failed to initialize\n");
+ return EFCT_HW_RTN_ERROR;
+ }
+ if (hw->sliport_healthcheck) {
+ rc = efct_hw_config_sli_port_health_check(hw, 0, 1);
+ if (rc != EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(hw->os, "Enable port Health check fail\n");
+ return rc;
+ }
+ }
+
+ /*
+ * Set FDT transfer hint, only works on Lancer
+ */
+ if (hw->sli.if_type == SLI4_INTF_IF_TYPE_2 &&
+ EFCT_HW_FDT_XFER_HINT != 0) {
+ /*
+ * Non-fatal error. In particular, we can disregard failure to
+ * set EFCT_HW_FDT_XFER_HINT on devices with legacy firmware
+ * that do not support EFCT_HW_FDT_XFER_HINT feature.
+ */
+ efct_hw_config_set_fdt_xfer_hint(hw, EFCT_HW_FDT_XFER_HINT);
+ }
+
+ /*
+ * Verify that we have not exceeded any queue sizes
+ */
+ if (hw->config.n_eq > sli->qinfo.max_qcount[SLI_QTYPE_EQ]) {
+ efc_log_err(hw->os, "requested %d EQ but %d allowed\n",
+ hw->config.n_eq,
+ sli->qinfo.max_qcount[SLI_QTYPE_EQ]);
+ return EFCT_HW_RTN_ERROR;
+ }
+ if (hw->config.n_cq > sli->qinfo.max_qcount[SLI_QTYPE_CQ]) {
+ efc_log_err(hw->os, "requested %d CQ but %d allowed\n",
+ hw->config.n_cq,
+ sli->qinfo.max_qcount[SLI_QTYPE_CQ]);
+ return EFCT_HW_RTN_ERROR;
+ }
+ if (hw->config.n_mq > sli->qinfo.max_qcount[SLI_QTYPE_MQ]) {
+ efc_log_err(hw->os, "requested %d MQ but %d allowed\n",
+ hw->config.n_mq,
+ sli->qinfo.max_qcount[SLI_QTYPE_MQ]);
+ return EFCT_HW_RTN_ERROR;
+ }
+ if (hw->config.n_rq > sli->qinfo.max_qcount[SLI_QTYPE_RQ]) {
+ efc_log_err(hw->os, "requested %d RQ but %d allowed\n",
+ hw->config.n_rq,
+ sli->qinfo.max_qcount[SLI_QTYPE_RQ]);
+ return EFCT_HW_RTN_ERROR;
+ }
+ if (hw->config.n_wq > sli->qinfo.max_qcount[SLI_QTYPE_WQ]) {
+ efc_log_err(hw->os, "requested %d WQ but %d allowed\n",
+ hw->config.n_wq,
+ sli->qinfo.max_qcount[SLI_QTYPE_WQ]);
+ return EFCT_HW_RTN_ERROR;
+ }
+
+ /* zero the hashes */
+ memset(hw->cq_hash, 0, sizeof(hw->cq_hash));
+ efc_log_debug(hw->os, "Max CQs %d, hash size = %d\n",
+ EFCT_HW_MAX_NUM_CQ, EFCT_HW_Q_HASH_SIZE);
+
+ memset(hw->rq_hash, 0, sizeof(hw->rq_hash));
+ efc_log_debug(hw->os, "Max RQs %d, hash size = %d\n",
+ EFCT_HW_MAX_NUM_RQ, EFCT_HW_Q_HASH_SIZE);
+
+ memset(hw->wq_hash, 0, sizeof(hw->wq_hash));
+ efc_log_debug(hw->os, "Max WQs %d, hash size = %d\n",
+ EFCT_HW_MAX_NUM_WQ, EFCT_HW_Q_HASH_SIZE);
+
+ rc = efct_hw_init_queues(hw, hw->qtop);
+ if (rc != EFCT_HW_RTN_SUCCESS)
+ return rc;
+
+ max_rpi = sli->extent[SLI_RSRC_RPI].size;
+ i = sli_fc_get_rpi_requirements(&hw->sli, max_rpi);
+ if (i) {
+ struct efc_dma payload_memory;
+
+ rc = EFCT_HW_RTN_ERROR;
+
+ if (hw->rnode_mem.size) {
+ dma_free_coherent(&efct->pcidev->dev,
+ hw->rnode_mem.size,
+ hw->rnode_mem.virt,
+ hw->rnode_mem.phys);
+ memset(&hw->rnode_mem, 0, sizeof(struct efc_dma));
+ }
+
+ hw->rnode_mem.size = i;
+ hw->rnode_mem.virt = dma_alloc_coherent(&efct->pcidev->dev,
+ hw->rnode_mem.size,
+ &hw->rnode_mem.phys,
+ GFP_DMA);
+ if (!hw->rnode_mem.virt) {
+ efc_log_err(hw->os,
+ "remote node memory allocation fail\n");
+ return EFCT_HW_RTN_NO_MEMORY;
+ }
+
+ payload_memory.size = 0;
+ if (!sli_cmd_post_hdr_templates(&hw->sli, buf,
+ SLI4_BMBX_SIZE,
+ &hw->rnode_mem,
+ U16_MAX,
+ &payload_memory)) {
+ rc = efct_hw_command(hw, buf, EFCT_CMD_POLL, NULL,
+ NULL);
+
+ if (payload_memory.size != 0) {
+ /*
+ * The command was non-embedded - need to
+ * free the dma buffer
+ */
+ dma_free_coherent(&efct->pcidev->dev,
+ payload_memory.size,
+ payload_memory.virt,
+ payload_memory.phys);
+ memset(&payload_memory, 0,
+ sizeof(struct efc_dma));
+ }
+ }
+
+ if (rc != EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(hw->os,
+ "header template registration failed\n");
+ return rc;
+ }
+ }
+
+ /* Allocate and post RQ buffers */
+ rc = efct_hw_rx_allocate(hw);
+ if (rc) {
+ efc_log_err(hw->os, "rx_allocate failed\n");
+ return rc;
+ }
+
+ /* Populate hw->seq_free_list */
+ if (!hw->seq_pool) {
+ u32 count = 0;
+ u32 i;
+
+ /*
+ * Sum up the total number of RQ entries, to use to allocate
+ * the sequence object pool
+ */
+ for (i = 0; i < hw->hw_rq_count; i++)
+ count += hw->hw_rq[i]->entry_count;
+
+ hw->seq_pool = efct_array_alloc(hw->os,
+ sizeof(struct efc_hw_sequence),
+ count);
+ if (!hw->seq_pool) {
+ efc_log_err(hw->os, "malloc seq_pool failed\n");
+ return EFCT_HW_RTN_NO_MEMORY;
+ }
+ }
+
+ if (efct_hw_rx_post(hw))
+ efc_log_err(hw->os, "WARNING - error posting RQ buffers\n");
+
+ /* Allocate rpi_ref if not previously allocated */
+ if (!hw->rpi_ref) {
+ hw->rpi_ref = kmalloc_array(max_rpi, sizeof(*hw->rpi_ref),
+ GFP_KERNEL);
+ if (!hw->rpi_ref)
+ return EFCT_HW_RTN_NO_MEMORY;
+
+ memset(hw->rpi_ref, 0, max_rpi * sizeof(*hw->rpi_ref));
+ }
+
+ for (i = 0; i < max_rpi; i++) {
+ atomic_set(&hw->rpi_ref[i].rpi_count, 0);
+ atomic_set(&hw->rpi_ref[i].rpi_attached, 0);
+ }
+
+ /*
+ * Register a FCFI to allow unsolicited frames to be routed to the
+ * driver
+ */
+ if (hw->hw_mrq_count) {
+ efc_log_info(hw->os, "using REG_FCFI MRQ\n");
+
+ rc = efct_hw_config_mrq(hw,
+ SLI4_CMD_REG_FCFI_SET_FCFI_MODE, 0);
+ if (rc != EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(hw->os,
+ "REG_FCFI_MRQ FCFI reg failed\n");
+ return rc;
+ }
+
+ rc = efct_hw_config_mrq(hw,
+ SLI4_CMD_REG_FCFI_SET_MRQ_MODE, 0);
+ if (rc != EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(hw->os,
+ "REG_FCFI_MRQ MRQ reg failed\n");
+ return rc;
+ }
+ } else {
+ u32 min_rq_count;
+
+ efc_log_info(hw->os, "using REG_FCFI standard\n");
+
+ /*
+ * Set the filter match/mask values from hw's
+ * filter_def values
+ */
+ for (i = 0; i < SLI4_CMD_REG_FCFI_NUM_RQ_CFG; i++) {
+ rq_cfg[i].rq_id = cpu_to_le16(0xffff);
+ rq_cfg[i].r_ctl_mask = (u8)hw->config.filter_def[i];
+ rq_cfg[i].r_ctl_match = (u8)
+ (hw->config.filter_def[i] >> 8);
+ rq_cfg[i].type_mask = (u8)
+ (hw->config.filter_def[i] >> 16);
+ rq_cfg[i].type_match = (u8)
+ (hw->config.filter_def[i] >> 24);
+ }
+
+ /*
+ * Update the rq_id's of the FCF configuration
+ * (don't update more than the number of rq_cfg
+ * elements)
+ */
+ min_rq_count = (hw->hw_rq_count <
+ SLI4_CMD_REG_FCFI_NUM_RQ_CFG)
+ ? hw->hw_rq_count :
+ SLI4_CMD_REG_FCFI_NUM_RQ_CFG;
+ for (i = 0; i < min_rq_count; i++) {
+ struct hw_rq *rq = hw->hw_rq[i];
+ u32 j;
+
+ for (j = 0; j < SLI4_CMD_REG_FCFI_NUM_RQ_CFG;
+ j++) {
+ u32 mask = (rq->filter_mask != 0) ?
+ rq->filter_mask : 1;
+
+ if (!(mask & (1U << j)))
+ continue;
+
+ rq_cfg[j].rq_id = cpu_to_le16(rq->hdr->id);
+ efct_logfcfi(hw, j, i, rq->hdr->id);
+ }
+ }
+
+ rc = EFCT_HW_RTN_ERROR;
+ if (!sli_cmd_reg_fcfi(&hw->sli, buf,
+ SLI4_BMBX_SIZE, 0,
+ rq_cfg)) {
+ rc = efct_hw_command(hw, buf, EFCT_CMD_POLL,
+ NULL, NULL);
+ }
+
+ if (rc != EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(hw->os,
+ "FCFI registration failed\n");
+ return rc;
+ }
+ hw->fcf_indicator =
+ le16_to_cpu(((struct sli4_cmd_reg_fcfi *)buf)->fcfi);
+ }
+
+ /*
+ * Allocate the WQ request tag pool, if not previously allocated
+ * (the request tag value is 16 bits, thus the pool allocation size
+ * of 64k)
+ */
+ rc = efct_hw_reqtag_init(hw);
+ if (rc) {
+ efc_log_err(hw->os, "efct_hw_reqtag_init failed %d\n", rc);
+ return rc;
+ }
+
+ rc = efct_hw_setup_io(hw);
+ if (rc) {
+ efc_log_err(hw->os, "IO allocation failure\n");
+ return rc;
+ }
+
+ rc = efct_hw_init_io(hw);
+ if (rc) {
+ efc_log_err(hw->os, "IO initialization failure\n");
+ return rc;
+ }
+
+ /* Set the DIF seed - only for lancer right now */
+ if (sli->if_type == SLI4_INTF_IF_TYPE_2 &&
+ efct_hw_set_dif_seed(hw) != EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(hw->os, "Failed to set DIF seed value\n");
+ return rc;
+ }
+
+ /*
+ * Arming the EQ allows (e.g.) interrupts when CQ completions write EQ
+ * entries
+ */
+ for (i = 0; i < hw->eq_count; i++)
+ sli_queue_arm(&hw->sli, &hw->eq[i], true);
+
+ /*
+ * Initialize RQ hash
+ */
+ for (i = 0; i < hw->rq_count; i++)
+ efct_hw_queue_hash_add(hw->rq_hash, hw->rq[i].id, i);
+
+ /*
+ * Initialize WQ hash
+ */
+ for (i = 0; i < hw->wq_count; i++)
+ efct_hw_queue_hash_add(hw->wq_hash, hw->wq[i].id, i);
+
+ /*
+ * Arming the CQ allows (e.g.) MQ completions to write CQ entries
+ */
+ for (i = 0; i < hw->cq_count; i++) {
+ efct_hw_queue_hash_add(hw->cq_hash, hw->cq[i].id, i);
+ sli_queue_arm(&hw->sli, &hw->cq[i], true);
+ }
+
+ /* record the fact that the queues are functional */
+ hw->state = EFCT_HW_STATE_ACTIVE;
+
+ /* finally kick off periodic timer to check for timed out target WQEs */
+ if (hw->config.emulate_tgt_wqe_timeout) {
+ timer_setup(&hw->wqe_timer, &target_wqe_timer_cb, 0);
+
+ mod_timer(&hw->wqe_timer, jiffies +
+ msecs_to_jiffies(EFCT_HW_WQ_TIMER_PERIOD_MS));
+ }
+ /*
+ * Allocate a HW IOs for send frame. Allocate one for each Class 1
+ * WQ, or if there are none of those, allocate one for WQ[0]
+ */
+ count = efct_varray_get_count(hw->wq_class_array[1]);
+ if (count > 0) {
+ struct hw_wq *wq;
+
+ for (i = 0; i < count; i++) {
+ wq = efct_varray_iter_next(hw->wq_class_array[1]);
+ wq->send_frame_io = efct_hw_io_alloc(hw);
+ if (!wq->send_frame_io)
+ efc_log_err(hw->os,
+ "alloc for send_frame_io failed\n");
+ }
+ } else {
+ hw->hw_wq[0]->send_frame_io = efct_hw_io_alloc(hw);
+ if (!hw->hw_wq[0]->send_frame_io)
+ efc_log_err(hw->os,
+ "alloc for send_frame_io failed\n");
+ }
+
+ /* Initialize send frame frame sequence id */
+ atomic_set(&hw->send_frame_seq_id, 0);
+
+ /* Initialize watchdog timer if enabled by user */
+ if (hw->watchdog_timeout) {
+ if (hw->watchdog_timeout < 1 ||
+ hw->watchdog_timeout > 65534)
+ efc_log_err(hw->os,
+ "WDT out of range: range is 1 - 65534\n");
+ else if (!efct_hw_config_watchdog_timer(hw))
+ efc_log_info(hw->os,
+ "WDT timer config with tmo = %d secs\n",
+ hw->watchdog_timeout);
+ }
+
+ return EFCT_HW_RTN_SUCCESS;
+}
+
+/**
+ * efct_hw_config_mrq() - Configure Multi-RQ
+ *
+ * @hw: Hardware context allocated by the caller.
+ * @mode: 1 to set MRQ filters and 0 to set FCFI index
+ * @fcf_index: valid in mode 0
+ *
+ * Returns 0 on success, or a non-zero value on failure.
+ */
+static int
+efct_hw_config_mrq(struct efct_hw *hw, u8 mode, u16 fcf_index)
+{
+ u8 buf[SLI4_BMBX_SIZE], mrq_bitmask = 0;
+ struct hw_rq *rq;
+ struct sli4_cmd_reg_fcfi_mrq *rsp = NULL;
+ u32 i, j;
+ struct sli4_cmd_rq_cfg rq_filter[SLI4_CMD_REG_FCFI_MRQ_NUM_RQ_CFG];
+ int rc;
+
+ if (mode == SLI4_CMD_REG_FCFI_SET_FCFI_MODE)
+ goto issue_cmd;
+
+ /* Set the filter match/mask values from hw's filter_def values */
+ for (i = 0; i < SLI4_CMD_REG_FCFI_NUM_RQ_CFG; i++) {
+ rq_filter[i].rq_id = cpu_to_le16(0xffff);
+ rq_filter[i].r_ctl_mask = (u8)hw->config.filter_def[i];
+ rq_filter[i].r_ctl_match = (u8)(hw->config.filter_def[i] >> 8);
+ rq_filter[i].type_mask = (u8)(hw->config.filter_def[i] >> 16);
+ rq_filter[i].type_match = (u8)(hw->config.filter_def[i] >> 24);
+ }
+
+ /* Accumulate counts for each filter type used, build rq_ids[] list */
+ for (i = 0; i < hw->hw_rq_count; i++) {
+ rq = hw->hw_rq[i];
+ for (j = 0; j < SLI4_CMD_REG_FCFI_MRQ_NUM_RQ_CFG; j++) {
+ if (!(rq->filter_mask & (1U << j)))
+ continue;
+
+ if (rq_filter[j].rq_id != cpu_to_le16(0xffff)) {
+ /*
+ * Already used. Bailout ifts not RQset
+ * case
+ */
+ if (!rq->is_mrq ||
+ rq_filter[j].rq_id !=
+ cpu_to_le16(rq->base_mrq_id)) {
+ efc_log_err(hw->os, "Wrong q top.\n");
+ return EFCT_HW_RTN_ERROR;
+ }
+ continue;
+ }
+
+ if (!rq->is_mrq) {
+ rq_filter[j].rq_id = cpu_to_le16(rq->hdr->id);
+ continue;
+ }
+
+ rq_filter[j].rq_id = cpu_to_le16(rq->base_mrq_id);
+ mrq_bitmask |= (1U << j);
+ }
+ }
+
+issue_cmd:
+ /* Invoke REG_FCFI_MRQ */
+ rc = sli_cmd_reg_fcfi_mrq(&hw->sli,
+ buf, /* buf */
+ SLI4_BMBX_SIZE, /* size */
+ mode, /* mode 1 */
+ fcf_index, /* fcf_index */
+ /* RQ selection policy*/
+ hw->config.rq_selection_policy,
+ mrq_bitmask, /* MRQ bitmask */
+ hw->hw_mrq_count, /* num_mrqs */
+ rq_filter);/* RQ filter */
+ if (rc) {
+ efc_log_err(hw->os,
+ "sli_cmd_reg_fcfi_mrq() failed: %d\n", rc);
+ return EFCT_HW_RTN_ERROR;
+ }
+
+ rc = efct_hw_command(hw, buf, EFCT_CMD_POLL, NULL, NULL);
+
+ rsp = (struct sli4_cmd_reg_fcfi_mrq *)buf;
+
+ if (rc != EFCT_HW_RTN_SUCCESS ||
+ le16_to_cpu(rsp->hdr.status)) {
+ efc_log_err(hw->os,
+ "FCFI MRQ reg failed. cmd = %x status = %x\n",
+ rsp->hdr.command,
+ le16_to_cpu(rsp->hdr.status));
+ return EFCT_HW_RTN_ERROR;
+ }
+
+ if (mode == SLI4_CMD_REG_FCFI_SET_FCFI_MODE)
+ hw->fcf_indicator = le16_to_cpu(rsp->fcfi);
+ return 0;
+}
+
+enum efct_hw_rtn
+efct_hw_set(struct efct_hw *hw, enum efct_hw_property prop, u32 value)
+{
+ enum efct_hw_rtn rc = EFCT_HW_RTN_SUCCESS;
+ struct sli4 *sli = &hw->sli;
+
+ switch (prop) {
+ case EFCT_HW_N_IO:
+ if (value > sli->extent[SLI_RSRC_XRI].size ||
+ value == 0) {
+ efc_log_test(hw->os,
+ "IO value out of range %d vs %d\n",
+ value,
+ sli->extent[SLI_RSRC_XRI].size);
+ rc = EFCT_HW_RTN_ERROR;
+ } else {
+ hw->config.n_io = value;
+ }
+ break;
+ case EFCT_HW_N_SGL:
+ value += SLI4_SGE_MAX_RESERVED;
+ if (value > sli_get_max_sgl(&hw->sli)) {
+ efc_log_test(hw->os,
+ "SGL value out of range %d vs %d\n",
+ value, sli_get_max_sgl(&hw->sli));
+ rc = EFCT_HW_RTN_ERROR;
+ } else {
+ hw->config.n_sgl = value;
+ }
+ break;
+ case EFCT_HW_TOPOLOGY:
+ switch (value) {
+ case EFCT_HW_TOPOLOGY_AUTO:
+ sli_set_topology(&hw->sli,
+ SLI4_READ_CFG_TOPO_FC);
+ break;
+ case EFCT_HW_TOPOLOGY_NPORT:
+ sli_set_topology(&hw->sli, SLI4_READ_CFG_TOPO_FC_DA);
+ break;
+ case EFCT_HW_TOPOLOGY_LOOP:
+ sli_set_topology(&hw->sli, SLI4_READ_CFG_TOPO_FC_AL);
+ break;
+ default:
+ efc_log_test(hw->os,
+ "unsupported topology %#x\n", value);
+ rc = EFCT_HW_RTN_ERROR;
+ }
+ hw->config.topology = value;
+ break;
+ case EFCT_HW_LINK_SPEED:
+
+ switch (value) {
+ case 0: /* Auto-speed negotiation */
+ hw->config.speed = FC_LINK_SPEED_AUTO_16_8_4;
+ break;
+ case 2000: /* FC speeds */
+ hw->config.speed = FC_LINK_SPEED_2G;
+ break;
+ case 4000:
+ hw->config.speed = FC_LINK_SPEED_4G;
+ break;
+ case 8000:
+ hw->config.speed = FC_LINK_SPEED_8G;
+ break;
+ case 16000:
+ hw->config.speed = FC_LINK_SPEED_16G;
+ break;
+ case 32000:
+ hw->config.speed = FC_LINK_SPEED_32G;
+ break;
+ default:
+ efc_log_test(hw->os, "unsupported speed %d\n", value);
+ rc = EFCT_HW_RTN_ERROR;
+ }
+ break;
+ case EFCT_HW_RQ_PROCESS_LIMIT: {
+ struct hw_rq *rq;
+ u32 i;
+
+ /* For each hw_rq object, set its parent CQ limit value */
+ for (i = 0; i < hw->hw_rq_count; i++) {
+ rq = hw->hw_rq[i];
+ hw->cq[rq->cq->instance].proc_limit = value;
+ }
+ break;
+ }
+ case EFCT_HW_RQ_DEFAULT_BUFFER_SIZE:
+ hw->config.rq_default_buffer_size = value;
+ break;
+ case EFCT_ESOC:
+ hw->config.esoc = value;
+ break;
+ case EFCT_HW_HIGH_LOGIN_MODE:
+ rc = sli_set_hlm(&hw->sli, value);
+ break;
+ case EFCT_HW_PREREGISTER_SGL:
+ rc = sli_set_sgl_preregister(&hw->sli, value);
+ break;
+ case EFCT_HW_EMULATE_TARGET_WQE_TIMEOUT:
+ hw->config.emulate_tgt_wqe_timeout = value;
+ break;
+ case EFCT_HW_BOUNCE:
+ hw->config.bounce = value;
+ break;
+ case EFCT_HW_RQ_SELECTION_POLICY:
+ hw->config.rq_selection_policy = value;
+ break;
+ case EFCT_HW_RR_QUANTA:
+ hw->config.rr_quanta = value;
+ break;
+ default:
+ efc_log_test(hw->os, "unsupported property %#x\n", prop);
+ rc = EFCT_HW_RTN_ERROR;
+ }
+
+ return rc;
+}
+
+enum efct_hw_rtn
+efct_hw_set_ptr(struct efct_hw *hw, enum efct_hw_property prop,
+ void *value)
+{
+ enum efct_hw_rtn rc = EFCT_HW_RTN_SUCCESS;
+
+ switch (prop) {
+ case EFCT_HW_FILTER_DEF: {
+ char *p = NULL;
+ char *token;
+ u32 idx = 0;
+
+ for (idx = 0; idx < ARRAY_SIZE(hw->config.filter_def); idx++)
+ hw->config.filter_def[idx] = 0;
+
+ p = kstrdup(value, GFP_KERNEL);
+ if (!p || !*p) {
+ efc_log_err(hw->os, "p is NULL\n");
+ break;
+ }
+
+ idx = 0;
+ while ((token = strsep(&p, ",")) && *token) {
+ if (kstrtou32(token, 0, &hw->config.filter_def[idx++]))
+ efc_log_err(hw->os, "kstrtoint failed\n");
+
+ if (!p || !*p)
+ break;
+
+ if (idx == ARRAY_SIZE(hw->config.filter_def))
+ break;
+ }
+ kfree(p);
+
+ break;
+ }
+ default:
+ efc_log_test(hw->os, "unsupported property %#x\n", prop);
+ rc = EFCT_HW_RTN_ERROR;
+ break;
+ }
+ return rc;
+}
+
+enum efct_hw_rtn
+efct_hw_get(struct efct_hw *hw, enum efct_hw_property prop,
+ u32 *value)
+{
+ enum efct_hw_rtn rc = EFCT_HW_RTN_SUCCESS;
+ int tmp;
+ struct sli4 *sli = &hw->sli;
+
+ if (!value)
+ return EFCT_HW_RTN_ERROR;
+
+ *value = 0;
+
+ switch (prop) {
+ case EFCT_HW_N_IO:
+ *value = hw->config.n_io;
+ break;
+ case EFCT_HW_N_SGL:
+ *value = (hw->config.n_sgl - SLI4_SGE_MAX_RESERVED);
+ break;
+ case EFCT_HW_MAX_IO:
+ *value = sli->extent[SLI_RSRC_XRI].size;
+ break;
+ case EFCT_HW_MAX_NODES:
+ *value = sli->extent[SLI_RSRC_RPI].size;
+ break;
+ case EFCT_HW_MAX_RQ_ENTRIES:
+ *value = hw->num_qentries[SLI_QTYPE_RQ];
+ break;
+ case EFCT_HW_RQ_DEFAULT_BUFFER_SIZE:
+ *value = hw->config.rq_default_buffer_size;
+ break;
+ case EFCT_HW_MAX_SGE:
+ *value = sli->sge_supported_length;
+ break;
+ case EFCT_HW_MAX_SGL:
+ *value = sli_get_max_sgl(&hw->sli);
+ break;
+ case EFCT_HW_TOPOLOGY:
+ /*
+ * Infer link.status based on link.speed.
+ * Report EFCT_HW_TOPOLOGY_NONE if the link is down.
+ */
+ if (hw->link.speed == 0) {
+ *value = EFCT_HW_TOPOLOGY_NONE;
+ break;
+ }
+ switch (hw->link.topology) {
+ case SLI_LINK_TOPO_NPORT:
+ *value = EFCT_HW_TOPOLOGY_NPORT;
+ break;
+ case SLI_LINK_TOPO_LOOP:
+ *value = EFCT_HW_TOPOLOGY_LOOP;
+ break;
+ case SLI_LINK_TOPO_NONE:
+ *value = EFCT_HW_TOPOLOGY_NONE;
+ break;
+ default:
+ efc_log_test(hw->os,
+ "unsupported topology %#x\n",
+ hw->link.topology);
+ rc = EFCT_HW_RTN_ERROR;
+ break;
+ }
+ break;
+ case EFCT_HW_CONFIG_TOPOLOGY:
+ *value = hw->config.topology;
+ break;
+ case EFCT_HW_LINK_SPEED:
+ *value = hw->link.speed;
+ break;
+ case EFCT_HW_LINK_CONFIG_SPEED:
+ switch (hw->config.speed) {
+ case FC_LINK_SPEED_10G:
+ *value = 10000;
+ break;
+ case FC_LINK_SPEED_AUTO_16_8_4:
+ *value = 0;
+ break;
+ case FC_LINK_SPEED_2G:
+ *value = 2000;
+ break;
+ case FC_LINK_SPEED_4G:
+ *value = 4000;
+ break;
+ case FC_LINK_SPEED_8G:
+ *value = 8000;
+ break;
+ case FC_LINK_SPEED_16G:
+ *value = 16000;
+ break;
+ case FC_LINK_SPEED_32G:
+ *value = 32000;
+ break;
+ default:
+ efc_log_test(hw->os,
+ "unsupported speed %#x\n",
+ hw->config.speed);
+ rc = EFCT_HW_RTN_ERROR;
+ break;
+ }
+ break;
+ case EFCT_HW_IF_TYPE:
+ *value = sli->if_type;
+ break;
+ case EFCT_HW_SLI_REV:
+ *value = sli->sli_rev;
+ break;
+ case EFCT_HW_SLI_FAMILY:
+ *value = sli->sli_family;
+ break;
+ case EFCT_HW_DIF_CAPABLE:
+ *value = sli->features & SLI4_REQFEAT_DIF;
+ break;
+ case EFCT_HW_DIF_SEED:
+ *value = hw->config.dif_seed;
+ break;
+ case EFCT_HW_DIF_MODE:
+ *value = hw->config.dif_mode;
+ break;
+ case EFCT_HW_DIF_MULTI_SEPARATE:
+ /* Lancer supports multiple DIF separates */
+ if (hw->sli.if_type == SLI4_INTF_IF_TYPE_2)
+ *value = true;
+ else
+ *value = false;
+ break;
+ case EFCT_HW_DUMP_MAX_SIZE:
+ *value = hw->dump_size;
+ break;
+ case EFCT_HW_DUMP_READY:
+ *value = sli_dump_is_ready(&hw->sli);
+ break;
+ case EFCT_HW_DUMP_PRESENT:
+ *value = sli_dump_is_present(&hw->sli);
+ break;
+ case EFCT_HW_RESET_REQUIRED:
+ tmp = sli_reset_required(&hw->sli);
+ if (tmp < 0)
+ rc = EFCT_HW_RTN_ERROR;
+ else
+ *value = tmp;
+ break;
+ case EFCT_HW_FW_ERROR:
+ *value = sli_fw_error_status(&hw->sli);
+ break;
+ case EFCT_HW_FW_READY:
+ *value = sli_fw_ready(&hw->sli);
+ break;
+ case EFCT_HW_HIGH_LOGIN_MODE:
+ *value = sli->features & SLI4_REQFEAT_HLM;
+ break;
+ case EFCT_HW_PREREGISTER_SGL:
+ *value = sli->sgl_pre_registration_required;
+ break;
+ case EFCT_HW_HW_REV1:
+ *value = sli->hw_rev[0];
+ break;
+ case EFCT_HW_HW_REV2:
+ *value = sli->hw_rev[1];
+ break;
+ case EFCT_HW_HW_REV3:
+ *value = sli->hw_rev[2];
+ break;
+ case EFCT_HW_LINK_MODULE_TYPE:
+ *value = sli->link_module_type;
+ break;
+ case EFCT_HW_EMULATE_TARGET_WQE_TIMEOUT:
+ *value = hw->config.emulate_tgt_wqe_timeout;
+ break;
+ case EFCT_HW_VPD_LEN:
+ *value = sli->vpd_length;
+ break;
+ case EFCT_HW_SEND_FRAME_CAPABLE:
+ *value = 0;
+ break;
+ case EFCT_HW_RQ_SELECTION_POLICY:
+ *value = hw->config.rq_selection_policy;
+ break;
+ case EFCT_HW_RR_QUANTA:
+ *value = hw->config.rr_quanta;
+ break;
+ case EFCT_HW_MAX_VPORTS:
+ *value = sli->extent[SLI_RSRC_VPI].size;
+ break;
+ default:
+ efc_log_test(hw->os, "unsupported property %#x\n", prop);
+ rc = EFCT_HW_RTN_ERROR;
+ }
+
+ return rc;
+}
+
+void *
+efct_hw_get_ptr(struct efct_hw *hw, enum efct_hw_property prop)
+{
+ void *rc = NULL;
+ struct sli4 *sli = &hw->sli;
+
+ switch (prop) {
+ case EFCT_HW_WWN_NODE:
+ rc = sli->wwnn;
+ break;
+ case EFCT_HW_WWN_PORT:
+ rc = sli->wwpn;
+ break;
+ case EFCT_HW_VPD:
+ /* make sure VPD length is non-zero */
+ if (sli->vpd_length)
+ rc = sli->vpd_data.virt;
+ break;
+ case EFCT_HW_FW_REV:
+ rc = sli->fw_name[0];
+ break;
+ case EFCT_HW_FW_REV2:
+ rc = sli->fw_name[1];
+ break;
+ case EFCT_HW_IPL:
+ rc = sli->ipl_name;
+ break;
+ case EFCT_HW_PORTNUM:
+ rc = sli->port_name;
+ break;
+ case EFCT_HW_BIOS_VERSION_STRING:
+ rc = sli->bios_version_string;
+ break;
+ default:
+ efc_log_test(hw->os, "unsupported property %#x\n", prop);
+ }
+
+ return rc;
+}
+
+uint64_t
+efct_get_wwn(struct efct_hw *hw, enum efct_hw_property prop)
+{
+ u8 *p = efct_hw_get_ptr(hw, prop);
+ u64 value = 0;
+
+ if (p) {
+ u32 i;
+
+ for (i = 0; i < sizeof(value); i++)
+ value = (value << 8) | p[i];
+ }
+
+ return value;
+}
diff --git a/drivers/scsi/elx/efct/efct_hw.h b/drivers/scsi/elx/efct/efct_hw.h
index ff6de91923fa..bbba73969de3 100644
--- a/drivers/scsi/elx/efct/efct_hw.h
+++ b/drivers/scsi/elx/efct/efct_hw.h
@@ -149,7 +149,6 @@ enum efct_hw_property {
EFCT_HW_ETH_LICENSE,
EFCT_HW_LINK_MODULE_TYPE,
EFCT_HW_NUM_CHUTES,
- EFCT_HW_WAR_VERSION,
/* enable driver timeouts for target WQEs */
EFCT_HW_EMULATE_TARGET_WQE_TIMEOUT,
EFCT_HW_LINK_CONFIG_SPEED,
@@ -849,4 +848,19 @@ struct efct_hw_grp_hdr {
u8 revision[32];
};
+extern enum efct_hw_rtn
+efct_hw_setup(struct efct_hw *hw, void *os, struct pci_dev *pdev);
+enum efct_hw_rtn efct_hw_init(struct efct_hw *hw);
+extern enum efct_hw_rtn
+efct_hw_get(struct efct_hw *hw, enum efct_hw_property prop, u32 *value);
+extern void *
+efct_hw_get_ptr(struct efct_hw *hw, enum efct_hw_property prop);
+extern enum efct_hw_rtn
+efct_hw_set(struct efct_hw *hw, enum efct_hw_property prop, u32 value);
+extern enum efct_hw_rtn
+efct_hw_set_ptr(struct efct_hw *hw, enum efct_hw_property prop,
+ void *value);
+extern uint64_t
+efct_get_wwn(struct efct_hw *hw, enum efct_hw_property prop);
+
#endif /* __EFCT_H__ */
diff --git a/drivers/scsi/elx/efct/efct_xport.c b/drivers/scsi/elx/efct/efct_xport.c
new file mode 100644
index 000000000000..e6d6f2000168
--- /dev/null
+++ b/drivers/scsi/elx/efct/efct_xport.c
@@ -0,0 +1,587 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 Broadcom. All Rights Reserved. The term
+ * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
+ */
+
+#include "efct_driver.h"
+#include "efct_unsol.h"
+
+/* Post node event callback argument. */
+struct efct_xport_post_node_event {
+ struct completion done;
+ atomic_t refcnt;
+ struct efc_node *node;
+ u32 evt;
+ void *context;
+};
+
+static struct dentry *efct_debugfs_root;
+static atomic_t efct_debugfs_count;
+
+static struct scsi_host_template efct_template = {
+ .module = THIS_MODULE,
+ .name = EFCT_DRIVER_NAME,
+ .supported_mode = MODE_TARGET,
+};
+
+/* globals */
+static struct fc_function_template efct_xport_functions;
+static struct fc_function_template efct_vport_functions;
+
+static struct scsi_transport_template *efct_xport_fc_tt;
+static struct scsi_transport_template *efct_vport_fc_tt;
+
+/*
+ * transport object is allocated,
+ * and associated with a device instance
+ */
+struct efct_xport *
+efct_xport_alloc(struct efct *efct)
+{
+ struct efct_xport *xport;
+
+ xport = kmalloc(sizeof(*xport), GFP_KERNEL);
+ if (!xport)
+ return xport;
+
+ memset(xport, 0, sizeof(*xport));
+ xport->efct = efct;
+ return xport;
+}
+
+static int
+efct_xport_init_debugfs(struct efct *efct)
+{
+ /* Setup efct debugfs root directory */
+ if (!efct_debugfs_root) {
+ efct_debugfs_root = debugfs_create_dir("efct", NULL);
+ atomic_set(&efct_debugfs_count, 0);
+ if (!efct_debugfs_root) {
+ efc_log_err(efct, "failed to create debugfs entry\n");
+ goto debugfs_fail;
+ }
+ }
+
+ /* Create a directory for sessions in root */
+ if (!efct->sess_debugfs_dir) {
+ efct->sess_debugfs_dir = debugfs_create_dir("sessions", NULL);
+ if (!efct->sess_debugfs_dir) {
+ efc_log_err(efct,
+ "failed to create debugfs entry for sessions\n");
+ goto debugfs_fail;
+ }
+ atomic_inc(&efct_debugfs_count);
+ }
+
+ return 0;
+
+debugfs_fail:
+ return -1;
+}
+
+static void efct_xport_delete_debugfs(struct efct *efct)
+{
+ /* Remove session debugfs directory */
+ debugfs_remove(efct->sess_debugfs_dir);
+ efct->sess_debugfs_dir = NULL;
+ atomic_dec(&efct_debugfs_count);
+
+ if (atomic_read(&efct_debugfs_count) == 0) {
+ /* remove root debugfs directory */
+ debugfs_remove(efct_debugfs_root);
+ efct_debugfs_root = NULL;
+ }
+}
+
+int
+efct_xport_attach(struct efct_xport *xport)
+{
+ struct efct *efct = xport->efct;
+ int rc;
+ u32 max_sgl;
+ u32 n_sgl;
+ u32 value;
+
+ xport->fcfi.hold_frames = true;
+ spin_lock_init(&xport->fcfi.pend_frames_lock);
+ INIT_LIST_HEAD(&xport->fcfi.pend_frames);
+
+ rc = efct_hw_setup(&efct->hw, efct, efct->pcidev);
+ if (rc) {
+ efc_log_err(efct, "%s: Can't setup hardware\n", efct->desc);
+ return -1;
+ }
+
+ efct_hw_set(&efct->hw, EFCT_HW_RQ_SELECTION_POLICY,
+ efct->rq_selection_policy);
+ efct_hw_get(&efct->hw, EFCT_HW_RQ_SELECTION_POLICY, &value);
+ efc_log_debug(efct, "RQ Selection Policy: %d\n", value);
+
+ efct_hw_set_ptr(&efct->hw, EFCT_HW_FILTER_DEF,
+ (void *)efct->filter_def);
+
+ efct_hw_get(&efct->hw, EFCT_HW_MAX_SGL, &max_sgl);
+ max_sgl -= SLI4_SGE_MAX_RESERVED;
+ n_sgl = (max_sgl > EFCT_FC_MAX_SGL) ? EFCT_FC_MAX_SGL : max_sgl;
+
+ /* Note: number of SGLs must be set for efc_node_create_pool */
+ if (efct_hw_set(&efct->hw, EFCT_HW_N_SGL, n_sgl) !=
+ EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(efct,
+ "%s: Can't set number of SGLs\n", efct->desc);
+ return -1;
+ }
+
+ efc_log_debug(efct, "%s: Configured for %d SGLs\n", efct->desc,
+ n_sgl);
+
+ xport->io_pool = efct_io_pool_create(efct, EFCT_NUM_SCSI_IOS, n_sgl);
+ if (!xport->io_pool) {
+ efc_log_err(efct, "Can't allocate IO pool\n");
+ return -1;
+ }
+
+ return 0;
+}
+
+static void
+efct_xport_config_stats_timer(struct efct *efct);
+
+static void
+efct_xport_stats_timer_cb(struct timer_list *t)
+{
+ struct efct_xport *xport = from_timer(xport, t, stats_timer);
+ struct efct *efct = xport->efct;
+
+ efct_xport_config_stats_timer(efct);
+}
+
+static void
+efct_xport_config_stats_timer(struct efct *efct)
+{
+ u32 timeout = 3 * 1000;
+ struct efct_xport *xport = NULL;
+
+ if (!efct) {
+ pr_err("%s: failed to locate EFCT device\n", __func__);
+ return;
+ }
+
+ xport = efct->xport;
+ efct_hw_get_link_stats(&efct->hw, 0, 0, 0,
+ efct_xport_async_link_stats_cb,
+ &xport->fc_xport_stats);
+ efct_hw_get_host_stats(&efct->hw, 0, efct_xport_async_host_stats_cb,
+ &xport->fc_xport_stats);
+
+ timer_setup(&xport->stats_timer,
+ &efct_xport_stats_timer_cb, 0);
+ mod_timer(&xport->stats_timer,
+ jiffies + msecs_to_jiffies(timeout));
+}
+
+int
+efct_xport_initialize(struct efct_xport *xport)
+{
+ struct efct *efct = xport->efct;
+ int rc;
+ u32 max_hw_io;
+ u32 max_sgl;
+ u32 rq_limit;
+
+ bool tgt_device_set = false;
+ bool hw_initialized = false;
+
+ efct_hw_get(&efct->hw, EFCT_HW_MAX_IO, &max_hw_io);
+ if (efct_hw_set(&efct->hw, EFCT_HW_N_IO, max_hw_io) !=
+ EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(efct, "%s: Can't set number of IOs\n",
+ efct->desc);
+ return -1;
+ }
+
+ efct_hw_get(&efct->hw, EFCT_HW_MAX_SGL, &max_sgl);
+ max_sgl -= SLI4_SGE_MAX_RESERVED;
+
+ efct_hw_get(&efct->hw, EFCT_HW_MAX_IO, &max_hw_io);
+
+ if (efct_hw_set(&efct->hw, EFCT_HW_TOPOLOGY, efct->topology) !=
+ EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(efct, "%s: Can't set the toplogy\n", efct->desc);
+ return -1;
+ }
+ efct_hw_set(&efct->hw, EFCT_HW_RQ_DEFAULT_BUFFER_SIZE,
+ EFCT_FC_RQ_SIZE_DEFAULT);
+
+ if (efct_hw_set(&efct->hw, EFCT_HW_LINK_SPEED, efct->speed) !=
+ EFCT_HW_RTN_SUCCESS) {
+ efc_log_err(efct, "%s: Can't set the link speed\n",
+ efct->desc);
+ return -1;
+ }
+
+ if (efct->target_io_timer_sec) {
+ efc_log_debug(efct, "setting target io timer=%d\n",
+ efct->target_io_timer_sec);
+ efct_hw_set(&efct->hw, EFCT_HW_EMULATE_TARGET_WQE_TIMEOUT,
+ true);
+ }
+
+ /* Initialize vport list */
+ INIT_LIST_HEAD(&xport->vport_list);
+ spin_lock_init(&xport->io_pending_lock);
+ INIT_LIST_HEAD(&xport->io_pending_list);
+ atomic_set(&xport->io_active_count, 0);
+ atomic_set(&xport->io_pending_count, 0);
+ atomic_set(&xport->io_total_free, 0);
+ atomic_set(&xport->io_total_pending, 0);
+ atomic_set(&xport->io_alloc_failed_count, 0);
+ atomic_set(&xport->io_pending_recursing, 0);
+ rc = efct_hw_init(&efct->hw);
+ if (rc) {
+ efc_log_err(efct, "efct_hw_init failure\n");
+ goto efct_xport_init_cleanup;
+ } else {
+ hw_initialized = true;
+ }
+
+ rq_limit = max_hw_io / 2;
+ if (efct_hw_set(&efct->hw, EFCT_HW_RQ_PROCESS_LIMIT, rq_limit) !=
+ EFCT_HW_RTN_SUCCESS)
+ efc_log_err(efct, "%s: Can't set the RQ process limit\n",
+ efct->desc);
+
+ rc = efct_scsi_tgt_new_device(efct);
+ if (rc) {
+ efc_log_err(efct, "failed to initialize target\n");
+ goto efct_xport_init_cleanup;
+ } else {
+ tgt_device_set = true;
+ }
+
+ rc = efct_scsi_new_device(efct);
+ if (rc) {
+ efc_log_err(efct, "failed to initialize initiator\n");
+ goto efct_xport_init_cleanup;
+ }
+
+ /* Get FC link and host statistics perodically*/
+ efct_xport_config_stats_timer(efct);
+
+ efct_xport_init_debugfs(efct);
+
+ return 0;
+
+efct_xport_init_cleanup:
+ if (tgt_device_set)
+ efct_scsi_tgt_del_device(efct);
+
+ if (hw_initialized) {
+ /* efct_hw_teardown can only execute after efct_hw_init */
+ efct_hw_teardown(&efct->hw);
+ }
+
+ return -1;
+}
+
+int
+efct_xport_status(struct efct_xport *xport, enum efct_xport_status cmd,
+ union efct_xport_stats_u *result)
+{
+ u32 rc = 0;
+ struct efct *efct = NULL;
+ union efct_xport_stats_u value;
+ enum efct_hw_rtn hw_rc;
+
+ efct = xport->efct;
+
+ switch (cmd) {
+ case EFCT_XPORT_CONFIG_PORT_STATUS:
+ if (xport->configured_link_state == 0) {
+ /*
+ * Initial state is offline. configured_link_state is
+ * set to online explicitly when port is brought online
+ */
+ xport->configured_link_state = EFCT_XPORT_PORT_OFFLINE;
+ }
+ result->value = xport->configured_link_state;
+ break;
+
+ case EFCT_XPORT_PORT_STATUS:
+ /* Determine port status based on link speed. */
+ hw_rc = efct_hw_get(&efct->hw, EFCT_HW_LINK_SPEED,
+ &value.value);
+ if (hw_rc == EFCT_HW_RTN_SUCCESS) {
+ if (value.value == 0)
+ result->value = 0;
+ else
+ result->value = 1;
+ rc = 0;
+ } else {
+ rc = -1;
+ }
+ break;
+
+ case EFCT_XPORT_LINK_SPEED: {
+ u32 speed;
+
+ result->value = 0;
+
+ rc = efct_hw_get(&efct->hw, EFCT_HW_LINK_SPEED, &speed);
+ if (rc == 0)
+ result->value = speed;
+ break;
+ }
+
+ case EFCT_XPORT_IS_SUPPORTED_LINK_SPEED: {
+ u32 speed;
+ u32 link_module_type;
+
+ speed = result->value;
+
+ rc = efct_hw_get(&efct->hw, EFCT_HW_LINK_MODULE_TYPE,
+ &link_module_type);
+ if (rc == 0) {
+ switch (speed) {
+ case 1000:
+ rc = (link_module_type &
+ EFCT_HW_LINK_MODULE_TYPE_1GB) != 0;
+ break;
+ case 2000:
+ rc = (link_module_type &
+ EFCT_HW_LINK_MODULE_TYPE_2GB) != 0;
+ break;
+ case 4000:
+ rc = (link_module_type &
+ EFCT_HW_LINK_MODULE_TYPE_4GB) != 0;
+ break;
+ case 8000:
+ rc = (link_module_type &
+ EFCT_HW_LINK_MODULE_TYPE_8GB) != 0;
+ break;
+ case 10000:
+ rc = (link_module_type &
+ EFCT_HW_LINK_MODULE_TYPE_10GB) != 0;
+ break;
+ case 16000:
+ rc = (link_module_type &
+ EFCT_HW_LINK_MODULE_TYPE_16GB) != 0;
+ break;
+ case 32000:
+ rc = (link_module_type &
+ EFCT_HW_LINK_MODULE_TYPE_32GB) != 0;
+ break;
+ default:
+ rc = 0;
+ break;
+ }
+ } else {
+ rc = 0;
+ }
+ break;
+ }
+ case EFCT_XPORT_LINK_STATISTICS:
+ memcpy((void *)result, &efct->xport->fc_xport_stats,
+ sizeof(union efct_xport_stats_u));
+ break;
+ case EFCT_XPORT_LINK_STAT_RESET: {
+ /* Create a completion to synchronize the stat reset process. */
+ init_completion(&result->stats.done);
+
+ /* First reset the link stats */
+ rc = efct_hw_get_link_stats(&efct->hw, 0, 1, 1,
+ efct_xport_link_stats_cb, result);
+
+ /* Wait for completion to be signaled when the cmd completes */
+ if (wait_for_completion_interruptible(&result->stats.done)) {
+ /* Undefined failure */
+ efc_log_test(efct, "sem wait failed\n");
+ rc = -ENXIO;
+ break;
+ }
+
+ /* Next reset the host stats */
+ rc = efct_hw_get_host_stats(&efct->hw, 1,
+ efct_xport_host_stats_cb, result);
+
+ /* Wait for completion to be signaled when the cmd completes */
+ if (wait_for_completion_interruptible(&result->stats.done)) {
+ /* Undefined failure */
+ efc_log_test(efct, "sem wait failed\n");
+ rc = -ENXIO;
+ break;
+ }
+ break;
+ }
+ default:
+ rc = -1;
+ break;
+ }
+
+ return rc;
+}
+
+int
+efct_scsi_new_device(struct efct *efct)
+{
+ struct Scsi_Host *shost = NULL;
+ int error = 0;
+ struct efct_vport *vport = NULL;
+ union efct_xport_stats_u speed;
+ u32 supported_speeds = 0;
+
+ shost = scsi_host_alloc(&efct_template, sizeof(*vport));
+ if (!shost) {
+ efc_log_err(efct, "failed to allocate Scsi_Host struct\n");
+ return -1;
+ }
+
+ /* save shost to initiator-client context */
+ efct->shost = shost;
+
+ /* save efct information to shost LLD-specific space */
+ vport = (struct efct_vport *)shost->hostdata;
+ vport->efct = efct;
+
+ /*
+ * Set initial can_queue value to the max SCSI IOs. This is the maximum
+ * global queue depth (as opposed to the per-LUN queue depth --
+ * .cmd_per_lun This may need to be adjusted for I+T mode.
+ */
+ shost->can_queue = efct_scsi_get_property(efct, EFCT_SCSI_MAX_IOS);
+ shost->max_cmd_len = 16; /* 16-byte CDBs */
+ shost->max_id = 0xffff;
+ shost->max_lun = 0xffffffff;
+
+ /*
+ * can only accept (from mid-layer) as many SGEs as we've
+ * pre-registered
+ */
+ shost->sg_tablesize = efct_scsi_get_property(efct, EFCT_SCSI_MAX_SGL);
+
+ /* attach FC Transport template to shost */
+ shost->transportt = efct_xport_fc_tt;
+ efc_log_debug(efct, "transport template=%p\n", efct_xport_fc_tt);
+
+ /* get pci_dev structure and add host to SCSI ML */
+ error = scsi_add_host_with_dma(shost, &efct->pcidev->dev,
+ &efct->pcidev->dev);
+ if (error) {
+ efc_log_test(efct, "failed scsi_add_host_with_dma\n");
+ return -1;
+ }
+
+ /* Set symbolic name for host port */
+ snprintf(fc_host_symbolic_name(shost),
+ sizeof(fc_host_symbolic_name(shost)),
+ "Emulex %s FV%s DV%s", efct->model,
+ efct->fw_version, efct->driver_version);
+
+ /* Set host port supported classes */
+ fc_host_supported_classes(shost) = FC_COS_CLASS3;
+
+ speed.value = 1000;
+ if (efct_xport_status(efct->xport, EFCT_XPORT_IS_SUPPORTED_LINK_SPEED,
+ &speed)) {
+ supported_speeds |= FC_PORTSPEED_1GBIT;
+ }
+ speed.value = 2000;
+ if (efct_xport_status(efct->xport, EFCT_XPORT_IS_SUPPORTED_LINK_SPEED,
+ &speed)) {
+ supported_speeds |= FC_PORTSPEED_2GBIT;
+ }
+ speed.value = 4000;
+ if (efct_xport_status(efct->xport, EFCT_XPORT_IS_SUPPORTED_LINK_SPEED,
+ &speed)) {
+ supported_speeds |= FC_PORTSPEED_4GBIT;
+ }
+ speed.value = 8000;
+ if (efct_xport_status(efct->xport, EFCT_XPORT_IS_SUPPORTED_LINK_SPEED,
+ &speed)) {
+ supported_speeds |= FC_PORTSPEED_8GBIT;
+ }
+ speed.value = 10000;
+ if (efct_xport_status(efct->xport, EFCT_XPORT_IS_SUPPORTED_LINK_SPEED,
+ &speed)) {
+ supported_speeds |= FC_PORTSPEED_10GBIT;
+ }
+ speed.value = 16000;
+ if (efct_xport_status(efct->xport, EFCT_XPORT_IS_SUPPORTED_LINK_SPEED,
+ &speed)) {
+ supported_speeds |= FC_PORTSPEED_16GBIT;
+ }
+ speed.value = 32000;
+ if (efct_xport_status(efct->xport, EFCT_XPORT_IS_SUPPORTED_LINK_SPEED,
+ &speed)) {
+ supported_speeds |= FC_PORTSPEED_32GBIT;
+ }
+
+ fc_host_supported_speeds(shost) = supported_speeds;
+
+ fc_host_node_name(shost) = efct_get_wwn(&efct->hw, EFCT_HW_WWN_NODE);
+ fc_host_port_name(shost) = efct_get_wwn(&efct->hw, EFCT_HW_WWN_PORT);
+ fc_host_max_npiv_vports(shost) = 128;
+
+ return 0;
+}
+
+struct scsi_transport_template *
+efct_attach_fc_transport(void)
+{
+ struct scsi_transport_template *efct_fc_template = NULL;
+
+ efct_fc_template = fc_attach_transport(&efct_xport_functions);
+
+ if (!efct_fc_template)
+ pr_err("failed to attach EFCT with fc transport\n");
+
+ return efct_fc_template;
+}
+
+struct scsi_transport_template *
+efct_attach_vport_fc_transport(void)
+{
+ struct scsi_transport_template *efct_fc_template = NULL;
+
+ efct_fc_template = fc_attach_transport(&efct_vport_functions);
+
+ if (!efct_fc_template)
+ pr_err("failed to attach EFCT with fc transport\n");
+
+ return efct_fc_template;
+}
+
+int
+efct_scsi_reg_fc_transport(void)
+{
+ /* attach to appropriate scsi_tranport_* module */
+ efct_xport_fc_tt = efct_attach_fc_transport();
+ if (!efct_xport_fc_tt) {
+ pr_err("%s: failed to attach to scsi_transport_*", __func__);
+ return -1;
+ }
+
+ efct_vport_fc_tt = efct_attach_vport_fc_transport();
+ if (!efct_vport_fc_tt) {
+ pr_err("%s: failed to attach to scsi_transport_*", __func__);
+ efct_release_fc_transport(efct_xport_fc_tt);
+ efct_xport_fc_tt = NULL;
+ return -1;
+ }
+
+ return 0;
+}
+
+int
+efct_scsi_release_fc_transport(void)
+{
+ /* detach from scsi_transport_* */
+ efct_release_fc_transport(efct_xport_fc_tt);
+ efct_xport_fc_tt = NULL;
+ if (efct_vport_fc_tt)
+ efct_release_fc_transport(efct_vport_fc_tt);
+ efct_vport_fc_tt = NULL;
+
+ return 0;
+}
diff --git a/drivers/scsi/elx/efct/efct_xport.h b/drivers/scsi/elx/efct/efct_xport.h
new file mode 100644
index 000000000000..c390aea8ff01
--- /dev/null
+++ b/drivers/scsi/elx/efct/efct_xport.h
@@ -0,0 +1,205 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+/*
+ * Copyright (C) 2019 Broadcom. All Rights Reserved. The term
+ * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
+ */
+
+#if !defined(__EFCT_XPORT_H__)
+#define __EFCT_XPORT_H__
+
+/* FCFI lookup/pending frames */
+struct efct_xport_fcfi {
+ /* lock to protect pending frames access*/
+ spinlock_t pend_frames_lock;
+ struct list_head pend_frames;
+ /* hold pending frames */
+ bool hold_frames;
+ /* count of pending frames that were processed */
+ u32 pend_frames_processed;
+};
+
+enum efct_xport_ctrl {
+ EFCT_XPORT_PORT_ONLINE = 1,
+ EFCT_XPORT_PORT_OFFLINE,
+ EFCT_XPORT_SHUTDOWN,
+ EFCT_XPORT_POST_NODE_EVENT,
+ EFCT_XPORT_WWNN_SET,
+ EFCT_XPORT_WWPN_SET,
+};
+
+enum efct_xport_status {
+ EFCT_XPORT_PORT_STATUS,
+ EFCT_XPORT_CONFIG_PORT_STATUS,
+ EFCT_XPORT_LINK_SPEED,
+ EFCT_XPORT_IS_SUPPORTED_LINK_SPEED,
+ EFCT_XPORT_LINK_STATISTICS,
+ EFCT_XPORT_LINK_STAT_RESET,
+ EFCT_XPORT_IS_QUIESCED
+};
+
+struct efct_xport_link_stats {
+ bool rec;
+ bool gec;
+ bool w02of;
+ bool w03of;
+ bool w04of;
+ bool w05of;
+ bool w06of;
+ bool w07of;
+ bool w08of;
+ bool w09of;
+ bool w10of;
+ bool w11of;
+ bool w12of;
+ bool w13of;
+ bool w14of;
+ bool w15of;
+ bool w16of;
+ bool w17of;
+ bool w18of;
+ bool w19of;
+ bool w20of;
+ bool w21of;
+ bool clrc;
+ bool clof1;
+ u32 link_failure_error_count;
+ u32 loss_of_sync_error_count;
+ u32 loss_of_signal_error_count;
+ u32 primitive_sequence_error_count;
+ u32 invalid_transmission_word_error_count;
+ u32 crc_error_count;
+ u32 primitive_sequence_event_timeout_count;
+ u32 elastic_buffer_overrun_error_count;
+ u32 arbitration_fc_al_timeout_count;
+ u32 advertised_receive_bufftor_to_buffer_credit;
+ u32 current_receive_buffer_to_buffer_credit;
+ u32 advertised_transmit_buffer_to_buffer_credit;
+ u32 current_transmit_buffer_to_buffer_credit;
+ u32 received_eofa_count;
+ u32 received_eofdti_count;
+ u32 received_eofni_count;
+ u32 received_soff_count;
+ u32 received_dropped_no_aer_count;
+ u32 received_dropped_no_available_rpi_resources_count;
+ u32 received_dropped_no_available_xri_resources_count;
+};
+
+struct efct_xport_host_stats {
+ bool cc;
+ u32 transmit_kbyte_count;
+ u32 receive_kbyte_count;
+ u32 transmit_frame_count;
+ u32 receive_frame_count;
+ u32 transmit_sequence_count;
+ u32 receive_sequence_count;
+ u32 total_exchanges_originator;
+ u32 total_exchanges_responder;
+ u32 receive_p_bsy_count;
+ u32 receive_f_bsy_count;
+ u32 dropped_frames_due_to_no_rq_buffer_count;
+ u32 empty_rq_timeout_count;
+ u32 dropped_frames_due_to_no_xri_count;
+ u32 empty_xri_pool_count;
+};
+
+struct efct_xport_host_statistics {
+ struct completion done;
+ struct efct_xport_link_stats link_stats;
+ struct efct_xport_host_stats host_stats;
+};
+
+union efct_xport_stats_u {
+ u32 value;
+ struct efct_xport_host_statistics stats;
+};
+
+struct efct_xport_fcp_stats {
+ u64 input_bytes;
+ u64 output_bytes;
+ u64 input_requests;
+ u64 output_requests;
+ u64 control_requests;
+};
+
+struct efct_xport {
+ struct efct *efct;
+ /* wwpn requested by user for primary sport */
+ u64 req_wwpn;
+ /* wwnn requested by user for primary sport */
+ u64 req_wwnn;
+
+ struct efct_xport_fcfi fcfi;
+
+ /* Nodes */
+ /* number of allocated nodes */
+ u32 nodes_count;
+ /* array of pointers to nodes */
+ struct efc_node **nodes;
+ /* linked list of free nodes */
+ struct list_head nodes_free_list;
+
+ /* Io pool and counts */
+ /* pointer to IO pool */
+ struct efct_io_pool *io_pool;
+ /* used to track how often IO pool is empty */
+ atomic_t io_alloc_failed_count;
+ /* lock for io_pending_list */
+ spinlock_t io_pending_lock;
+ /* list of IOs waiting for HW resources
+ * lock: xport->io_pending_lock
+ * link: efct_io_s->io_pending_link
+ */
+ struct list_head io_pending_list;
+ /* count of totals IOS allocated */
+ atomic_t io_total_alloc;
+ /* count of totals IOS free'd */
+ atomic_t io_total_free;
+ /* count of totals IOS that were pended */
+ atomic_t io_total_pending;
+ /* count of active IOS */
+ atomic_t io_active_count;
+ /* count of pending IOS */
+ atomic_t io_pending_count;
+ /* non-zero if efct_scsi_check_pending is executing */
+ atomic_t io_pending_recursing;
+
+ /* vport */
+ /* list of VPORTS (NPIV) */
+ struct list_head vport_list;
+
+ /* Port */
+ /* requested link state */
+ u32 configured_link_state;
+
+ /* Timer for Statistics */
+ struct timer_list stats_timer;
+ union efct_xport_stats_u fc_xport_stats;
+ struct efct_xport_fcp_stats fcp_stats;
+};
+
+struct efct_rport_data {
+ struct efc_node *node;
+};
+
+extern struct efct_xport *
+efct_xport_alloc(struct efct *efct);
+extern int
+efct_xport_attach(struct efct_xport *xport);
+extern int
+efct_xport_initialize(struct efct_xport *xport);
+extern int
+efct_xport_detach(struct efct_xport *xport);
+extern int
+efct_xport_control(struct efct_xport *xport, enum efct_xport_ctrl cmd, ...);
+extern int
+efct_xport_status(struct efct_xport *xport, enum efct_xport_status cmd,
+ union efct_xport_stats_u *result);
+extern void
+efct_xport_free(struct efct_xport *xport);
+
+struct scsi_transport_template *efct_attach_fc_transport(void);
+struct scsi_transport_template *efct_attach_vport_fc_transport(void);
+void
+efct_release_fc_transport(struct scsi_transport_template *transport_template);
+
+#endif /* __EFCT_XPORT_H__ */
--
2.13.7
next prev parent reply other threads:[~2019-12-20 22:37 UTC|newest]
Thread overview: 77+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-12-20 22:36 [PATCH v2 00/32] [NEW] efct: Broadcom (Emulex) FC Target driver James Smart
2019-12-20 22:36 ` [PATCH v2 01/32] elx: libefc_sli: SLI-4 register offsets and field definitions James Smart
2020-01-08 7:11 ` Hannes Reinecke
2020-01-09 0:59 ` James Smart
2019-12-20 22:36 ` [PATCH v2 02/32] elx: libefc_sli: SLI Descriptors and Queue entries James Smart
2020-01-08 7:24 ` Hannes Reinecke
2020-01-09 1:00 ` James Smart
2019-12-20 22:36 ` [PATCH v2 03/32] elx: libefc_sli: Data structures and defines for mbox commands James Smart
2020-01-08 7:32 ` Hannes Reinecke
2020-01-09 1:03 ` James Smart
2019-12-20 22:36 ` [PATCH v2 04/32] elx: libefc_sli: queue create/destroy/parse routines James Smart
2020-01-08 7:45 ` Hannes Reinecke
2020-01-09 1:04 ` James Smart
2019-12-20 22:36 ` [PATCH v2 05/32] elx: libefc_sli: Populate and post different WQEs James Smart
2020-01-08 7:54 ` Hannes Reinecke
2020-01-09 1:04 ` James Smart
2019-12-20 22:36 ` [PATCH v2 06/32] elx: libefc_sli: bmbx routines and SLI config commands James Smart
2020-01-08 8:05 ` Hannes Reinecke
2019-12-20 22:36 ` [PATCH v2 07/32] elx: libefc_sli: APIs to setup SLI library James Smart
2020-01-08 8:22 ` Hannes Reinecke
2020-01-09 1:29 ` James Smart
2019-12-20 22:36 ` [PATCH v2 08/32] elx: libefc: Generic state machine framework James Smart
2020-01-09 7:05 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 09/32] elx: libefc: Emulex FC discovery library APIs and definitions James Smart
2020-01-09 7:16 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 10/32] elx: libefc: FC Domain state machine interfaces James Smart
2020-01-09 7:27 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 11/32] elx: libefc: SLI and FC PORT " James Smart
2020-01-09 7:34 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 12/32] elx: libefc: Remote node " James Smart
2020-01-09 8:31 ` Hannes Reinecke
2020-01-09 9:57 ` Daniel Wagner
2019-12-20 22:37 ` [PATCH v2 13/32] elx: libefc: Fabric " James Smart
2020-01-09 8:34 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 14/32] elx: libefc: FC node ELS and state handling James Smart
2020-01-09 8:39 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 15/32] elx: efct: Data structures and defines for hw operations James Smart
2020-01-09 8:41 ` Hannes Reinecke
2019-12-20 22:37 ` James Smart [this message]
2020-01-09 9:01 ` [PATCH v2 16/32] elx: efct: Driver initialization routines Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 17/32] elx: efct: Hardware queues creation and deletion James Smart
2020-01-09 9:10 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 18/32] elx: efct: RQ buffer, memory pool allocation and deallocation APIs James Smart
2020-01-09 9:13 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 19/32] elx: efct: Hardware IO and SGL initialization James Smart
2020-01-09 9:22 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 20/32] elx: efct: Hardware queues processing James Smart
2020-01-09 9:24 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 21/32] elx: efct: Unsolicited FC frame processing routines James Smart
2020-01-09 9:26 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 22/32] elx: efct: Extended link Service IO handling James Smart
2020-01-09 9:38 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 23/32] elx: efct: SCSI IO handling routines James Smart
2020-01-09 9:41 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 24/32] elx: efct: LIO backend interface routines James Smart
2020-01-09 3:56 ` Bart Van Assche
2019-12-20 22:37 ` [PATCH v2 25/32] elx: efct: Hardware IO submission routines James Smart
2020-01-09 9:52 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 26/32] elx: efct: link statistics and SFP data James Smart
2020-01-09 10:12 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 27/32] elx: efct: xport and hardware teardown routines James Smart
2020-01-09 10:14 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 28/32] elx: efct: IO timeout handling routines James Smart
2020-01-09 11:27 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 29/32] elx: efct: Firmware update, async link processing James Smart
2020-01-09 11:45 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 30/32] elx: efct: scsi_transport_fc host interface support James Smart
2020-01-09 11:46 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 31/32] elx: efct: Add Makefile and Kconfig for efct driver James Smart
2019-12-20 23:17 ` Randy Dunlap
2020-01-09 11:47 ` Hannes Reinecke
2019-12-20 22:37 ` [PATCH v2 32/32] elx: efct: Tie into kernel Kconfig and build process James Smart
2019-12-24 7:45 ` kbuild test robot
2019-12-24 21:01 ` Nathan Chancellor
2019-12-25 16:09 ` James Smart
2020-01-09 11:47 ` Hannes Reinecke
2019-12-29 18:27 ` [PATCH v2 00/32] [NEW] efct: Broadcom (Emulex) FC Target driver Sebastian Herbszt
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20191220223723.26563-17-jsmart2021@gmail.com \
--to=jsmart2021@gmail.com \
--cc=bvanassche@acm.org \
--cc=dwagner@suse.de \
--cc=linux-scsi@vger.kernel.org \
--cc=maier@linux.ibm.com \
--cc=ram.vegesna@broadcom.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).