All of lore.kernel.org
 help / color / mirror / Atom feed
* WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
       [not found] <530228679.31424535.1488695516036.JavaMail.zimbra@redhat.com>
@ 2017-03-05  6:41     ` Yi Zhang
  0 siblings, 0 replies; 10+ messages in thread
From: Yi Zhang @ 2017-03-05  6:41 UTC (permalink / raw)
  To: linux-rdma-u79uwXL29TY76Z2rM5mHXA,
	linux-nvme-IAPFreCvJWM7uuMidbF8XUB+6BGkLq7r

Hi

I get bellow WARNING when trying to connect nvmet with nvmecli.

Steps I used:
On target:
1. Use nvmetcli setup nvmet with below json 
{
  "hosts": [
    {
      "nqn": "hostnqn"
    }
  ], 
  "ports": [
    {
      "addr": {
        "adrfam": "ipv4", 
        "traddr": "172.31.40.4", 
        "treq": "not specified", 
        "trsvcid": "1023", 
        "trtype": "rdma"
      }, 
      "portid": 2, 
      "referrals": [], 
      "subsystems": [
        "testnqn"
      ]
    }
  ], 
  "subsystems": [
    {
      "allowed_hosts": [], 
      "attr": {
        "allow_any_host": "1"
      }, 
      "namespaces": [
        {
          "device": {
            "nguid": "ef90689c-6c46-d44c-89c1-4067801309a8", 
            "path": "/dev/nullb0"
          }, 
          "enable": 1, 
          "nsid": 1
        }
      ], 
      "nqn": "testnqn"
    }
  ]
}
2. Initiator side:
#nvme connect-all -t rdma -a 172.31.40.4 -s 1023

3. check the kernel log on both target/initiator side.

kernel log:
[  242.494533] ocrdma0:Using VLAN with PFC is recommended
[  242.520244] ocrdma0:Using VLAN 0 for this connection
[  242.652599] ocrdma0:Using VLAN with PFC is recommended
[  242.676365] ocrdma0:Using VLAN 0 for this connection
[  242.700476] ocrdma0:Using VLAN with PFC is recommended
[  242.723497] ocrdma0:Using VLAN 0 for this connection
[  242.812331] nvme nvme0: new ctrl: NQN "nqn.2014-08.org.nvmexpress.discovery", addr 172.31.40.4:1023
[  242.854149] ocrdma0:Using VLAN with PFC is recommended
[  242.854149] ocrdma0:Using VLAN 0 for this connection
[  242.854662] ------------[ cut here ]------------
[  242.854671] WARNING: CPU: 2 PID: 158 at drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
[  242.854671] failed to drain send queue: -22
[  242.854671] Modules linked in: nvme_rdma nvme_fabrics nvme_core sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ocrdma ib_core ipmi_ssif intel_rapl x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_intel intel_cstate intel_uncore ipmi_si ipmi_devintf hpilo hpwdt intel_rapl_perf gpio_ich iTCO_wdt iTCO_vendor_support pcspkr ie31200_edac shpchp sg edac_core acpi_power_meter lpc_ich ipmi_msghandler pcc_cpufreq acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod tg3 ptp mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfill
 rect sysimgblt
[  242.854694]  fb_sys_fops ttm drm ahci libahci libata crc32c_intel be2net fjes serio_raw i2c_core pps_core dm_mirror dm_region_hash dm_log dm_mod
[  242.854699] CPU: 2 PID: 158 Comm: kworker/2:2 Not tainted 4.10.0 #2
[  242.854699] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 09/01/2013
[  242.854702] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  242.854702] Call Trace:
[  242.854706]  dump_stack+0x63/0x87
[  242.854711]  __warn+0xd1/0xf0
[  242.854712]  warn_slowpath_fmt+0x5f/0x80
[  242.854715]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  242.854718]  __ib_drain_sq+0x182/0x1c0 [ib_core]
[  242.854721]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  242.854724]  ib_drain_sq+0x25/0x30 [ib_core]
[  242.854727]  ib_drain_qp+0x12/0x30 [ib_core]
[  242.854728]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  242.854729]  nvme_rdma_destroy_admin_queue+0x4d/0xc0 [nvme_rdma]
[  242.854730]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  242.854731]  __nvme_rdma_remove_ctrl+0x84/0x90 [nvme_rdma]
[  242.854732]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  242.854733]  process_one_work+0x165/0x410
[  242.854734]  worker_thread+0x27f/0x4c0
[  242.854735]  kthread+0x101/0x140
[  242.854736]  ? rescuer_thread+0x3b0/0x3b0
[  242.854737]  ? kthread_park+0x90/0x90
[  242.854751]  ret_from_fork+0x2c/0x40
[  242.854752] ---[ end trace 87cf24af407fe063 ]---
[  242.855216] ------------[ cut here ]------------
[  242.855220] WARNING: CPU: 2 PID: 158 at drivers/infiniband/core/verbs.c:2010 __ib_drain_rq+0x177/0x1c0 [ib_core]
[  242.855220] failed to drain recv queue: -22
[  242.855220] Modules linked in: nvme_rdma nvme_fabrics nvme_core sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ocrdma ib_core ipmi_ssif intel_rapl x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_intel intel_cstate intel_uncore ipmi_si ipmi_devintf hpilo hpwdt intel_rapl_perf gpio_ich iTCO_wdt iTCO_vendor_support pcspkr ie31200_edac shpchp sg edac_core acpi_power_meter lpc_ich ipmi_msghandler pcc_cpufreq acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod tg3 ptp mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfill
 rect sysimgblt
[  242.855232]  fb_sys_fops ttm drm ahci libahci libata crc32c_intel be2net fjes serio_raw i2c_core pps_core dm_mirror dm_region_hash dm_log dm_mod
[  242.855235] CPU: 2 PID: 158 Comm: kworker/2:2 Tainted: G        W       4.10.0 #2
[  242.855235] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 09/01/2013
[  242.855237] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  242.855237] Call Trace:
[  242.855238]  dump_stack+0x63/0x87
[  242.855239]  __warn+0xd1/0xf0
[  242.855240]  warn_slowpath_fmt+0x5f/0x80
[  242.855242]  ? ocrdma_post_recv+0x127/0x140 [ocrdma]
[  242.855243]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  242.855246]  __ib_drain_rq+0x177/0x1c0 [ib_core]
[  242.855268]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  242.855271]  ib_drain_rq+0x25/0x30 [ib_core]
[  242.855273]  ib_drain_qp+0x24/0x30 [ib_core]
[  242.855274]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  242.855275]  nvme_rdma_destroy_admin_queue+0x4d/0xc0 [nvme_rdma]
[  242.855276]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  242.855277]  __nvme_rdma_remove_ctrl+0x84/0x90 [nvme_rdma]
[  242.855278]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  242.855279]  process_one_work+0x165/0x410
[  242.855280]  worker_thread+0x27f/0x4c0
[  242.855282]  kthread+0x101/0x140
[  242.855283]  ? rescuer_thread+0x3b0/0x3b0
[  242.855283]  ? kthread_park+0x90/0x90
[  242.855284]  ret_from_fork+0x2c/0x40
[  242.855285] ---[ end trace 87cf24af407fe064 ]---
[  244.970302] ocrdma0:Using VLAN with PFC is recommended
[  244.970302] ocrdma0:Using VLAN 0 for this connection
[  245.074424] ocrdma0:Using VLAN with PFC is recommended
[  245.098195] ocrdma0:Using VLAN 0 for this connection
[  245.122976] ocrdma0:Using VLAN with PFC is recommended
[  245.147078] ocrdma0:Using VLAN 0 for this connection
[  245.183682] nvme nvme0: creating 8 I/O queues.
[  245.242477] ocrdma0:Using VLAN with PFC is recommended
[  245.266737] ocrdma0:Using VLAN 0 for this connection
[  245.388146] ocrdma0:Using VLAN with PFC is recommended
[  245.411813] ocrdma0:Using VLAN 0 for this connection
[  245.436101] ocrdma0:Using VLAN with PFC is recommended
[  245.459634] ocrdma0:Using VLAN 0 for this connection
[  245.485689] ocrdma0:Using VLAN with PFC is recommended
[  245.512216] ocrdma0:Using VLAN 0 for this connection
[  245.662331] ocrdma0:Using VLAN with PFC is recommended
[  245.687422] ocrdma0:Using VLAN 0 for this connection
[  245.713057] ocrdma0:Using VLAN with PFC is recommended
[  245.738019] ocrdma0:Using VLAN 0 for this connection
[  245.766147] ocrdma0:Using VLAN with PFC is recommended
[  245.791930] ocrdma0:Using VLAN 0 for this connection
[  245.824506] ocrdma0:Using VLAN with PFC is recommended
[  245.849968] ocrdma0:Using VLAN 0 for this connection
[  245.876203] ocrdma0:Using VLAN with PFC is recommended
[  245.900670] ocrdma0:Using VLAN 0 for this connection
[  245.926549] ocrdma0:Using VLAN with PFC is recommended
[  245.949892] ocrdma0:Using VLAN 0 for this connection
[  246.102218] ocrdma0:Using VLAN with PFC is recommended
[  246.125597] ocrdma0:Using VLAN 0 for this connection
[  246.150668] ocrdma0:Using VLAN with PFC is recommended
[  246.174000] ocrdma0:Using VLAN 0 for this connection
[  246.199945] ocrdma0:Using VLAN with PFC is recommended
[  246.223247] ocrdma0:Using VLAN 0 for this connection
[  246.253534] ocrdma0:Using VLAN with PFC is recommended
[  246.278590] ocrdma0:Using VLAN 0 for this connection
[  246.303311] ocrdma0:Using VLAN with PFC is recommended
[  246.329326] ocrdma0:Using VLAN 0 for this connection
[  246.358014] ocrdma0:Using VLAN with PFC is recommended
[  246.382120] ocrdma0:Using VLAN 0 for this connection
[  246.537506] ocrdma0:Using VLAN with PFC is recommended
[  246.560855] ocrdma0:Using VLAN 0 for this connection
[  246.585018] ocrdma0:Using VLAN with PFC is recommended
[  246.608393] ocrdma0:Using VLAN 0 for this connection
[  246.634275] ocrdma0:Using VLAN with PFC is recommended
[  246.658760] ocrdma0:Using VLAN 0 for this connection
[  246.779071] ocrdma0:Using VLAN with PFC is recommended
[  246.803813] ocrdma0:Using VLAN 0 for this connection
[  246.829115] ocrdma0:Using VLAN with PFC is recommended
[  246.853383] ocrdma0:Using VLAN 0 for this connection
[  246.880570] ocrdma0:Using VLAN with PFC is recommended
[  246.905561] ocrdma0:Using VLAN 0 for this connection
[  247.033429] ocrdma0:Using VLAN with PFC is recommended
[  247.060779] ocrdma0:Using VLAN 0 for this connection
[  247.087520] ocrdma0:Using VLAN with PFC is recommended
[  247.111723] ocrdma0:Using VLAN 0 for this connection
[  247.701156] nvme nvme0: new ctrl: NQN "testnqn", addr 172.31.40.4:1023


Best Regards,
  Yi Zhang


--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 10+ messages in thread

* WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
@ 2017-03-05  6:41     ` Yi Zhang
  0 siblings, 0 replies; 10+ messages in thread
From: Yi Zhang @ 2017-03-05  6:41 UTC (permalink / raw)


Hi

I get bellow WARNING when trying to connect nvmet with nvmecli.

Steps I used:
On target:
1. Use nvmetcli setup nvmet with below json 
{
  "hosts": [
    {
      "nqn": "hostnqn"
    }
  ], 
  "ports": [
    {
      "addr": {
        "adrfam": "ipv4", 
        "traddr": "172.31.40.4", 
        "treq": "not specified", 
        "trsvcid": "1023", 
        "trtype": "rdma"
      }, 
      "portid": 2, 
      "referrals": [], 
      "subsystems": [
        "testnqn"
      ]
    }
  ], 
  "subsystems": [
    {
      "allowed_hosts": [], 
      "attr": {
        "allow_any_host": "1"
      }, 
      "namespaces": [
        {
          "device": {
            "nguid": "ef90689c-6c46-d44c-89c1-4067801309a8", 
            "path": "/dev/nullb0"
          }, 
          "enable": 1, 
          "nsid": 1
        }
      ], 
      "nqn": "testnqn"
    }
  ]
}
2. Initiator side:
#nvme connect-all -t rdma -a 172.31.40.4 -s 1023

3. check the kernel log on both target/initiator side.

kernel log:
[  242.494533] ocrdma0:Using VLAN with PFC is recommended
[  242.520244] ocrdma0:Using VLAN 0 for this connection
[  242.652599] ocrdma0:Using VLAN with PFC is recommended
[  242.676365] ocrdma0:Using VLAN 0 for this connection
[  242.700476] ocrdma0:Using VLAN with PFC is recommended
[  242.723497] ocrdma0:Using VLAN 0 for this connection
[  242.812331] nvme nvme0: new ctrl: NQN "nqn.2014-08.org.nvmexpress.discovery", addr 172.31.40.4:1023
[  242.854149] ocrdma0:Using VLAN with PFC is recommended
[  242.854149] ocrdma0:Using VLAN 0 for this connection
[  242.854662] ------------[ cut here ]------------
[  242.854671] WARNING: CPU: 2 PID: 158 at drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
[  242.854671] failed to drain send queue: -22
[  242.854671] Modules linked in: nvme_rdma nvme_fabrics nvme_core sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ocrdma ib_core ipmi_ssif intel_rapl x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_intel intel_cstate intel_uncore ipmi_si ipmi_devintf hpilo hpwdt intel_rapl_perf gpio_ich iTCO_wdt iTCO_vendor_support pcspkr ie31200_edac shpchp sg edac_core acpi_power_meter lpc_ich ipmi_msghandler pcc_cpufreq acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod tg3 ptp mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt
[  242.854694]  fb_sys_fops ttm drm ahci libahci libata crc32c_intel be2net fjes serio_raw i2c_core pps_core dm_mirror dm_region_hash dm_log dm_mod
[  242.854699] CPU: 2 PID: 158 Comm: kworker/2:2 Not tainted 4.10.0 #2
[  242.854699] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 09/01/2013
[  242.854702] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  242.854702] Call Trace:
[  242.854706]  dump_stack+0x63/0x87
[  242.854711]  __warn+0xd1/0xf0
[  242.854712]  warn_slowpath_fmt+0x5f/0x80
[  242.854715]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  242.854718]  __ib_drain_sq+0x182/0x1c0 [ib_core]
[  242.854721]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  242.854724]  ib_drain_sq+0x25/0x30 [ib_core]
[  242.854727]  ib_drain_qp+0x12/0x30 [ib_core]
[  242.854728]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  242.854729]  nvme_rdma_destroy_admin_queue+0x4d/0xc0 [nvme_rdma]
[  242.854730]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  242.854731]  __nvme_rdma_remove_ctrl+0x84/0x90 [nvme_rdma]
[  242.854732]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  242.854733]  process_one_work+0x165/0x410
[  242.854734]  worker_thread+0x27f/0x4c0
[  242.854735]  kthread+0x101/0x140
[  242.854736]  ? rescuer_thread+0x3b0/0x3b0
[  242.854737]  ? kthread_park+0x90/0x90
[  242.854751]  ret_from_fork+0x2c/0x40
[  242.854752] ---[ end trace 87cf24af407fe063 ]---
[  242.855216] ------------[ cut here ]------------
[  242.855220] WARNING: CPU: 2 PID: 158 at drivers/infiniband/core/verbs.c:2010 __ib_drain_rq+0x177/0x1c0 [ib_core]
[  242.855220] failed to drain recv queue: -22
[  242.855220] Modules linked in: nvme_rdma nvme_fabrics nvme_core sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm ocrdma ib_core ipmi_ssif intel_rapl x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_intel intel_cstate intel_uncore ipmi_si ipmi_devintf hpilo hpwdt intel_rapl_perf gpio_ich iTCO_wdt iTCO_vendor_support pcspkr ie31200_edac shpchp sg edac_core acpi_power_meter lpc_ich ipmi_msghandler pcc_cpufreq acpi_cpufreq nfsd auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod tg3 ptp mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt
[  242.855232]  fb_sys_fops ttm drm ahci libahci libata crc32c_intel be2net fjes serio_raw i2c_core pps_core dm_mirror dm_region_hash dm_log dm_mod
[  242.855235] CPU: 2 PID: 158 Comm: kworker/2:2 Tainted: G        W       4.10.0 #2
[  242.855235] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 09/01/2013
[  242.855237] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  242.855237] Call Trace:
[  242.855238]  dump_stack+0x63/0x87
[  242.855239]  __warn+0xd1/0xf0
[  242.855240]  warn_slowpath_fmt+0x5f/0x80
[  242.855242]  ? ocrdma_post_recv+0x127/0x140 [ocrdma]
[  242.855243]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  242.855246]  __ib_drain_rq+0x177/0x1c0 [ib_core]
[  242.855268]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  242.855271]  ib_drain_rq+0x25/0x30 [ib_core]
[  242.855273]  ib_drain_qp+0x24/0x30 [ib_core]
[  242.855274]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  242.855275]  nvme_rdma_destroy_admin_queue+0x4d/0xc0 [nvme_rdma]
[  242.855276]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  242.855277]  __nvme_rdma_remove_ctrl+0x84/0x90 [nvme_rdma]
[  242.855278]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  242.855279]  process_one_work+0x165/0x410
[  242.855280]  worker_thread+0x27f/0x4c0
[  242.855282]  kthread+0x101/0x140
[  242.855283]  ? rescuer_thread+0x3b0/0x3b0
[  242.855283]  ? kthread_park+0x90/0x90
[  242.855284]  ret_from_fork+0x2c/0x40
[  242.855285] ---[ end trace 87cf24af407fe064 ]---
[  244.970302] ocrdma0:Using VLAN with PFC is recommended
[  244.970302] ocrdma0:Using VLAN 0 for this connection
[  245.074424] ocrdma0:Using VLAN with PFC is recommended
[  245.098195] ocrdma0:Using VLAN 0 for this connection
[  245.122976] ocrdma0:Using VLAN with PFC is recommended
[  245.147078] ocrdma0:Using VLAN 0 for this connection
[  245.183682] nvme nvme0: creating 8 I/O queues.
[  245.242477] ocrdma0:Using VLAN with PFC is recommended
[  245.266737] ocrdma0:Using VLAN 0 for this connection
[  245.388146] ocrdma0:Using VLAN with PFC is recommended
[  245.411813] ocrdma0:Using VLAN 0 for this connection
[  245.436101] ocrdma0:Using VLAN with PFC is recommended
[  245.459634] ocrdma0:Using VLAN 0 for this connection
[  245.485689] ocrdma0:Using VLAN with PFC is recommended
[  245.512216] ocrdma0:Using VLAN 0 for this connection
[  245.662331] ocrdma0:Using VLAN with PFC is recommended
[  245.687422] ocrdma0:Using VLAN 0 for this connection
[  245.713057] ocrdma0:Using VLAN with PFC is recommended
[  245.738019] ocrdma0:Using VLAN 0 for this connection
[  245.766147] ocrdma0:Using VLAN with PFC is recommended
[  245.791930] ocrdma0:Using VLAN 0 for this connection
[  245.824506] ocrdma0:Using VLAN with PFC is recommended
[  245.849968] ocrdma0:Using VLAN 0 for this connection
[  245.876203] ocrdma0:Using VLAN with PFC is recommended
[  245.900670] ocrdma0:Using VLAN 0 for this connection
[  245.926549] ocrdma0:Using VLAN with PFC is recommended
[  245.949892] ocrdma0:Using VLAN 0 for this connection
[  246.102218] ocrdma0:Using VLAN with PFC is recommended
[  246.125597] ocrdma0:Using VLAN 0 for this connection
[  246.150668] ocrdma0:Using VLAN with PFC is recommended
[  246.174000] ocrdma0:Using VLAN 0 for this connection
[  246.199945] ocrdma0:Using VLAN with PFC is recommended
[  246.223247] ocrdma0:Using VLAN 0 for this connection
[  246.253534] ocrdma0:Using VLAN with PFC is recommended
[  246.278590] ocrdma0:Using VLAN 0 for this connection
[  246.303311] ocrdma0:Using VLAN with PFC is recommended
[  246.329326] ocrdma0:Using VLAN 0 for this connection
[  246.358014] ocrdma0:Using VLAN with PFC is recommended
[  246.382120] ocrdma0:Using VLAN 0 for this connection
[  246.537506] ocrdma0:Using VLAN with PFC is recommended
[  246.560855] ocrdma0:Using VLAN 0 for this connection
[  246.585018] ocrdma0:Using VLAN with PFC is recommended
[  246.608393] ocrdma0:Using VLAN 0 for this connection
[  246.634275] ocrdma0:Using VLAN with PFC is recommended
[  246.658760] ocrdma0:Using VLAN 0 for this connection
[  246.779071] ocrdma0:Using VLAN with PFC is recommended
[  246.803813] ocrdma0:Using VLAN 0 for this connection
[  246.829115] ocrdma0:Using VLAN with PFC is recommended
[  246.853383] ocrdma0:Using VLAN 0 for this connection
[  246.880570] ocrdma0:Using VLAN with PFC is recommended
[  246.905561] ocrdma0:Using VLAN 0 for this connection
[  247.033429] ocrdma0:Using VLAN with PFC is recommended
[  247.060779] ocrdma0:Using VLAN 0 for this connection
[  247.087520] ocrdma0:Using VLAN with PFC is recommended
[  247.111723] ocrdma0:Using VLAN 0 for this connection
[  247.701156] nvme nvme0: new ctrl: NQN "testnqn", addr 172.31.40.4:1023


Best Regards,
  Yi Zhang

^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
  2017-03-05  6:41     ` Yi Zhang
@ 2017-03-06 11:42         ` Sagi Grimberg
  -1 siblings, 0 replies; 10+ messages in thread
From: Sagi Grimberg @ 2017-03-06 11:42 UTC (permalink / raw)
  To: Yi Zhang, linux-rdma-u79uwXL29TY76Z2rM5mHXA,
	linux-nvme-IAPFreCvJWM7uuMidbF8XUB+6BGkLq7r, Devesh Sharma


> 2. Initiator side:
> #nvme connect-all -t rdma -a 172.31.40.4 -s 1023
>
> 3. check the kernel log on both target/initiator side.
>
> kernel log:
> [  242.494533] ocrdma0:Using VLAN with PFC is recommended
> [  242.520244] ocrdma0:Using VLAN 0 for this connection
> [  242.652599] ocrdma0:Using VLAN with PFC is recommended
> [  242.676365] ocrdma0:Using VLAN 0 for this connection
> [  242.700476] ocrdma0:Using VLAN with PFC is recommended
> [  242.723497] ocrdma0:Using VLAN 0 for this connection
> [  242.812331] nvme nvme0: new ctrl: NQN "nqn.2014-08.org.nvmexpress.discovery", addr 172.31.40.4:1023
> [  242.854149] ocrdma0:Using VLAN with PFC is recommended
> [  242.854149] ocrdma0:Using VLAN 0 for this connection
> [  242.854662] ------------[ cut here ]------------
> [  242.854671] WARNING: CPU: 2 PID: 158 at drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]



I suspect that ib_drain_sq is not supported on ocrdma. From looking at
the code it seems that ocrdma fails and send post when the qp is not
in RTS. I think that if this is the case ocrdma needs to implement
its own qp drain logic similar to cxgb4, ns, i40e etc.

Devesh (CC'd), has anyone tested ib_drain_qp on ocrdma?
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 10+ messages in thread

* WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
@ 2017-03-06 11:42         ` Sagi Grimberg
  0 siblings, 0 replies; 10+ messages in thread
From: Sagi Grimberg @ 2017-03-06 11:42 UTC (permalink / raw)



> 2. Initiator side:
> #nvme connect-all -t rdma -a 172.31.40.4 -s 1023
>
> 3. check the kernel log on both target/initiator side.
>
> kernel log:
> [  242.494533] ocrdma0:Using VLAN with PFC is recommended
> [  242.520244] ocrdma0:Using VLAN 0 for this connection
> [  242.652599] ocrdma0:Using VLAN with PFC is recommended
> [  242.676365] ocrdma0:Using VLAN 0 for this connection
> [  242.700476] ocrdma0:Using VLAN with PFC is recommended
> [  242.723497] ocrdma0:Using VLAN 0 for this connection
> [  242.812331] nvme nvme0: new ctrl: NQN "nqn.2014-08.org.nvmexpress.discovery", addr 172.31.40.4:1023
> [  242.854149] ocrdma0:Using VLAN with PFC is recommended
> [  242.854149] ocrdma0:Using VLAN 0 for this connection
> [  242.854662] ------------[ cut here ]------------
> [  242.854671] WARNING: CPU: 2 PID: 158 at drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]



I suspect that ib_drain_sq is not supported on ocrdma. From looking at
the code it seems that ocrdma fails and send post when the qp is not
in RTS. I think that if this is the case ocrdma needs to implement
its own qp drain logic similar to cxgb4, ns, i40e etc.

Devesh (CC'd), has anyone tested ib_drain_qp on ocrdma?

^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
  2017-03-06 11:42         ` Sagi Grimberg
@ 2017-03-06 12:00             ` Devesh Sharma
  -1 siblings, 0 replies; 10+ messages in thread
From: Devesh Sharma @ 2017-03-06 12:00 UTC (permalink / raw)
  To: Sagi Grimberg
  Cc: Yi Zhang, linux-rdma,
	linux-nvme-IAPFreCvJWM7uuMidbF8XUB+6BGkLq7r, Devesh Sharma

Hi Sagi

On Mon, Mar 6, 2017 at 5:12 PM, Sagi Grimberg <sagi-NQWnxTmZq1alnMjI0IkVqw@public.gmane.org> wrote:
>
>> 2. Initiator side:
>> #nvme connect-all -t rdma -a 172.31.40.4 -s 1023
>>
>> 3. check the kernel log on both target/initiator side.
>>
>> kernel log:
>> [  242.494533] ocrdma0:Using VLAN with PFC is recommended
>> [  242.520244] ocrdma0:Using VLAN 0 for this connection
>> [  242.652599] ocrdma0:Using VLAN with PFC is recommended
>> [  242.676365] ocrdma0:Using VLAN 0 for this connection
>> [  242.700476] ocrdma0:Using VLAN with PFC is recommended
>> [  242.723497] ocrdma0:Using VLAN 0 for this connection
>> [  242.812331] nvme nvme0: new ctrl: NQN
>> "nqn.2014-08.org.nvmexpress.discovery", addr 172.31.40.4:1023
>> [  242.854149] ocrdma0:Using VLAN with PFC is recommended
>> [  242.854149] ocrdma0:Using VLAN 0 for this connection
>> [  242.854662] ------------[ cut here ]------------
>> [  242.854671] WARNING: CPU: 2 PID: 158 at
>> drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
>
>
>
>
> I suspect that ib_drain_sq is not supported on ocrdma. From looking at
> the code it seems that ocrdma fails and send post when the qp is not
> in RTS. I think that if this is the case ocrdma needs to implement
> its own qp drain logic similar to cxgb4, ns, i40e etc.
>
> Devesh (CC'd), has anyone tested ib_drain_qp on ocrdma?
No, ib_drain_sq and ib_drain_rq are not implemented as of now in ocrdma.
--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 10+ messages in thread

* WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
@ 2017-03-06 12:00             ` Devesh Sharma
  0 siblings, 0 replies; 10+ messages in thread
From: Devesh Sharma @ 2017-03-06 12:00 UTC (permalink / raw)


Hi Sagi

On Mon, Mar 6, 2017@5:12 PM, Sagi Grimberg <sagi@grimberg.me> wrote:
>
>> 2. Initiator side:
>> #nvme connect-all -t rdma -a 172.31.40.4 -s 1023
>>
>> 3. check the kernel log on both target/initiator side.
>>
>> kernel log:
>> [  242.494533] ocrdma0:Using VLAN with PFC is recommended
>> [  242.520244] ocrdma0:Using VLAN 0 for this connection
>> [  242.652599] ocrdma0:Using VLAN with PFC is recommended
>> [  242.676365] ocrdma0:Using VLAN 0 for this connection
>> [  242.700476] ocrdma0:Using VLAN with PFC is recommended
>> [  242.723497] ocrdma0:Using VLAN 0 for this connection
>> [  242.812331] nvme nvme0: new ctrl: NQN
>> "nqn.2014-08.org.nvmexpress.discovery", addr 172.31.40.4:1023
>> [  242.854149] ocrdma0:Using VLAN with PFC is recommended
>> [  242.854149] ocrdma0:Using VLAN 0 for this connection
>> [  242.854662] ------------[ cut here ]------------
>> [  242.854671] WARNING: CPU: 2 PID: 158 at
>> drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
>
>
>
>
> I suspect that ib_drain_sq is not supported on ocrdma. From looking at
> the code it seems that ocrdma fails and send post when the qp is not
> in RTS. I think that if this is the case ocrdma needs to implement
> its own qp drain logic similar to cxgb4, ns, i40e etc.
>
> Devesh (CC'd), has anyone tested ib_drain_qp on ocrdma?
No, ib_drain_sq and ib_drain_rq are not implemented as of now in ocrdma.

^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
  2017-03-05  6:41     ` Yi Zhang
@ 2017-03-25  1:44         ` Doug Ledford
  -1 siblings, 0 replies; 10+ messages in thread
From: Doug Ledford @ 2017-03-25  1:44 UTC (permalink / raw)
  To: Yi Zhang, linux-rdma-u79uwXL29TY76Z2rM5mHXA,
	linux-nvme-IAPFreCvJWM7uuMidbF8XUB+6BGkLq7r

On Sun, 2017-03-05 at 01:41 -0500, Yi Zhang wrote:
> Hi
> 
> I get bellow WARNING when trying to connect nvmet with nvmecli.
> 
> Steps I used:
> On target:
> 1. Use nvmetcli setup nvmet with below json 
> {
>   "hosts": [
>     {
>       "nqn": "hostnqn"
>     }
>   ], 
>   "ports": [
>     {
>       "addr": {
>         "adrfam": "ipv4", 
>         "traddr": "172.31.40.4", 

For testing ocrdma, you need to use the 172.31.45.0 network vlan.  For
testing other RoCE cards in our lab, you can use either the 43 or 45
vlans, but you shouldn't use the untagged interface for RoCE tests.

-- 
Doug Ledford <dledford-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
    GPG KeyID: B826A3330E572FDD
   
Key fingerprint = AE6B 1BDA 122B 23B4 265B  1274 B826 A333 0E57 2FDD

--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 10+ messages in thread

* WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
@ 2017-03-25  1:44         ` Doug Ledford
  0 siblings, 0 replies; 10+ messages in thread
From: Doug Ledford @ 2017-03-25  1:44 UTC (permalink / raw)


On Sun, 2017-03-05@01:41 -0500, Yi Zhang wrote:
> Hi
> 
> I get bellow WARNING when trying to connect nvmet with nvmecli.
> 
> Steps I used:
> On target:
> 1. Use nvmetcli setup nvmet with below json?
> {
> ? "hosts": [
> ????{
> ??????"nqn": "hostnqn"
> ????}
> ? ],?
> ? "ports": [
> ????{
> ??????"addr": {
> ????????"adrfam": "ipv4",?
> ????????"traddr": "172.31.40.4",?

For testing ocrdma, you need to use the 172.31.45.0 network vlan. ?For
testing other RoCE cards in our lab, you can use either the 43 or 45
vlans, but you shouldn't use the untagged interface for RoCE tests.

-- 
Doug Ledford <dledford at redhat.com>
? ? GPG KeyID: B826A3330E572FDD
? ?
Key fingerprint = AE6B 1BDA 122B 23B4 265B ?1274 B826 A333 0E57 2FDD

^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
  2017-03-25  1:44         ` Doug Ledford
@ 2017-03-27  3:18             ` Yi Zhang
  -1 siblings, 0 replies; 10+ messages in thread
From: Yi Zhang @ 2017-03-27  3:18 UTC (permalink / raw)
  To: Doug Ledford, linux-rdma-u79uwXL29TY76Z2rM5mHXA,
	linux-nvme-IAPFreCvJWM7uuMidbF8XUB+6BGkLq7r



On 03/25/2017 09:44 AM, Doug Ledford wrote:
> On Sun, 2017-03-05 at 01:41 -0500, Yi Zhang wrote:
>> Hi
>>
>> I get bellow WARNING when trying to connect nvmet with nvmecli.
>>
>> Steps I used:
>> On target:
>> 1. Use nvmetcli setup nvmet with below json
>> {
>>    "hosts": [
>>      {
>>        "nqn": "hostnqn"
>>      }
>>    ],
>>    "ports": [
>>      {
>>        "addr": {
>>          "adrfam": "ipv4",
>>          "traddr": "172.31.40.4",
> For testing ocrdma, you need to use the 172.31.45.0 network vlan.  For
> testing other RoCE cards in our lab, you can use either the 43 or 45
> vlans, but you shouldn't use the untagged interface for RoCE tests.
>
Hi Doug
Thanks for your reminder.
I've tried .45 network vlan with latest upstream, still can get below 
WARNING.

[  232.040920] nvme nvme0: new ctrl: NQN 
"nqn.2014-08.org.nvmexpress.discovery", addr 172.31.45.4:4420
[  232.088354] ------------[ cut here ]------------
[  232.111557] WARNING: CPU: 1 PID: 156 at 
drivers/infiniband/core/verbs.c:1969 __ib_drain_sq+0x16a/0x1b0 [ib_core]
[  232.157379] failed to drain send queue: -22
[  232.176836] Modules linked in: nvme_rdma nvme_fabrics nvme_core 
sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod 
ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp 
scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm 
ib_cm iw_cm ocrdma ib_core intel_rapl x86_pkg_temp_thermal 
intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul 
ipmi_ssif crc32_pclmul gpio_ich hpilo hpwdt iTCO_wdt iTCO_vendor_support 
ghash_clmulni_intel intel_cstate intel_uncore pcc_cpufreq 
intel_rapl_perf ie31200_edac sg shpchp acpi_power_meter acpi_cpufreq 
pcspkr ipmi_si ipmi_devintf edac_core ipmi_msghandler lpc_ich nfsd 
auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod 
mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt 
fb_sys_fops
[  232.511718]  ttm tg3 ahci libahci drm ptp libata crc32c_intel be2net 
i2c_core serio_raw pps_core dm_mirror dm_region_hash dm_log dm_mod
[  232.566744] CPU: 1 PID: 156 Comm: kworker/1:2 Not tainted 4.11.0-rc4 #1
[  232.600307] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 
09/01/2013
[  232.634737] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  232.666109] Call Trace:
[  232.677051]  dump_stack+0x63/0x87
[  232.691930]  __warn+0xd1/0xf0
[  232.705262]  warn_slowpath_fmt+0x5f/0x80
[  232.723238]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  232.747605]  __ib_drain_sq+0x16a/0x1b0 [ib_core]
[  232.768021]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  232.789971]  ib_drain_sq+0x25/0x30 [ib_core]
[  232.809144]  ib_drain_qp+0x12/0x30 [ib_core]
[  232.828273]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  232.855757]  nvme_rdma_destroy_admin_queue+0x60/0xb0 [nvme_rdma]
[  232.882317]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  232.908993]  __nvme_rdma_remove_ctrl+0x8c/0x90 [nvme_rdma]
[  232.933602]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  232.958178]  process_one_work+0x165/0x410
[  232.976871]  worker_thread+0x27f/0x4c0
[  232.993930]  kthread+0x101/0x140
[  233.008415]  ? rescuer_thread+0x3b0/0x3b0
[  233.026434]  ? kthread_park+0x90/0x90
[  233.042937]  ret_from_fork+0x2c/0x40
[  233.059136] ---[ end trace 2da8cf1943c3a50f ]---
[  233.105785] ------------[ cut here ]------------
[  233.128911] WARNING: CPU: 1 PID: 156 at 
drivers/infiniband/core/verbs.c:2003 __ib_drain_rq+0x15f/0x1b0 [ib_core]
[  233.178754] failed to drain recv queue: -22
[  233.197710] Modules linked in: nvme_rdma nvme_fabrics nvme_core 
sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod 
ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp 
scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm 
ib_cm iw_cm ocrdma ib_core intel_rapl x86_pkg_temp_thermal 
intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul 
ipmi_ssif crc32_pclmul gpio_ich hpilo hpwdt iTCO_wdt iTCO_vendor_support 
ghash_clmulni_intel intel_cstate intel_uncore pcc_cpufreq 
intel_rapl_perf ie31200_edac sg shpchp acpi_power_meter acpi_cpufreq 
pcspkr ipmi_si ipmi_devintf edac_core ipmi_msghandler lpc_ich nfsd 
auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod 
mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt 
fb_sys_fops
[  233.526083]  ttm tg3 ahci libahci drm ptp libata crc32c_intel be2net 
i2c_core serio_raw pps_core dm_mirror dm_region_hash dm_log dm_mod
[  233.585357] CPU: 1 PID: 156 Comm: kworker/1:2 Tainted: G W       
4.11.0-rc4 #1
[  233.623342] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 
09/01/2013
[  233.658771] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  233.691614] Call Trace:
[  233.704155]  dump_stack+0x63/0x87
[  233.720829]  __warn+0xd1/0xf0
[  233.736661]  warn_slowpath_fmt+0x5f/0x80
[  233.756792]  ? ocrdma_post_recv+0x127/0x140 [ocrdma]
[  233.781768]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  233.806196]  __ib_drain_rq+0x15f/0x1b0 [ib_core]
[  233.827474]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  233.849871]  ib_drain_rq+0x25/0x30 [ib_core]
[  233.869660]  ib_drain_qp+0x24/0x30 [ib_core]
[  233.889312]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  233.918097]  nvme_rdma_destroy_admin_queue+0x60/0xb0 [nvme_rdma]
[  233.945818]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  233.970971]  __nvme_rdma_remove_ctrl+0x8c/0x90 [nvme_rdma]
[  233.996145]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  234.021216]  process_one_work+0x165/0x410
[  234.040244]  worker_thread+0x27f/0x4c0
[  234.057532]  kthread+0x101/0x140
[  234.072476]  ? rescuer_thread+0x3b0/0x3b0
[  234.090911]  ? kthread_park+0x90/0x90
[  234.107890]  ret_from_fork+0x2c/0x40
[  234.124422] ---[ end trace 2da8cf1943c3a510 ]---
[  234.491570] nvme nvme0: creating 8 I/O queues.
[  235.099065] nvme nvme0: new ctrl: NQN "testnqn", addr 172.31.45.4:4420


--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo-u79uwXL29TY76Z2rM5mHXA@public.gmane.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 10+ messages in thread

* WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core]
@ 2017-03-27  3:18             ` Yi Zhang
  0 siblings, 0 replies; 10+ messages in thread
From: Yi Zhang @ 2017-03-27  3:18 UTC (permalink / raw)




On 03/25/2017 09:44 AM, Doug Ledford wrote:
> On Sun, 2017-03-05@01:41 -0500, Yi Zhang wrote:
>> Hi
>>
>> I get bellow WARNING when trying to connect nvmet with nvmecli.
>>
>> Steps I used:
>> On target:
>> 1. Use nvmetcli setup nvmet with below json
>> {
>>    "hosts": [
>>      {
>>        "nqn": "hostnqn"
>>      }
>>    ],
>>    "ports": [
>>      {
>>        "addr": {
>>          "adrfam": "ipv4",
>>          "traddr": "172.31.40.4",
> For testing ocrdma, you need to use the 172.31.45.0 network vlan.  For
> testing other RoCE cards in our lab, you can use either the 43 or 45
> vlans, but you shouldn't use the untagged interface for RoCE tests.
>
Hi Doug
Thanks for your reminder.
I've tried .45 network vlan with latest upstream, still can get below 
WARNING.

[  232.040920] nvme nvme0: new ctrl: NQN 
"nqn.2014-08.org.nvmexpress.discovery", addr 172.31.45.4:4420
[  232.088354] ------------[ cut here ]------------
[  232.111557] WARNING: CPU: 1 PID: 156 at 
drivers/infiniband/core/verbs.c:1969 __ib_drain_sq+0x16a/0x1b0 [ib_core]
[  232.157379] failed to drain send queue: -22
[  232.176836] Modules linked in: nvme_rdma nvme_fabrics nvme_core 
sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod 
ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp 
scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm 
ib_cm iw_cm ocrdma ib_core intel_rapl x86_pkg_temp_thermal 
intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul 
ipmi_ssif crc32_pclmul gpio_ich hpilo hpwdt iTCO_wdt iTCO_vendor_support 
ghash_clmulni_intel intel_cstate intel_uncore pcc_cpufreq 
intel_rapl_perf ie31200_edac sg shpchp acpi_power_meter acpi_cpufreq 
pcspkr ipmi_si ipmi_devintf edac_core ipmi_msghandler lpc_ich nfsd 
auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod 
mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt 
fb_sys_fops
[  232.511718]  ttm tg3 ahci libahci drm ptp libata crc32c_intel be2net 
i2c_core serio_raw pps_core dm_mirror dm_region_hash dm_log dm_mod
[  232.566744] CPU: 1 PID: 156 Comm: kworker/1:2 Not tainted 4.11.0-rc4 #1
[  232.600307] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 
09/01/2013
[  232.634737] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  232.666109] Call Trace:
[  232.677051]  dump_stack+0x63/0x87
[  232.691930]  __warn+0xd1/0xf0
[  232.705262]  warn_slowpath_fmt+0x5f/0x80
[  232.723238]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  232.747605]  __ib_drain_sq+0x16a/0x1b0 [ib_core]
[  232.768021]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  232.789971]  ib_drain_sq+0x25/0x30 [ib_core]
[  232.809144]  ib_drain_qp+0x12/0x30 [ib_core]
[  232.828273]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  232.855757]  nvme_rdma_destroy_admin_queue+0x60/0xb0 [nvme_rdma]
[  232.882317]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  232.908993]  __nvme_rdma_remove_ctrl+0x8c/0x90 [nvme_rdma]
[  232.933602]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  232.958178]  process_one_work+0x165/0x410
[  232.976871]  worker_thread+0x27f/0x4c0
[  232.993930]  kthread+0x101/0x140
[  233.008415]  ? rescuer_thread+0x3b0/0x3b0
[  233.026434]  ? kthread_park+0x90/0x90
[  233.042937]  ret_from_fork+0x2c/0x40
[  233.059136] ---[ end trace 2da8cf1943c3a50f ]---
[  233.105785] ------------[ cut here ]------------
[  233.128911] WARNING: CPU: 1 PID: 156 at 
drivers/infiniband/core/verbs.c:2003 __ib_drain_rq+0x15f/0x1b0 [ib_core]
[  233.178754] failed to drain recv queue: -22
[  233.197710] Modules linked in: nvme_rdma nvme_fabrics nvme_core 
sch_mqprio 8021q garp mrp stp llc rpcrdma ib_isert iscsi_target_mod 
ib_iser libiscsi scsi_transport_iscsi ib_srpt target_core_mod ib_srp 
scsi_transport_srp ib_ipoib rdma_ucm ib_ucm ib_uverbs ib_umad rdma_cm 
ib_cm iw_cm ocrdma ib_core intel_rapl x86_pkg_temp_thermal 
intel_powerclamp coretemp kvm_intel kvm irqbypass crct10dif_pclmul 
ipmi_ssif crc32_pclmul gpio_ich hpilo hpwdt iTCO_wdt iTCO_vendor_support 
ghash_clmulni_intel intel_cstate intel_uncore pcc_cpufreq 
intel_rapl_perf ie31200_edac sg shpchp acpi_power_meter acpi_cpufreq 
pcspkr ipmi_si ipmi_devintf edac_core ipmi_msghandler lpc_ich nfsd 
auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c sd_mod 
mgag200 i2c_algo_bit drm_kms_helper syscopyarea sysfillrect sysimgblt 
fb_sys_fops
[  233.526083]  ttm tg3 ahci libahci drm ptp libata crc32c_intel be2net 
i2c_core serio_raw pps_core dm_mirror dm_region_hash dm_log dm_mod
[  233.585357] CPU: 1 PID: 156 Comm: kworker/1:2 Tainted: G W       
4.11.0-rc4 #1
[  233.623342] Hardware name: HP ProLiant DL320e Gen8 v2, BIOS P80 
09/01/2013
[  233.658771] Workqueue: nvme_rdma_wq nvme_rdma_del_ctrl_work [nvme_rdma]
[  233.691614] Call Trace:
[  233.704155]  dump_stack+0x63/0x87
[  233.720829]  __warn+0xd1/0xf0
[  233.736661]  warn_slowpath_fmt+0x5f/0x80
[  233.756792]  ? ocrdma_post_recv+0x127/0x140 [ocrdma]
[  233.781768]  ? ocrdma_mbx_modify_qp+0x23b/0x370 [ocrdma]
[  233.806196]  __ib_drain_rq+0x15f/0x1b0 [ib_core]
[  233.827474]  ? ib_sg_to_pages+0x1a0/0x1a0 [ib_core]
[  233.849871]  ib_drain_rq+0x25/0x30 [ib_core]
[  233.869660]  ib_drain_qp+0x24/0x30 [ib_core]
[  233.889312]  nvme_rdma_stop_and_free_queue+0x27/0x40 [nvme_rdma]
[  233.918097]  nvme_rdma_destroy_admin_queue+0x60/0xb0 [nvme_rdma]
[  233.945818]  nvme_rdma_shutdown_ctrl+0xd4/0xe0 [nvme_rdma]
[  233.970971]  __nvme_rdma_remove_ctrl+0x8c/0x90 [nvme_rdma]
[  233.996145]  nvme_rdma_del_ctrl_work+0x1a/0x20 [nvme_rdma]
[  234.021216]  process_one_work+0x165/0x410
[  234.040244]  worker_thread+0x27f/0x4c0
[  234.057532]  kthread+0x101/0x140
[  234.072476]  ? rescuer_thread+0x3b0/0x3b0
[  234.090911]  ? kthread_park+0x90/0x90
[  234.107890]  ret_from_fork+0x2c/0x40
[  234.124422] ---[ end trace 2da8cf1943c3a510 ]---
[  234.491570] nvme nvme0: creating 8 I/O queues.
[  235.099065] nvme nvme0: new ctrl: NQN "testnqn", addr 172.31.45.4:4420

^ permalink raw reply	[flat|nested] 10+ messages in thread

end of thread, other threads:[~2017-03-27  3:18 UTC | newest]

Thread overview: 10+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
     [not found] <530228679.31424535.1488695516036.JavaMail.zimbra@redhat.com>
     [not found] ` <530228679.31424535.1488695516036.JavaMail.zimbra-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
2017-03-05  6:41   ` WARNING on drivers/infiniband/core/verbs.c:1975 __ib_drain_sq+0x182/0x1c0 [ib_core] Yi Zhang
2017-03-05  6:41     ` Yi Zhang
     [not found]     ` <1457714996.31424605.1488696068958.JavaMail.zimbra-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
2017-03-06 11:42       ` Sagi Grimberg
2017-03-06 11:42         ` Sagi Grimberg
     [not found]         ` <3ce3cfdb-a452-ff0c-fa52-2ad0546c8b8f-NQWnxTmZq1alnMjI0IkVqw@public.gmane.org>
2017-03-06 12:00           ` Devesh Sharma
2017-03-06 12:00             ` Devesh Sharma
2017-03-25  1:44       ` Doug Ledford
2017-03-25  1:44         ` Doug Ledford
     [not found]         ` <1490406266.2404.34.camel-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
2017-03-27  3:18           ` Yi Zhang
2017-03-27  3:18             ` Yi Zhang

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.