From: Yi Zhang <yi.zhang@redhat.com>
To: Parav Pandit <parav@mellanox.com>,
"linux-nvme@lists.infradead.org" <linux-nvme@lists.infradead.org>
Cc: Daniel Jurgens <danielj@mellanox.com>,
"linux-rdma@vger.kernel.org" <linux-rdma@vger.kernel.org>,
Devesh Sharma <devesh.sharma@broadcom.com>,
"selvin.xavier@broadcom.com" <selvin.xavier@broadcom.com>
Subject: Re: regression: nvme rdma with bnxt_re0 broken
Date: Fri, 12 Jul 2019 09:53:17 +0800 [thread overview]
Message-ID: <66d43fd8-18e8-8b9d-90e3-ee2804d56889@redhat.com> (raw)
In-Reply-To: <AM0PR05MB4866070FBADCCABD1F84E42ED1F30@AM0PR05MB4866.eurprd05.prod.outlook.com>
Hi Parav
Here is the info, let me know if it's enough, thanks.
[root@rdma-perf-07 ~]$ echo -n "module ib_core +p" >
/sys/kernel/debug/dynamic_debug/control
[root@rdma-perf-07 ~]$ ifdown bnxt_roce
Device 'bnxt_roce' successfully disconnected.
[root@rdma-perf-07 ~]$ ifup bnxt_roce
Connection successfully activated (D-Bus active path:
/org/freedesktop/NetworkManager/ActiveConnection/16)
[root@rdma-perf-07 ~]$ sh a.sh
DEV PORT INDEX GID IPv4 VER DEV
--- ---- ----- --- ------------ --- ---
bnxt_re0 1 0 fe80:0000:0000:0000:020a:f7ff:fee3:6e32
v1 bnxt_roce
bnxt_re0 1 1 fe80:0000:0000:0000:020a:f7ff:fee3:6e32
v2 bnxt_roce
bnxt_re0 1 10 0000:0000:0000:0000:0000:ffff:ac1f:2bbb
172.31.43.187 v1 bnxt_roce.43
bnxt_re0 1 11 0000:0000:0000:0000:0000:ffff:ac1f:2bbb
172.31.43.187 v2 bnxt_roce.43
bnxt_re0 1 2 fe80:0000:0000:0000:020a:f7ff:fee3:6e32
v1 bnxt_roce.45
bnxt_re0 1 3 fe80:0000:0000:0000:020a:f7ff:fee3:6e32
v2 bnxt_roce.45
bnxt_re0 1 4 fe80:0000:0000:0000:020a:f7ff:fee3:6e32
v1 bnxt_roce.43
bnxt_re0 1 5 fe80:0000:0000:0000:020a:f7ff:fee3:6e32
v2 bnxt_roce.43
bnxt_re0 1 6 0000:0000:0000:0000:0000:ffff:ac1f:28bb
172.31.40.187 v1 bnxt_roce
bnxt_re0 1 7 0000:0000:0000:0000:0000:ffff:ac1f:28bb
172.31.40.187 v2 bnxt_roce
bnxt_re0 1 8 0000:0000:0000:0000:0000:ffff:ac1f:2dbb
172.31.45.187 v1 bnxt_roce.45
bnxt_re0 1 9 0000:0000:0000:0000:0000:ffff:ac1f:2dbb
172.31.45.187 v2 bnxt_roce.45
bnxt_re1 1 0 fe80:0000:0000:0000:020a:f7ff:fee3:6e33
v1 lom_2
bnxt_re1 1 1 fe80:0000:0000:0000:020a:f7ff:fee3:6e33
v2 lom_2
cxgb4_0 1 0 0007:433b:f5b0:0000:0000:0000:0000:0000 v1
cxgb4_0 2 0 0007:433b:f5b8:0000:0000:0000:0000:0000 v1
hfi1_0 1 0 fe80:0000:0000:0000:0011:7501:0109:6c60 v1
hfi1_0 1 1 fe80:0000:0000:0000:0006:6a00:0000:0005 v1
mlx5_0 1 0 fe80:0000:0000:0000:506b:4b03:00f3:8a38 v1
n_gids_found=19
[root@rdma-perf-07 ~]$ dmesg | tail -15
[ 19.744421] IPv6: ADDRCONF(NETDEV_CHANGE): mlx5_ib0.8002: link
becomes ready
[ 19.758371] IPv6: ADDRCONF(NETDEV_CHANGE): mlx5_ib0.8004: link
becomes ready
[ 20.010469] hfi1 0000:d8:00.0: hfi1_0: Switching to NO_DMA_RTAIL
[ 20.440580] IPv6: ADDRCONF(NETDEV_CHANGE): mlx5_ib0.8006: link
becomes ready
[ 21.098510] bnxt_en 0000:19:00.0 bnxt_roce: Too many traffic classes
requested: 8. Max supported is 2.
[ 21.324341] bnxt_en 0000:19:00.0 bnxt_roce: Too many traffic classes
requested: 8. Max supported is 2.
[ 22.058647] IPv6: ADDRCONF(NETDEV_CHANGE): hfi1_opa0: link becomes ready
[ 211.407329] _ib_cache_gid_del: can't delete gid
fe80:0000:0000:0000:020a:f7ff:fee3:6e32 error=-22
[ 211.407334] _ib_cache_gid_del: can't delete gid
fe80:0000:0000:0000:020a:f7ff:fee3:6e32 error=-22
[ 211.425275] infiniband bnxt_re0: del_gid port=1 index=6 gid
0000:0000:0000:0000:0000:ffff:ac1f:28bb
[ 211.425280] infiniband bnxt_re0: free_gid_entry_locked port=1 index=6
gid 0000:0000:0000:0000:0000:ffff:ac1f:28bb
[ 211.425292] infiniband bnxt_re0: del_gid port=1 index=7 gid
0000:0000:0000:0000:0000:ffff:ac1f:28bb
[ 211.425461] infiniband bnxt_re0: free_gid_entry_locked port=1 index=7
gid 0000:0000:0000:0000:0000:ffff:ac1f:28bb
[ 225.474061] infiniband bnxt_re0: store_gid_entry port=1 index=6 gid
0000:0000:0000:0000:0000:ffff:ac1f:28bb
[ 225.474075] infiniband bnxt_re0: store_gid_entry port=1 index=7 gid
0000:0000:0000:0000:0000:ffff:ac1f:28bb
On 7/12/19 12:18 AM, Parav Pandit wrote:
> Sagi,
>
> This is better one to cc to linux-rdma.
>
> + Devesh, Selvin.
>
>> -----Original Message-----
>> From: Parav Pandit
>> Sent: Thursday, July 11, 2019 6:25 PM
>> To: Yi Zhang <yi.zhang@redhat.com>; linux-nvme@lists.infradead.org
>> Cc: Daniel Jurgens <danielj@mellanox.com>
>> Subject: RE: regression: nvme rdma with bnxt_re0 broken
>>
>> Hi Yi Zhang,
>>
>>> -----Original Message-----
>>> From: Yi Zhang <yi.zhang@redhat.com>
>>> Sent: Thursday, July 11, 2019 3:17 PM
>>> To: linux-nvme@lists.infradead.org
>>> Cc: Daniel Jurgens <danielj@mellanox.com>; Parav Pandit
>>> <parav@mellanox.com>
>>> Subject: regression: nvme rdma with bnxt_re0 broken
>>>
>>> Hello
>>>
>>> 'nvme connect' failed when use bnxt_re0 on latest upstream build[1],
>>> by bisecting I found it was introduced from v5.2.0-rc1 with [2], it
>>> works after I revert it.
>>> Let me know if you need more info, thanks.
>>>
>>> [1]
>>> [root@rdma-perf-07 ~]$ nvme connect -t rdma -a 172.31.40.125 -s 4420
>>> -n testnqn Failed to write to /dev/nvme-fabrics: Bad address
>>>
>>> [root@rdma-perf-07 ~]$ dmesg
>>> [ 476.320742] bnxt_en 0000:19:00.0: QPLIB: cmdq[0x4b9]=0x15 status
>>> 0x5 [ 476.327103] infiniband bnxt_re0: Failed to allocate HW AH [
>>> 476.332525] nvme nvme2: rdma_connect failed (-14).
>>> [ 476.343552] nvme nvme2: rdma connection establishment failed (-14)
>>>
>>> [root@rdma-perf-07 ~]$ lspci | grep -i Broadcom
>>> 01:00.0 Ethernet controller: Broadcom Inc. and subsidiaries NetXtreme
>>> BCM5720 2-port Gigabit Ethernet PCIe
>>> 01:00.1 Ethernet controller: Broadcom Inc. and subsidiaries NetXtreme
>>> BCM5720 2-port Gigabit Ethernet PCIe
>>> 18:00.0 RAID bus controller: Broadcom / LSI MegaRAID SAS-3 3008 [Fury]
>>> (rev
>>> 02)
>>> 19:00.0 Ethernet controller: Broadcom Inc. and subsidiaries BCM57412
>>> NetXtreme-E 10Gb RDMA Ethernet Controller (rev 01)
>>> 19:00.1 Ethernet controller: Broadcom Inc. and subsidiaries BCM57412
>>> NetXtreme-E 10Gb RDMA Ethernet Controller (rev 01)
>>>
>>>
>>> [2]
>>> commit 823b23da71132b80d9f41ab667c68b112455f3b6
>>> Author: Parav Pandit <parav@mellanox.com>
>>> Date: Wed Apr 10 11:23:03 2019 +0300
>>>
>>> IB/core: Allow vlan link local address based RoCE GIDs
>>>
>>> IPv6 link local address for a VLAN netdevice has nothing to do with its
>>> resemblance with the default GID, because VLAN link local GID is in
>>> different layer 2 domain.
>>>
>>> Now that RoCE MAD packet processing and route resolution consider the
>>> right GID index, there is no need for an unnecessary check which prevents
>>> the addition of vlan based IPv6 link local GIDs.
>>>
>>> Signed-off-by: Parav Pandit <parav@mellanox.com>
>>> Reviewed-by: Daniel Jurgens <danielj@mellanox.com>
>>> Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
>>> Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
>>>
>>>
>>>
>>> Best Regards,
>>> Yi Zhang
>>>
>> I need some more information from you to debug this issue as I don’t have the
>> hw.
>> The highlighted patch added support for IPv6 link local address for vlan. I am
>> unsure how this can affect IPv4 AH creation for which there is failure.
>>
>> 1. Before you assign the IP address to the netdevice, Please do, echo -n
>> "module ib_core +p" > /sys/kernel/debug/dynamic_debug/control
>>
>> Please share below output before doing nvme connect.
>> 2. Output of script [1]
>> $ show_gids script
>> If getting this script is problematic, share the output of,
>>
>> $ cat /sys/class/infiniband/bnxt_re0/ports/1/gids/*
>> $ cat /sys/class/infiniband/bnxt_re0/ports/1/gid_attrs/ndevs/*
>> $ ip link show
>> $ip addr show
>> $ dmesg
>>
>> [1] https://community.mellanox.com/s/article/understanding-show-gids-
>> script#jive_content_id_The_Script
>>
>> I suspect that driver's assumption about GID indices might have gone wrong
>> here in drivers/infiniband/hw/bnxt_re/ib_verbs.c.
>> Lets see about results to confirm that.
> _______________________________________________
> Linux-nvme mailing list
> Linux-nvme@lists.infradead.org
> http://lists.infradead.org/mailman/listinfo/linux-nvme
WARNING: multiple messages have this Message-ID (diff)
From: yi.zhang@redhat.com (Yi Zhang)
Subject: regression: nvme rdma with bnxt_re0 broken
Date: Fri, 12 Jul 2019 09:53:17 +0800 [thread overview]
Message-ID: <66d43fd8-18e8-8b9d-90e3-ee2804d56889@redhat.com> (raw)
In-Reply-To: <AM0PR05MB4866070FBADCCABD1F84E42ED1F30@AM0PR05MB4866.eurprd05.prod.outlook.com>
Hi Parav
Here is the info, let me know if it's enough, thanks.
[root at rdma-perf-07 ~]$ echo -n "module ib_core +p" >
/sys/kernel/debug/dynamic_debug/control
[root at rdma-perf-07 ~]$ ifdown bnxt_roce
Device 'bnxt_roce' successfully disconnected.
[root at rdma-perf-07 ~]$ ifup bnxt_roce
Connection successfully activated (D-Bus active path:
/org/freedesktop/NetworkManager/ActiveConnection/16)
[root at rdma-perf-07 ~]$ sh a.sh
DEV??? PORT??? INDEX??? GID??? ??? ??? ??? ??? IPv4 ??? ??? VER DEV
---??? ----??? -----??? ---??? ??? ??? ??? ??? ------------ ---??? ---
bnxt_re0??? 1??? 0??? fe80:0000:0000:0000:020a:f7ff:fee3:6e32 ??? ???
v1??? bnxt_roce
bnxt_re0??? 1??? 1??? fe80:0000:0000:0000:020a:f7ff:fee3:6e32 ??? ???
v2??? bnxt_roce
bnxt_re0??? 1??? 10??? 0000:0000:0000:0000:0000:ffff:ac1f:2bbb
172.31.43.187 ??? v1??? bnxt_roce.43
bnxt_re0??? 1??? 11??? 0000:0000:0000:0000:0000:ffff:ac1f:2bbb
172.31.43.187 ??? v2??? bnxt_roce.43
bnxt_re0??? 1??? 2??? fe80:0000:0000:0000:020a:f7ff:fee3:6e32 ??? ???
v1??? bnxt_roce.45
bnxt_re0??? 1??? 3??? fe80:0000:0000:0000:020a:f7ff:fee3:6e32 ??? ???
v2??? bnxt_roce.45
bnxt_re0??? 1??? 4??? fe80:0000:0000:0000:020a:f7ff:fee3:6e32 ??? ???
v1??? bnxt_roce.43
bnxt_re0??? 1??? 5??? fe80:0000:0000:0000:020a:f7ff:fee3:6e32 ??? ???
v2??? bnxt_roce.43
bnxt_re0??? 1??? 6??? 0000:0000:0000:0000:0000:ffff:ac1f:28bb
172.31.40.187 ??? v1??? bnxt_roce
bnxt_re0??? 1??? 7??? 0000:0000:0000:0000:0000:ffff:ac1f:28bb
172.31.40.187 ??? v2??? bnxt_roce
bnxt_re0??? 1??? 8??? 0000:0000:0000:0000:0000:ffff:ac1f:2dbb
172.31.45.187 ??? v1??? bnxt_roce.45
bnxt_re0??? 1??? 9??? 0000:0000:0000:0000:0000:ffff:ac1f:2dbb
172.31.45.187 ??? v2??? bnxt_roce.45
bnxt_re1??? 1??? 0??? fe80:0000:0000:0000:020a:f7ff:fee3:6e33 ??? ???
v1??? lom_2
bnxt_re1??? 1??? 1??? fe80:0000:0000:0000:020a:f7ff:fee3:6e33 ??? ???
v2??? lom_2
cxgb4_0??? 1??? 0??? 0007:433b:f5b0:0000:0000:0000:0000:0000 ??? ??? v1
cxgb4_0??? 2??? 0??? 0007:433b:f5b8:0000:0000:0000:0000:0000 ??? ??? v1
hfi1_0??? 1??? 0??? fe80:0000:0000:0000:0011:7501:0109:6c60 ??? v1
hfi1_0??? 1??? 1??? fe80:0000:0000:0000:0006:6a00:0000:0005 ??? v1
mlx5_0??? 1??? 0??? fe80:0000:0000:0000:506b:4b03:00f3:8a38 ??? v1
n_gids_found=19
[root at rdma-perf-07 ~]$ dmesg | tail -15
[?? 19.744421] IPv6: ADDRCONF(NETDEV_CHANGE): mlx5_ib0.8002: link
becomes ready
[?? 19.758371] IPv6: ADDRCONF(NETDEV_CHANGE): mlx5_ib0.8004: link
becomes ready
[?? 20.010469] hfi1 0000:d8:00.0: hfi1_0: Switching to NO_DMA_RTAIL
[?? 20.440580] IPv6: ADDRCONF(NETDEV_CHANGE): mlx5_ib0.8006: link
becomes ready
[?? 21.098510] bnxt_en 0000:19:00.0 bnxt_roce: Too many traffic classes
requested: 8. Max supported is 2.
[?? 21.324341] bnxt_en 0000:19:00.0 bnxt_roce: Too many traffic classes
requested: 8. Max supported is 2.
[?? 22.058647] IPv6: ADDRCONF(NETDEV_CHANGE): hfi1_opa0: link becomes ready
[? 211.407329] _ib_cache_gid_del: can't delete gid
fe80:0000:0000:0000:020a:f7ff:fee3:6e32 error=-22
[? 211.407334] _ib_cache_gid_del: can't delete gid
fe80:0000:0000:0000:020a:f7ff:fee3:6e32 error=-22
[? 211.425275] infiniband bnxt_re0: del_gid port=1 index=6 gid
0000:0000:0000:0000:0000:ffff:ac1f:28bb
[? 211.425280] infiniband bnxt_re0: free_gid_entry_locked port=1 index=6
gid 0000:0000:0000:0000:0000:ffff:ac1f:28bb
[? 211.425292] infiniband bnxt_re0: del_gid port=1 index=7 gid
0000:0000:0000:0000:0000:ffff:ac1f:28bb
[? 211.425461] infiniband bnxt_re0: free_gid_entry_locked port=1 index=7
gid 0000:0000:0000:0000:0000:ffff:ac1f:28bb
[? 225.474061] infiniband bnxt_re0: store_gid_entry port=1 index=6 gid
0000:0000:0000:0000:0000:ffff:ac1f:28bb
[? 225.474075] infiniband bnxt_re0: store_gid_entry port=1 index=7 gid
0000:0000:0000:0000:0000:ffff:ac1f:28bb
On 7/12/19 12:18 AM, Parav Pandit wrote:
> Sagi,
>
> This is better one to cc to linux-rdma.
>
> + Devesh, Selvin.
>
>> -----Original Message-----
>> From: Parav Pandit
>> Sent: Thursday, July 11, 2019 6:25 PM
>> To: Yi Zhang <yi.zhang at redhat.com>; linux-nvme at lists.infradead.org
>> Cc: Daniel Jurgens <danielj at mellanox.com>
>> Subject: RE: regression: nvme rdma with bnxt_re0 broken
>>
>> Hi Yi Zhang,
>>
>>> -----Original Message-----
>>> From: Yi Zhang <yi.zhang at redhat.com>
>>> Sent: Thursday, July 11, 2019 3:17 PM
>>> To: linux-nvme at lists.infradead.org
>>> Cc: Daniel Jurgens <danielj at mellanox.com>; Parav Pandit
>>> <parav at mellanox.com>
>>> Subject: regression: nvme rdma with bnxt_re0 broken
>>>
>>> Hello
>>>
>>> 'nvme connect' failed when use bnxt_re0 on latest upstream build[1],
>>> by bisecting I found it was introduced from v5.2.0-rc1 with [2], it
>>> works after I revert it.
>>> Let me know if you need more info, thanks.
>>>
>>> [1]
>>> [root at rdma-perf-07 ~]$ nvme connect -t rdma -a 172.31.40.125 -s 4420
>>> -n testnqn Failed to write to /dev/nvme-fabrics: Bad address
>>>
>>> [root at rdma-perf-07 ~]$ dmesg
>>> [ 476.320742] bnxt_en 0000:19:00.0: QPLIB: cmdq[0x4b9]=0x15 status
>>> 0x5 [ 476.327103] infiniband bnxt_re0: Failed to allocate HW AH [
>>> 476.332525] nvme nvme2: rdma_connect failed (-14).
>>> [ 476.343552] nvme nvme2: rdma connection establishment failed (-14)
>>>
>>> [root at rdma-perf-07 ~]$ lspci | grep -i Broadcom
>>> 01:00.0 Ethernet controller: Broadcom Inc. and subsidiaries NetXtreme
>>> BCM5720 2-port Gigabit Ethernet PCIe
>>> 01:00.1 Ethernet controller: Broadcom Inc. and subsidiaries NetXtreme
>>> BCM5720 2-port Gigabit Ethernet PCIe
>>> 18:00.0 RAID bus controller: Broadcom / LSI MegaRAID SAS-3 3008 [Fury]
>>> (rev
>>> 02)
>>> 19:00.0 Ethernet controller: Broadcom Inc. and subsidiaries BCM57412
>>> NetXtreme-E 10Gb RDMA Ethernet Controller (rev 01)
>>> 19:00.1 Ethernet controller: Broadcom Inc. and subsidiaries BCM57412
>>> NetXtreme-E 10Gb RDMA Ethernet Controller (rev 01)
>>>
>>>
>>> [2]
>>> commit 823b23da71132b80d9f41ab667c68b112455f3b6
>>> Author: Parav Pandit <parav at mellanox.com>
>>> Date: Wed Apr 10 11:23:03 2019 +0300
>>>
>>> IB/core: Allow vlan link local address based RoCE GIDs
>>>
>>> IPv6 link local address for a VLAN netdevice has nothing to do with its
>>> resemblance with the default GID, because VLAN link local GID is in
>>> different layer 2 domain.
>>>
>>> Now that RoCE MAD packet processing and route resolution consider the
>>> right GID index, there is no need for an unnecessary check which prevents
>>> the addition of vlan based IPv6 link local GIDs.
>>>
>>> Signed-off-by: Parav Pandit <parav at mellanox.com>
>>> Reviewed-by: Daniel Jurgens <danielj at mellanox.com>
>>> Signed-off-by: Leon Romanovsky <leonro at mellanox.com>
>>> Signed-off-by: Jason Gunthorpe <jgg at mellanox.com>
>>>
>>>
>>>
>>> Best Regards,
>>> Yi Zhang
>>>
>> I need some more information from you to debug this issue as I don?t have the
>> hw.
>> The highlighted patch added support for IPv6 link local address for vlan. I am
>> unsure how this can affect IPv4 AH creation for which there is failure.
>>
>> 1. Before you assign the IP address to the netdevice, Please do, echo -n
>> "module ib_core +p" > /sys/kernel/debug/dynamic_debug/control
>>
>> Please share below output before doing nvme connect.
>> 2. Output of script [1]
>> $ show_gids script
>> If getting this script is problematic, share the output of,
>>
>> $ cat /sys/class/infiniband/bnxt_re0/ports/1/gids/*
>> $ cat /sys/class/infiniband/bnxt_re0/ports/1/gid_attrs/ndevs/*
>> $ ip link show
>> $ip addr show
>> $ dmesg
>>
>> [1] https://community.mellanox.com/s/article/understanding-show-gids-
>> script#jive_content_id_The_Script
>>
>> I suspect that driver's assumption about GID indices might have gone wrong
>> here in drivers/infiniband/hw/bnxt_re/ib_verbs.c.
>> Lets see about results to confirm that.
> _______________________________________________
> Linux-nvme mailing list
> Linux-nvme at lists.infradead.org
> http://lists.infradead.org/mailman/listinfo/linux-nvme
next prev parent reply other threads:[~2019-07-12 1:53 UTC|newest]
Thread overview: 38+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <1310083272.27124086.1562836112586.JavaMail.zimbra@redhat.com>
2019-07-11 9:47 ` regression: nvme rdma with bnxt_re0 broken Yi Zhang
2019-07-11 12:54 ` Parav Pandit
2019-07-11 16:18 ` Parav Pandit
2019-07-11 16:18 ` Parav Pandit
2019-07-12 1:53 ` Yi Zhang [this message]
2019-07-12 1:53 ` Yi Zhang
2019-07-12 2:49 ` Parav Pandit
2019-07-12 2:49 ` Parav Pandit
2019-07-12 3:45 ` Selvin Xavier
2019-07-12 3:45 ` Selvin Xavier
2019-07-12 9:28 ` Parav Pandit
2019-07-12 9:28 ` Parav Pandit
2019-07-12 9:39 ` Parav Pandit
2019-07-12 9:39 ` Parav Pandit
2019-07-12 9:49 ` Parav Pandit
2019-07-12 9:49 ` Parav Pandit
2019-07-12 11:41 ` Yi Zhang
2019-07-12 11:41 ` Yi Zhang
2019-07-12 12:52 ` Parav Pandit
2019-07-12 12:52 ` Parav Pandit
2019-07-12 15:40 ` Jason Gunthorpe
2019-07-12 15:40 ` Jason Gunthorpe
2019-07-12 16:29 ` Selvin Xavier
2019-07-12 16:29 ` Selvin Xavier
2019-07-12 17:42 ` Jason Gunthorpe
2019-07-12 17:42 ` Jason Gunthorpe
2019-07-13 7:51 ` Selvin Xavier
2019-07-13 7:51 ` Selvin Xavier
2019-07-13 12:12 ` Jason Gunthorpe
2019-07-13 12:12 ` Jason Gunthorpe
2019-07-12 16:18 ` Selvin Xavier
2019-07-12 16:18 ` Selvin Xavier
2019-07-13 7:56 ` Yi Zhang
2019-07-13 7:56 ` Yi Zhang
2019-07-13 16:00 ` Selvin Xavier
2019-07-13 16:00 ` Selvin Xavier
2019-07-11 16:13 ` Sagi Grimberg
2019-07-11 16:13 ` Sagi Grimberg
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=66d43fd8-18e8-8b9d-90e3-ee2804d56889@redhat.com \
--to=yi.zhang@redhat.com \
--cc=danielj@mellanox.com \
--cc=devesh.sharma@broadcom.com \
--cc=linux-nvme@lists.infradead.org \
--cc=linux-rdma@vger.kernel.org \
--cc=parav@mellanox.com \
--cc=selvin.xavier@broadcom.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.