Linux-RDMA Archive on lore.kernel.org
 help / color / Atom feed
* [PATCH rdma-next 0/3] Optimize SGL registration
@ 2019-10-06 15:59 Leon Romanovsky
  2019-10-06 15:59 ` [PATCH mlx5-next 1/3] net/mlx5: Expose optimal performance scatter entries capability Leon Romanovsky
                   ` (2 more replies)
  0 siblings, 3 replies; 9+ messages in thread
From: Leon Romanovsky @ 2019-10-06 15:59 UTC (permalink / raw)
  To: Doug Ledford, Jason Gunthorpe
  Cc: Leon Romanovsky, RDMA mailing list, Or Gerlitz, Yamin Friedman,
	Saeed Mahameed, linux-netdev

From: Leon Romanovsky <leonro@mellanox.com>

Hi,

This series from Yamin implements long standing "TODO" existed in rw.c.

Thanks

Yamin Friedman (3):
  net/mlx5: Expose optimal performance scatter entries capability
  RDMA/mlx5: Add capability for max sge to get optimized performance
  RDMA/rw: Support threshold for registration vs scattering to local
    pages

 drivers/infiniband/core/rw.c      | 12 ++++++------
 drivers/infiniband/hw/mlx5/main.c |  2 ++
 include/linux/mlx5/mlx5_ifc.h     |  2 +-
 include/rdma/ib_verbs.h           |  2 ++
 4 files changed, 11 insertions(+), 7 deletions(-)

--
2.20.1


^ permalink raw reply	[flat|nested] 9+ messages in thread

* [PATCH mlx5-next 1/3] net/mlx5: Expose optimal performance scatter entries capability
  2019-10-06 15:59 [PATCH rdma-next 0/3] Optimize SGL registration Leon Romanovsky
@ 2019-10-06 15:59 ` Leon Romanovsky
  2019-10-06 15:59 ` [PATCH rdma-next 2/3] RDMA/mlx5: Add capability for max sge to get optimized performance Leon Romanovsky
  2019-10-06 15:59 ` [PATCH rdma-next 3/3] RDMA/rw: Support threshold for registration vs scattering to local pages Leon Romanovsky
  2 siblings, 0 replies; 9+ messages in thread
From: Leon Romanovsky @ 2019-10-06 15:59 UTC (permalink / raw)
  To: Doug Ledford, Jason Gunthorpe
  Cc: Leon Romanovsky, RDMA mailing list, Or Gerlitz, Yamin Friedman,
	Saeed Mahameed, linux-netdev

From: Yamin Friedman <yaminf@mellanox.com>

Expose maximum scatter entries per RDMA READ for optimal performance.

Signed-off-by: Yamin Friedman <yaminf@mellanox.com>
Reviewed-by: Or Gerlitz <ogerlitz@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
---
 include/linux/mlx5/mlx5_ifc.h | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/include/linux/mlx5/mlx5_ifc.h b/include/linux/mlx5/mlx5_ifc.h
index 138c50d5a353..c0bfb1d90dd2 100644
--- a/include/linux/mlx5/mlx5_ifc.h
+++ b/include/linux/mlx5/mlx5_ifc.h
@@ -1153,7 +1153,7 @@ struct mlx5_ifc_cmd_hca_cap_bits {
 	u8         log_max_srq[0x5];
 	u8         reserved_at_b0[0x10];
 
-	u8         reserved_at_c0[0x8];
+	u8         max_sgl_for_optimized_performance[0x8];
 	u8         log_max_cq_sz[0x8];
 	u8         reserved_at_d0[0xb];
 	u8         log_max_cq[0x5];
-- 
2.20.1


^ permalink raw reply	[flat|nested] 9+ messages in thread

* [PATCH rdma-next 2/3] RDMA/mlx5: Add capability for max sge to get optimized performance
  2019-10-06 15:59 [PATCH rdma-next 0/3] Optimize SGL registration Leon Romanovsky
  2019-10-06 15:59 ` [PATCH mlx5-next 1/3] net/mlx5: Expose optimal performance scatter entries capability Leon Romanovsky
@ 2019-10-06 15:59 ` Leon Romanovsky
  2019-10-07  7:13   ` Christoph Hellwig
  2019-10-06 15:59 ` [PATCH rdma-next 3/3] RDMA/rw: Support threshold for registration vs scattering to local pages Leon Romanovsky
  2 siblings, 1 reply; 9+ messages in thread
From: Leon Romanovsky @ 2019-10-06 15:59 UTC (permalink / raw)
  To: Doug Ledford, Jason Gunthorpe
  Cc: Leon Romanovsky, RDMA mailing list, Or Gerlitz, Yamin Friedman,
	Saeed Mahameed, linux-netdev

From: Yamin Friedman <yaminf@mellanox.com>

Allows the IB device to provide a value of maximum scatter gather entries
per RDMA READ.

In certain cases it may be preferable for a device to perform UMR memory
registration rather than have many scatter entries in a single RDMA READ.
This provides a significant performance increase in devices capable of
using different memory registration schemes based on the number of scatter
gather entries. This general capability allows each device vendor to fine
tune when it is better to use memory registration.

Signed-off-by: Yamin Friedman <yaminf@mellanox.com>
Reviewed-by: Or Gerlitz <ogerlitz@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
---
 drivers/infiniband/hw/mlx5/main.c | 2 ++
 include/rdma/ib_verbs.h           | 2 ++
 2 files changed, 4 insertions(+)

diff --git a/drivers/infiniband/hw/mlx5/main.c b/drivers/infiniband/hw/mlx5/main.c
index fa23c8e7043b..39d54e285ae9 100644
--- a/drivers/infiniband/hw/mlx5/main.c
+++ b/drivers/infiniband/hw/mlx5/main.c
@@ -1012,6 +1012,8 @@ static int mlx5_ib_query_device(struct ib_device *ibdev,
 		1 << MLX5_CAP_GEN(mdev, log_max_klm_list_size);
 	props->max_pi_fast_reg_page_list_len =
 		props->max_fast_reg_page_list_len / 2;
+	props->max_sgl_rd =
+		MLX5_CAP_GEN(mdev, max_sgl_for_optimized_performance);
 	get_atomic_caps_qp(dev, props);
 	props->masked_atomic_cap   = IB_ATOMIC_NONE;
 	props->max_mcast_grp	   = 1 << MLX5_CAP_GEN(mdev, log_max_mcg);
diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h
index 4f671378dbfc..60fd98a9b7e8 100644
--- a/include/rdma/ib_verbs.h
+++ b/include/rdma/ib_verbs.h
@@ -445,6 +445,8 @@ struct ib_device_attr {
 	struct ib_tm_caps	tm_caps;
 	struct ib_cq_caps       cq_caps;
 	u64			max_dm_size;
+	/* Max entries for sgl for optimized performance per READ */
+	u32			max_sgl_rd;
 };
 
 enum ib_mtu {
-- 
2.20.1


^ permalink raw reply	[flat|nested] 9+ messages in thread

* [PATCH rdma-next 3/3] RDMA/rw: Support threshold for registration vs scattering to local pages
  2019-10-06 15:59 [PATCH rdma-next 0/3] Optimize SGL registration Leon Romanovsky
  2019-10-06 15:59 ` [PATCH mlx5-next 1/3] net/mlx5: Expose optimal performance scatter entries capability Leon Romanovsky
  2019-10-06 15:59 ` [PATCH rdma-next 2/3] RDMA/mlx5: Add capability for max sge to get optimized performance Leon Romanovsky
@ 2019-10-06 15:59 ` Leon Romanovsky
  2019-10-07  6:58   ` Christoph Hellwig
  2 siblings, 1 reply; 9+ messages in thread
From: Leon Romanovsky @ 2019-10-06 15:59 UTC (permalink / raw)
  To: Doug Ledford, Jason Gunthorpe
  Cc: Leon Romanovsky, RDMA mailing list, Or Gerlitz, Yamin Friedman,
	Saeed Mahameed, linux-netdev

From: Yamin Friedman <yaminf@mellanox.com>

If there are more scatter entries than the recommended limit provided by
the ib device, UMR registration is used. This will provide optimal
performance when performing large RDMA READs over devices that advertise
the threshold capability.

With ConnectX-5 running NVMeoF RDMA with FIO single QP 128KB writes:
Without use of cap: 70Gb/sec
With use of cap: 84Gb/sec

Signed-off-by: Yamin Friedman <yaminf@mellanox.com>
Reviewed-by: Or Gerlitz <ogerlitz@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
---
 drivers/infiniband/core/rw.c | 12 ++++++------
 1 file changed, 6 insertions(+), 6 deletions(-)

diff --git a/drivers/infiniband/core/rw.c b/drivers/infiniband/core/rw.c
index 5337393d4dfe..ecff40efcb88 100644
--- a/drivers/infiniband/core/rw.c
+++ b/drivers/infiniband/core/rw.c
@@ -20,9 +20,7 @@ module_param_named(force_mr, rdma_rw_force_mr, bool, 0);
 MODULE_PARM_DESC(force_mr, "Force usage of MRs for RDMA READ/WRITE operations");
 
 /*
- * Check if the device might use memory registration.  This is currently only
- * true for iWarp devices. In the future we can hopefully fine tune this based
- * on HCA driver input.
+ * Check if the device might use memory registration.
  */
 static inline bool rdma_rw_can_use_mr(struct ib_device *dev, u8 port_num)
 {
@@ -30,6 +28,8 @@ static inline bool rdma_rw_can_use_mr(struct ib_device *dev, u8 port_num)
 		return true;
 	if (unlikely(rdma_rw_force_mr))
 		return true;
+	if (dev->attrs.max_sgl_rd)
+		return true;
 	return false;
 }
 
@@ -37,9 +37,6 @@ static inline bool rdma_rw_can_use_mr(struct ib_device *dev, u8 port_num)
  * Check if the device will use memory registration for this RW operation.
  * We currently always use memory registrations for iWarp RDMA READs, and
  * have a debug option to force usage of MRs.
- *
- * XXX: In the future we can hopefully fine tune this based on HCA driver
- * input.
  */
 static inline bool rdma_rw_io_needs_mr(struct ib_device *dev, u8 port_num,
 		enum dma_data_direction dir, int dma_nents)
@@ -48,6 +45,9 @@ static inline bool rdma_rw_io_needs_mr(struct ib_device *dev, u8 port_num,
 		return true;
 	if (unlikely(rdma_rw_force_mr))
 		return true;
+	if (dev->attrs.max_sgl_rd && dir == DMA_FROM_DEVICE
+	    && dma_nents > dev->attrs.max_sgl_rd)
+		return true;
 	return false;
 }
 
-- 
2.20.1


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH rdma-next 3/3] RDMA/rw: Support threshold for registration vs scattering to local pages
  2019-10-06 15:59 ` [PATCH rdma-next 3/3] RDMA/rw: Support threshold for registration vs scattering to local pages Leon Romanovsky
@ 2019-10-07  6:58   ` Christoph Hellwig
  2019-10-07  7:54     ` Leon Romanovsky
  0 siblings, 1 reply; 9+ messages in thread
From: Christoph Hellwig @ 2019-10-07  6:58 UTC (permalink / raw)
  To: Leon Romanovsky
  Cc: Doug Ledford, Jason Gunthorpe, Leon Romanovsky,
	RDMA mailing list, Or Gerlitz, Yamin Friedman, Saeed Mahameed,
	linux-netdev

>  /*
> - * Check if the device might use memory registration.  This is currently only
> - * true for iWarp devices. In the future we can hopefully fine tune this based
> - * on HCA driver input.
> + * Check if the device might use memory registration.
>   */

Please keep the important bits of this comments instead of just
removing them.

>  {
> @@ -30,6 +28,8 @@ static inline bool rdma_rw_can_use_mr(struct ib_device *dev, u8 port_num)
>  		return true;
>  	if (unlikely(rdma_rw_force_mr))
>  		return true;
> +	if (dev->attrs.max_sgl_rd)
> +		return true;

Logically this should go before the rdma_rw_force_mr check.

>  	if (unlikely(rdma_rw_force_mr))
>  		return true;
> +	if (dev->attrs.max_sgl_rd && dir == DMA_FROM_DEVICE
> +	    && dma_nents > dev->attrs.max_sgl_rd)

Wrong indendation.  The && belongs on the first line.  And again, this
logically belongs before the rdma_rw_force_mr check.

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH rdma-next 2/3] RDMA/mlx5: Add capability for max sge to get optimized performance
  2019-10-06 15:59 ` [PATCH rdma-next 2/3] RDMA/mlx5: Add capability for max sge to get optimized performance Leon Romanovsky
@ 2019-10-07  7:13   ` Christoph Hellwig
  2019-10-07  7:54     ` Leon Romanovsky
  0 siblings, 1 reply; 9+ messages in thread
From: Christoph Hellwig @ 2019-10-07  7:13 UTC (permalink / raw)
  To: Leon Romanovsky
  Cc: Doug Ledford, Jason Gunthorpe, Leon Romanovsky,
	RDMA mailing list, Or Gerlitz, Yamin Friedman, Saeed Mahameed,
	linux-netdev

On Sun, Oct 06, 2019 at 06:59:54PM +0300, Leon Romanovsky wrote:
> diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h
> index 4f671378dbfc..60fd98a9b7e8 100644
> --- a/include/rdma/ib_verbs.h
> +++ b/include/rdma/ib_verbs.h
> @@ -445,6 +445,8 @@ struct ib_device_attr {
>  	struct ib_tm_caps	tm_caps;
>  	struct ib_cq_caps       cq_caps;
>  	u64			max_dm_size;
> +	/* Max entries for sgl for optimized performance per READ */
> +	u32			max_sgl_rd;

This either needs to go into what current is patch 3 or an entirely
separate one.

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH rdma-next 2/3] RDMA/mlx5: Add capability for max sge to get optimized performance
  2019-10-07  7:13   ` Christoph Hellwig
@ 2019-10-07  7:54     ` Leon Romanovsky
  0 siblings, 0 replies; 9+ messages in thread
From: Leon Romanovsky @ 2019-10-07  7:54 UTC (permalink / raw)
  To: Christoph Hellwig
  Cc: Doug Ledford, Jason Gunthorpe, RDMA mailing list, Or Gerlitz,
	Yamin Friedman, Saeed Mahameed, linux-netdev

On Mon, Oct 07, 2019 at 12:13:02AM -0700, Christoph Hellwig wrote:
> On Sun, Oct 06, 2019 at 06:59:54PM +0300, Leon Romanovsky wrote:
> > diff --git a/include/rdma/ib_verbs.h b/include/rdma/ib_verbs.h
> > index 4f671378dbfc..60fd98a9b7e8 100644
> > --- a/include/rdma/ib_verbs.h
> > +++ b/include/rdma/ib_verbs.h
> > @@ -445,6 +445,8 @@ struct ib_device_attr {
> >  	struct ib_tm_caps	tm_caps;
> >  	struct ib_cq_caps       cq_caps;
> >  	u64			max_dm_size;
> > +	/* Max entries for sgl for optimized performance per READ */
> > +	u32			max_sgl_rd;
>
> This either needs to go into what current is patch 3 or an entirely
> separate one.

I'll reshuffle.

Thanks

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH rdma-next 3/3] RDMA/rw: Support threshold for registration vs scattering to local pages
  2019-10-07  6:58   ` Christoph Hellwig
@ 2019-10-07  7:54     ` Leon Romanovsky
  2019-10-07 10:07       ` Max Gurtovoy
  0 siblings, 1 reply; 9+ messages in thread
From: Leon Romanovsky @ 2019-10-07  7:54 UTC (permalink / raw)
  To: Christoph Hellwig
  Cc: Doug Ledford, Jason Gunthorpe, RDMA mailing list, Or Gerlitz,
	Yamin Friedman, Saeed Mahameed, linux-netdev

On Sun, Oct 06, 2019 at 11:58:25PM -0700, Christoph Hellwig wrote:
> >  /*
> > - * Check if the device might use memory registration.  This is currently only
> > - * true for iWarp devices. In the future we can hopefully fine tune this based
> > - * on HCA driver input.
> > + * Check if the device might use memory registration.
> >   */
>
> Please keep the important bits of this comments instead of just
> removing them.
>
> >  {
> > @@ -30,6 +28,8 @@ static inline bool rdma_rw_can_use_mr(struct ib_device *dev, u8 port_num)
> >  		return true;
> >  	if (unlikely(rdma_rw_force_mr))
> >  		return true;
> > +	if (dev->attrs.max_sgl_rd)
> > +		return true;
>
> Logically this should go before the rdma_rw_force_mr check.
>
> >  	if (unlikely(rdma_rw_force_mr))
> >  		return true;
> > +	if (dev->attrs.max_sgl_rd && dir == DMA_FROM_DEVICE
> > +	    && dma_nents > dev->attrs.max_sgl_rd)
>
> Wrong indendation.  The && belongs on the first line.  And again, this
> logically belongs before the rdma_rw_force_mr check.

I'll fix.

Thanks

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH rdma-next 3/3] RDMA/rw: Support threshold for registration vs scattering to local pages
  2019-10-07  7:54     ` Leon Romanovsky
@ 2019-10-07 10:07       ` Max Gurtovoy
  0 siblings, 0 replies; 9+ messages in thread
From: Max Gurtovoy @ 2019-10-07 10:07 UTC (permalink / raw)
  To: Leon Romanovsky, Christoph Hellwig
  Cc: Doug Ledford, Jason Gunthorpe, RDMA mailing list, Or Gerlitz,
	Yamin Friedman, Saeed Mahameed, linux-netdev


On 10/7/2019 10:54 AM, Leon Romanovsky wrote:
> On Sun, Oct 06, 2019 at 11:58:25PM -0700, Christoph Hellwig wrote:
>>>   /*
>>> - * Check if the device might use memory registration.  This is currently only
>>> - * true for iWarp devices. In the future we can hopefully fine tune this based
>>> - * on HCA driver input.
>>> + * Check if the device might use memory registration.
>>>    */
>> Please keep the important bits of this comments instead of just
>> removing them.
>>
>>>   {
>>> @@ -30,6 +28,8 @@ static inline bool rdma_rw_can_use_mr(struct ib_device *dev, u8 port_num)
>>>   		return true;
>>>   	if (unlikely(rdma_rw_force_mr))
>>>   		return true;
>>> +	if (dev->attrs.max_sgl_rd)
>>> +		return true;
>> Logically this should go before the rdma_rw_force_mr check.
>>
>>>   	if (unlikely(rdma_rw_force_mr))
>>>   		return true;
>>> +	if (dev->attrs.max_sgl_rd && dir == DMA_FROM_DEVICE
>>> +	    && dma_nents > dev->attrs.max_sgl_rd)
>> Wrong indendation.  The && belongs on the first line.  And again, this
>> logically belongs before the rdma_rw_force_mr check.
> I'll fix.
>
> Thanks

The above comments looks reasonable.

Nice optimization Yamin.


^ permalink raw reply	[flat|nested] 9+ messages in thread

end of thread, back to index

Thread overview: 9+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-10-06 15:59 [PATCH rdma-next 0/3] Optimize SGL registration Leon Romanovsky
2019-10-06 15:59 ` [PATCH mlx5-next 1/3] net/mlx5: Expose optimal performance scatter entries capability Leon Romanovsky
2019-10-06 15:59 ` [PATCH rdma-next 2/3] RDMA/mlx5: Add capability for max sge to get optimized performance Leon Romanovsky
2019-10-07  7:13   ` Christoph Hellwig
2019-10-07  7:54     ` Leon Romanovsky
2019-10-06 15:59 ` [PATCH rdma-next 3/3] RDMA/rw: Support threshold for registration vs scattering to local pages Leon Romanovsky
2019-10-07  6:58   ` Christoph Hellwig
2019-10-07  7:54     ` Leon Romanovsky
2019-10-07 10:07       ` Max Gurtovoy

Linux-RDMA Archive on lore.kernel.org

Archives are clonable:
	git clone --mirror https://lore.kernel.org/linux-rdma/0 linux-rdma/git/0.git

	# If you have public-inbox 1.1+ installed, you may
	# initialize and index your mirror using the following commands:
	public-inbox-init -V2 linux-rdma linux-rdma/ https://lore.kernel.org/linux-rdma \
		linux-rdma@vger.kernel.org linux-rdma@archiver.kernel.org
	public-inbox-index linux-rdma

Example config snippet for mirrors

Newsgroup available over NNTP:
	nntp://nntp.lore.kernel.org/org.kernel.vger.linux-rdma


AGPL code for this site: git clone https://public-inbox.org/ public-inbox