From: Yishai Hadas <yishaih@nvidia.com>
To: <linux-rdma@vger.kernel.org>
Cc: <jgg@nvidia.com>, <yishaih@nvidia.com>, <maorg@nvidia.com>,
<markzhang@nvidia.com>, <edwards@nvidia.com>,
Ido Kalir <idok@nvidia.com>
Subject: [PATCH rdma-core 18/27] pyverbs: Support DevX UMEM registration
Date: Tue, 20 Jul 2021 11:16:38 +0300 [thread overview]
Message-ID: <20210720081647.1980-19-yishaih@nvidia.com> (raw)
In-Reply-To: <20210720081647.1980-1-yishaih@nvidia.com>
From: Edward Srouji <edwards@nvidia.com>
Add support to register a DevX UMEM and UMEM extended, a user memory to
be used by the DevX interface for DMA.
Reviewed-by: Ido Kalir <idok@nvidia.com>
Signed-off-by: Edward Srouji <edwards@nvidia.com>
---
pyverbs/providers/mlx5/libmlx5.pxd | 21 +++++++--
pyverbs/providers/mlx5/mlx5dv.pxd | 8 ++++
pyverbs/providers/mlx5/mlx5dv.pyx | 93 ++++++++++++++++++++++++++++++++++++--
3 files changed, 116 insertions(+), 6 deletions(-)
diff --git a/pyverbs/providers/mlx5/libmlx5.pxd b/pyverbs/providers/mlx5/libmlx5.pxd
index 8dd0141..66b8705 100644
--- a/pyverbs/providers/mlx5/libmlx5.pxd
+++ b/pyverbs/providers/mlx5/libmlx5.pxd
@@ -212,6 +212,17 @@ cdef extern from 'infiniband/mlx5dv.h':
uint8_t fm_ce_se
uint32_t imm
+ cdef struct mlx5dv_devx_umem:
+ uint32_t umem_id;
+
+ cdef struct mlx5dv_devx_umem_in:
+ void *addr
+ size_t size
+ uint32_t access
+ uint64_t pgsz_bitmap
+ uint64_t comp_mask
+
+
void mlx5dv_set_ctrl_seg(mlx5_wqe_ctrl_seg *seg, uint16_t pi, uint8_t opcode,
uint8_t opmod, uint32_t qp_num, uint8_t fm_ce_se,
uint8_t ds, uint8_t signature, uint32_t imm)
@@ -298,11 +309,15 @@ cdef extern from 'infiniband/mlx5dv.h':
int mlx5dv_map_ah_to_qp(v.ibv_ah *ah, uint32_t qp_num)
# DevX APIs
- mlx5dv_devx_uar *mlx5dv_devx_alloc_uar(v.ibv_context *context,
- uint32_t flags)
+ mlx5dv_devx_uar *mlx5dv_devx_alloc_uar(v.ibv_context *context, uint32_t flags)
void mlx5dv_devx_free_uar(mlx5dv_devx_uar *devx_uar)
int mlx5dv_devx_general_cmd(v.ibv_context *context, const void *in_,
- size_t inlen, void *out, size_t outlen);
+ size_t inlen, void *out, size_t outlen)
+ mlx5dv_devx_umem *mlx5dv_devx_umem_reg(v.ibv_context *ctx, void *addr,
+ size_t size, unsigned long access)
+ mlx5dv_devx_umem *mlx5dv_devx_umem_reg_ex(v.ibv_context *ctx,
+ mlx5dv_devx_umem_in *umem_in)
+ int mlx5dv_devx_umem_dereg(mlx5dv_devx_umem *umem)
# Mkey setters
void mlx5dv_wr_mkey_configure(mlx5dv_qp_ex *mqp, mlx5dv_mkey *mkey,
diff --git a/pyverbs/providers/mlx5/mlx5dv.pxd b/pyverbs/providers/mlx5/mlx5dv.pxd
index 18b208e..154a117 100644
--- a/pyverbs/providers/mlx5/mlx5dv.pxd
+++ b/pyverbs/providers/mlx5/mlx5dv.pxd
@@ -11,6 +11,8 @@ from pyverbs.cq cimport CQEX
cdef class Mlx5Context(Context):
+ cdef object devx_umems
+ cdef add_ref(self, obj)
cpdef close(self)
cdef class Mlx5DVContextAttr(PyverbsObject):
@@ -69,3 +71,9 @@ cdef class Wqe(PyverbsCM):
cdef void *addr
cdef int is_user_addr
cdef object segments
+
+cdef class Mlx5UMEM(PyverbsCM):
+ cdef dv.mlx5dv_devx_umem *umem
+ cdef Context context
+ cdef void *addr
+ cdef object is_user_addr
diff --git a/pyverbs/providers/mlx5/mlx5dv.pyx b/pyverbs/providers/mlx5/mlx5dv.pyx
index 07dd7db..d16aed1 100644
--- a/pyverbs/providers/mlx5/mlx5dv.pyx
+++ b/pyverbs/providers/mlx5/mlx5dv.pyx
@@ -2,10 +2,11 @@
# Copyright (c) 2019 Mellanox Technologies, Inc. All rights reserved. See COPYING file
from libc.stdint cimport uintptr_t, uint8_t, uint16_t, uint32_t
-from libc.stdlib cimport calloc, free, malloc
-from libc.string cimport memcpy
+from libc.string cimport memcpy, memset
+from libc.stdlib cimport calloc, free
from posix.mman cimport munmap
import logging
+import weakref
from pyverbs.providers.mlx5.mlx5dv_mkey cimport Mlx5MrInterleaved, Mlx5Mkey, \
Mlx5MkeyConfAttr, Mlx5SigBlockAttr
@@ -13,6 +14,7 @@ from pyverbs.pyverbs_error import PyverbsUserError, PyverbsRDMAError, PyverbsErr
from pyverbs.providers.mlx5.mlx5dv_sched cimport Mlx5dvSchedLeaf
cimport pyverbs.providers.mlx5.mlx5dv_enums as dve
cimport pyverbs.providers.mlx5.libmlx5 as dv
+from pyverbs.mem_alloc import posix_memalign
from pyverbs.qp cimport QPInitAttrEx, QPEx
from pyverbs.base import PyverbsRDMAErrno
from pyverbs.base cimport close_weakrefs
@@ -156,6 +158,7 @@ cdef class Mlx5Context(Context):
if self.context == NULL:
raise PyverbsRDMAErrno('Failed to open mlx5 context on {dev}'
.format(dev=self.name))
+ self.devx_umems = weakref.WeakSet()
def query_mlx5_device(self, comp_mask=-1):
"""
@@ -259,12 +262,21 @@ cdef class Mlx5Context(Context):
free(clock_info)
return ns_time
+ cdef add_ref(self, obj):
+ try:
+ Context.add_ref(self, obj)
+ except PyverbsError:
+ if isinstance(obj, Mlx5UMEM):
+ self.devx_umems.add(obj)
+ else:
+ raise PyverbsError('Unrecognized object type')
+
def __dealloc__(self):
self.close()
cpdef close(self):
if self.context != NULL:
- close_weakrefs([self.pps])
+ close_weakrefs([self.pps, self.devx_umems])
super(Mlx5Context, self).close()
@@ -1307,3 +1319,78 @@ cdef class Wqe(PyverbsCM):
if not self.is_user_addr:
free(self.addr)
self.addr = NULL
+
+
+cdef class Mlx5UMEM(PyverbsCM):
+ def __init__(self, Context context not None, size, addr=None, alignment=64,
+ access=0, pgsz_bitmap=0, comp_mask=0):
+ """
+ User memory object to be used by the DevX interface.
+ If pgsz_bitmap or comp_mask were passed, the extended umem registration
+ will be used.
+ :param context: RDMA device context to create the action on
+ :param size: The size of the addr buffer (or the internal buffer to be
+ allocated if addr is None)
+ :param alignment: The alignment of the internally allocated buffer
+ (Valid if addr is None)
+ :param addr: The memory start address to register (if None, the address
+ will be allocated internally)
+ :param access: The desired memory protection attributes (default: 0)
+ :param pgsz_bitmap: Represents the required page sizes
+ :param comp_mask: Compatibility mask
+ """
+ super().__init__()
+ cdef dv.mlx5dv_devx_umem_in umem_in
+
+ if addr is not None:
+ self.addr = <void*><uintptr_t>addr
+ self.is_user_addr = True
+ else:
+ self.addr = <void*><uintptr_t>posix_memalign(size, alignment)
+ memset(self.addr, 0, size)
+ self.is_user_addr = False
+
+ if pgsz_bitmap or comp_mask:
+ umem_in.addr = self.addr
+ umem_in.size = size
+ umem_in.access = access
+ umem_in.pgsz_bitmap = pgsz_bitmap
+ umem_in.comp_mask = comp_mask
+ self.umem = dv.mlx5dv_devx_umem_reg_ex(context.context, &umem_in)
+ else:
+ self.umem = dv.mlx5dv_devx_umem_reg(context.context, self.addr,
+ size, access)
+ if self.umem == NULL:
+ raise PyverbsRDMAErrno("Failed to register a UMEM.")
+ self.context = context
+ self.context.add_ref(self)
+
+ def __dealloc__(self):
+ self.close()
+
+ cpdef close(self):
+ if self.umem != NULL:
+ self.logger.debug('Closing Mlx5UMEM')
+ rc = dv.mlx5dv_devx_umem_dereg(self.umem)
+ try:
+ if rc:
+ raise PyverbsError("Failed to dereg UMEM.", rc)
+ finally:
+ if not self.is_user_addr:
+ free(self.addr)
+ self.umem = NULL
+ self.context = None
+
+ def __str__(self):
+ print_format = '{:20}: {:<20}\n'
+ return print_format.format('umem id', self.umem_id) + \
+ print_format.format('reg addr', self.umem_addr)
+
+ @property
+ def umem_id(self):
+ return self.umem.umem_id
+
+ @property
+ def umem_addr(self):
+ if self.addr:
+ return <uintptr_t><void*>self.addr
--
1.8.3.1
next prev parent reply other threads:[~2021-07-20 8:20 UTC|newest]
Thread overview: 36+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-07-20 8:16 [PATCH rdma-core 00/27] Introduce mlx5 user space driver over VFIO Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 01/27] Update kernel headers Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 02/27] mlx5: Introduce mlx5dv_get_vfio_device_list() Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 03/27] mlx5: Enable debug functionality for vfio Yishai Hadas
2021-07-20 8:51 ` Leon Romanovsky
2021-07-20 9:27 ` Yishai Hadas
2021-07-20 12:27 ` Leon Romanovsky
2021-07-20 14:57 ` Yishai Hadas
2021-07-21 7:05 ` Gal Pressman
2021-07-21 7:58 ` Yishai Hadas
2021-07-21 8:51 ` Gal Pressman
2021-07-20 8:16 ` [PATCH rdma-core 04/27] util: Add interval_set support Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 05/27] verbs: Enable verbs_open_device() to work over non sysfs devices Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 06/27] mlx5: Setup mlx5 vfio context Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 07/27] mlx5: Add mlx5_vfio_cmd_exec() support Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 08/27] mlx5: vfio setup function support Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 09/27] mlx5: vfio setup basic caps Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 10/27] mlx5: Support fast teardown over vfio Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 11/27] mlx5: Enable interrupt command mode " Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 12/27] mlx5: Introduce vfio APIs to process events Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 13/27] mlx5: VFIO poll_health support Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 14/27] mlx5: Implement basic verbs operation for PD and MR over vfio Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 15/27] mlx5: Set DV context ops Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 16/27] mlx5: Support initial DEVX/DV APIs over vfio Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 17/27] mlx5: Implement mlx5dv devx_obj " Yishai Hadas
2021-07-20 8:16 ` Yishai Hadas [this message]
2021-07-20 8:16 ` [PATCH rdma-core 19/27] pyverbs/mlx5: Support EQN querying Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 20/27] pyverbs/mlx5: Support more DevX objects Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 21/27] pyverbs: Add auxiliary memory functions Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 22/27] pyverbs/mlx5: Add support to extract mlx5dv objects Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 23/27] pyverbs/mlx5: Wrap mlx5_cqe64 struct and add enums Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 24/27] tests: Add MAC address to the tests' args Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 25/27] tests: Add mlx5 DevX data path test Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 26/27] pyverbs/mlx5: Support mlx5 devices over VFIO Yishai Hadas
2021-07-20 8:16 ` [PATCH rdma-core 27/27] tests: Add a test for mlx5 " Yishai Hadas
2021-08-01 8:00 ` [PATCH rdma-core 00/27] Introduce mlx5 user space driver " Yishai Hadas
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20210720081647.1980-19-yishaih@nvidia.com \
--to=yishaih@nvidia.com \
--cc=edwards@nvidia.com \
--cc=idok@nvidia.com \
--cc=jgg@nvidia.com \
--cc=linux-rdma@vger.kernel.org \
--cc=maorg@nvidia.com \
--cc=markzhang@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).