From: "Björn Töpel" <bjorn.topel@intel.com>
To: Magnus Karlsson <magnus.karlsson@intel.com>,
ast@kernel.org, daniel@iogearbox.net, netdev@vger.kernel.org,
jonathan.lemon@gmail.com, maximmi@mellanox.com
Cc: bpf@vger.kernel.org, jeffrey.t.kirsher@intel.com,
anthony.l.nguyen@intel.com, maciej.fijalkowski@intel.com,
maciejromanfijalkowski@gmail.com, cristian.dumitrescu@intel.com
Subject: Re: [PATCH bpf-next v4 10/14] xsk: add shared umem support between queue ids
Date: Tue, 28 Jul 2020 09:15:27 +0200 [thread overview]
Message-ID: <f6b54e3c-a4e6-afc1-86cc-48a19a1c7372@intel.com> (raw)
In-Reply-To: <1595307848-20719-11-git-send-email-magnus.karlsson@intel.com>
On 2020-07-21 07:04, Magnus Karlsson wrote:
> Add support to share a umem between queue ids on the same
> device. This mode can be invoked with the XDP_SHARED_UMEM bind
> flag. Previously, sharing was only supported within the same
> queue id and device, and you shared one set of fill and
> completion rings. However, note that when sharing a umem between
> queue ids, you need to create a fill ring and a completion ring
> and tie them to the socket before you do the bind with the
> XDP_SHARED_UMEM flag. This so that the single-producer
> single-consumer semantics can be upheld.
>
> Signed-off-by: Magnus Karlsson <magnus.karlsson@intel.com>
Acked-by: Björn Töpel <bjorn.topel@intel.com>
> ---
> include/net/xsk_buff_pool.h | 2 ++
> net/xdp/xsk.c | 44 ++++++++++++++++++++++++++++++--------------
> net/xdp/xsk_buff_pool.c | 26 ++++++++++++++++++++++++--
> 3 files changed, 56 insertions(+), 16 deletions(-)
>
> diff --git a/include/net/xsk_buff_pool.h b/include/net/xsk_buff_pool.h
> index b4d6307..4d699dd 100644
> --- a/include/net/xsk_buff_pool.h
> +++ b/include/net/xsk_buff_pool.h
> @@ -75,6 +75,8 @@ struct xsk_buff_pool *xp_create_and_assign_umem(struct xdp_sock *xs,
> struct xdp_umem *umem);
> int xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *dev,
> u16 queue_id, u16 flags);
> +int xp_assign_dev_shared(struct xsk_buff_pool *pool, struct xdp_umem *umem,
> + struct net_device *dev, u16 queue_id);
> void xp_destroy(struct xsk_buff_pool *pool);
> void xp_release(struct xdp_buff_xsk *xskb);
> void xp_get_pool(struct xsk_buff_pool *pool);
> diff --git a/net/xdp/xsk.c b/net/xdp/xsk.c
> index d0ff5e8..e897755 100644
> --- a/net/xdp/xsk.c
> +++ b/net/xdp/xsk.c
> @@ -689,12 +689,6 @@ static int xsk_bind(struct socket *sock, struct sockaddr *addr, int addr_len)
> goto out_unlock;
> }
>
> - if (xs->fq_tmp || xs->cq_tmp) {
> - /* Do not allow setting your own fq or cq. */
> - err = -EINVAL;
> - goto out_unlock;
> - }
> -
> sock = xsk_lookup_xsk_from_fd(sxdp->sxdp_shared_umem_fd);
> if (IS_ERR(sock)) {
> err = PTR_ERR(sock);
> @@ -707,15 +701,41 @@ static int xsk_bind(struct socket *sock, struct sockaddr *addr, int addr_len)
> sockfd_put(sock);
> goto out_unlock;
> }
> - if (umem_xs->dev != dev || umem_xs->queue_id != qid) {
> + if (umem_xs->dev != dev) {
> err = -EINVAL;
> sockfd_put(sock);
> goto out_unlock;
> }
>
> - /* Share the buffer pool with the other socket. */
> - xp_get_pool(umem_xs->pool);
> - xs->pool = umem_xs->pool;
> + if (umem_xs->queue_id != qid) {
> + /* Share the umem with another socket on another qid */
> + xs->pool = xp_create_and_assign_umem(xs,
> + umem_xs->umem);
> + if (!xs->pool) {
> + sockfd_put(sock);
> + goto out_unlock;
> + }
> +
> + err = xp_assign_dev_shared(xs->pool, umem_xs->umem,
> + dev, qid);
> + if (err) {
> + xp_destroy(xs->pool);
> + sockfd_put(sock);
> + goto out_unlock;
> + }
> + } else {
> + /* Share the buffer pool with the other socket. */
> + if (xs->fq_tmp || xs->cq_tmp) {
> + /* Do not allow setting your own fq or cq. */
> + err = -EINVAL;
> + sockfd_put(sock);
> + goto out_unlock;
> + }
> +
> + xp_get_pool(umem_xs->pool);
> + xs->pool = umem_xs->pool;
> + }
> +
> xdp_get_umem(umem_xs->umem);
> WRITE_ONCE(xs->umem, umem_xs->umem);
> sockfd_put(sock);
> @@ -847,10 +867,6 @@ static int xsk_setsockopt(struct socket *sock, int level, int optname,
> mutex_unlock(&xs->mutex);
> return -EBUSY;
> }
> - if (!xs->umem) {
> - mutex_unlock(&xs->mutex);
> - return -EINVAL;
> - }
>
> q = (optname == XDP_UMEM_FILL_RING) ? &xs->fq_tmp :
> &xs->cq_tmp;
> diff --git a/net/xdp/xsk_buff_pool.c b/net/xdp/xsk_buff_pool.c
> index ca74a3e..688dc36 100644
> --- a/net/xdp/xsk_buff_pool.c
> +++ b/net/xdp/xsk_buff_pool.c
> @@ -123,8 +123,8 @@ static void xp_disable_drv_zc(struct xsk_buff_pool *pool)
> }
> }
>
> -int xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *netdev,
> - u16 queue_id, u16 flags)
> +static int __xp_assign_dev(struct xsk_buff_pool *pool,
> + struct net_device *netdev, u16 queue_id, u16 flags)
> {
> bool force_zc, force_copy;
> struct netdev_bpf bpf;
> @@ -193,6 +193,28 @@ int xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *netdev,
> return err;
> }
>
> +int xp_assign_dev(struct xsk_buff_pool *pool, struct net_device *dev,
> + u16 queue_id, u16 flags)
> +{
> + return __xp_assign_dev(pool, dev, queue_id, flags);
> +}
> +
> +int xp_assign_dev_shared(struct xsk_buff_pool *pool, struct xdp_umem *umem,
> + struct net_device *dev, u16 queue_id)
> +{
> + u16 flags;
> +
> + /* One fill and completion ring required for each queue id. */
> + if (!pool->fq || !pool->cq)
> + return -EINVAL;
> +
> + flags = umem->zc ? XDP_ZEROCOPY : XDP_COPY;
> + if (pool->uses_need_wakeup)
> + flags |= XDP_USE_NEED_WAKEUP;
> +
> + return __xp_assign_dev(pool, dev, queue_id, flags);
> +}
> +
> void xp_clear_dev(struct xsk_buff_pool *pool)
> {
> if (!pool->netdev)
>
next prev parent reply other threads:[~2020-07-28 7:15 UTC|newest]
Thread overview: 34+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-07-21 5:03 [PATCH bpf-next v4 00/14] xsk: support shared umems between devices and queues Magnus Karlsson
2020-07-21 5:03 ` [PATCH bpf-next v4 01/14] xsk: i40e: ice: ixgbe: mlx5: pass buffer pool to driver instead of umem Magnus Karlsson
2020-07-28 7:04 ` Björn Töpel
2020-07-21 5:03 ` [PATCH bpf-next v4 02/14] xsk: i40e: ice: ixgbe: mlx5: rename xsk zero-copy driver interfaces Magnus Karlsson
2020-07-28 7:04 ` Björn Töpel
2020-07-21 5:03 ` [PATCH bpf-next v4 03/14] xsk: create and free buffer pool independently from umem Magnus Karlsson
2020-07-28 7:05 ` Björn Töpel
2020-07-21 5:03 ` [PATCH bpf-next v4 04/14] xsk: move fill and completion rings to buffer pool Magnus Karlsson
2020-07-28 7:05 ` Björn Töpel
2020-07-21 5:03 ` [PATCH bpf-next v4 05/14] xsk: move queue_id, dev and need_wakeup " Magnus Karlsson
2020-07-28 7:09 ` Björn Töpel
2020-07-29 13:20 ` Magnus Karlsson
2020-07-28 9:21 ` Maxim Mikityanskiy
2020-07-29 13:21 ` Magnus Karlsson
2020-07-21 5:04 ` [PATCH bpf-next v4 06/14] xsk: move xsk_tx_list and its lock " Magnus Karlsson
2020-07-28 7:10 ` Björn Töpel
2020-07-21 5:04 ` [PATCH bpf-next v4 07/14] xsk: move addrs from buffer pool to umem Magnus Karlsson
2020-07-28 7:11 ` Björn Töpel
2020-07-21 5:04 ` [PATCH bpf-next v4 08/14] xsk: enable sharing of dma mappings Magnus Karlsson
2020-07-28 7:14 ` Björn Töpel
2020-07-28 8:59 ` Maxim Mikityanskiy
2020-07-29 13:22 ` Magnus Karlsson
2020-07-21 5:04 ` [PATCH bpf-next v4 09/14] xsk: rearrange internal structs for better performance Magnus Karlsson
2020-07-28 7:14 ` Björn Töpel
2020-07-21 5:04 ` [PATCH bpf-next v4 10/14] xsk: add shared umem support between queue ids Magnus Karlsson
2020-07-28 7:15 ` Björn Töpel [this message]
2020-07-21 5:04 ` [PATCH bpf-next v4 11/14] xsk: add shared umem support between devices Magnus Karlsson
2020-07-28 7:15 ` Björn Töpel
2020-07-21 5:04 ` [PATCH bpf-next v4 12/14] libbpf: support shared umems between queues and devices Magnus Karlsson
2020-07-28 7:18 ` Björn Töpel
2020-07-21 5:04 ` [PATCH bpf-next v4 13/14] samples/bpf: add new sample xsk_fwd.c Magnus Karlsson
2020-07-28 7:18 ` Björn Töpel
2020-07-21 5:04 ` [PATCH bpf-next v4 14/14] xsk: documentation for XDP_SHARED_UMEM between queues and netdevs Magnus Karlsson
2020-07-28 7:18 ` Björn Töpel
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=f6b54e3c-a4e6-afc1-86cc-48a19a1c7372@intel.com \
--to=bjorn.topel@intel.com \
--cc=anthony.l.nguyen@intel.com \
--cc=ast@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=cristian.dumitrescu@intel.com \
--cc=daniel@iogearbox.net \
--cc=jeffrey.t.kirsher@intel.com \
--cc=jonathan.lemon@gmail.com \
--cc=maciej.fijalkowski@intel.com \
--cc=maciejromanfijalkowski@gmail.com \
--cc=magnus.karlsson@intel.com \
--cc=maximmi@mellanox.com \
--cc=netdev@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).