All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH v3 bpf-next 0/2] Better handling of xskmap entries
@ 2019-05-31 18:57 Jonathan Lemon
  2019-05-31 18:57 ` [PATCH v3 bpf-next 1/2] bpf: Allow bpf_map_lookup_elem() on an xskmap Jonathan Lemon
  2019-05-31 18:57 ` [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs Jonathan Lemon
  0 siblings, 2 replies; 8+ messages in thread
From: Jonathan Lemon @ 2019-05-31 18:57 UTC (permalink / raw)
  To: netdev; +Cc: kernel-team, bjorn.topel, magnus.karlsson

Respin of v2 to fix stupid error.

Testing with samples/bpf/xdpsock:
[root@kerneltest004.06.atn2 ~]# ./xdpsock  -i eth0 -q 0

 sock0@eth0:0 rxdrop	
                pps         pkts        1.00       
rx              17          17         
tx              0           0          


Jonathan Lemon (2):
  bpf: Allow bpf_map_lookup_elem() on an xskmap
  libbpf: remove qidconf and better support external bpf programs.

 include/net/xdp_sock.h                        |  6 +-
 kernel/bpf/verifier.c                         |  6 +-
 kernel/bpf/xskmap.c                           |  4 +-
 tools/lib/bpf/xsk.c                           | 79 ++++---------------
 .../bpf/verifier/prevent_map_lookup.c         | 15 ----
 5 files changed, 27 insertions(+), 83 deletions(-)

-- 
2.17.1


^ permalink raw reply	[flat|nested] 8+ messages in thread

* [PATCH v3 bpf-next 1/2] bpf: Allow bpf_map_lookup_elem() on an xskmap
  2019-05-31 18:57 [PATCH v3 bpf-next 0/2] Better handling of xskmap entries Jonathan Lemon
@ 2019-05-31 18:57 ` Jonathan Lemon
  2019-06-01 23:06   ` Song Liu
  2019-05-31 18:57 ` [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs Jonathan Lemon
  1 sibling, 1 reply; 8+ messages in thread
From: Jonathan Lemon @ 2019-05-31 18:57 UTC (permalink / raw)
  To: netdev; +Cc: kernel-team, bjorn.topel, magnus.karlsson

Currently, the AF_XDP code uses a separate map in order to
determine if an xsk is bound to a queue.  Instead of doing this,
have bpf_map_lookup_elem() return the queue_id, as a way of
indicating that there is a valid entry at the map index.

Rearrange some xdp_sock members to eliminate structure holes.

Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com>
Acked-by: Björn Töpel <bjorn.topel@intel.com>
---
 include/net/xdp_sock.h                            |  6 +++---
 kernel/bpf/verifier.c                             |  6 +++++-
 kernel/bpf/xskmap.c                               |  4 +++-
 .../selftests/bpf/verifier/prevent_map_lookup.c   | 15 ---------------
 4 files changed, 11 insertions(+), 20 deletions(-)

diff --git a/include/net/xdp_sock.h b/include/net/xdp_sock.h
index d074b6d60f8a..7d84b1da43d2 100644
--- a/include/net/xdp_sock.h
+++ b/include/net/xdp_sock.h
@@ -57,12 +57,12 @@ struct xdp_sock {
 	struct net_device *dev;
 	struct xdp_umem *umem;
 	struct list_head flush_node;
-	u16 queue_id;
-	struct xsk_queue *tx ____cacheline_aligned_in_smp;
-	struct list_head list;
+	u32 queue_id;
 	bool zc;
 	/* Protects multiple processes in the control path */
 	struct mutex mutex;
+	struct xsk_queue *tx ____cacheline_aligned_in_smp;
+	struct list_head list;
 	/* Mutual exclusion of NAPI TX thread and sendmsg error paths
 	 * in the SKB destructor callback.
 	 */
diff --git a/kernel/bpf/verifier.c b/kernel/bpf/verifier.c
index 2778417e6e0c..91c730f85e92 100644
--- a/kernel/bpf/verifier.c
+++ b/kernel/bpf/verifier.c
@@ -2905,10 +2905,14 @@ static int check_map_func_compatibility(struct bpf_verifier_env *env,
 	 * appear.
 	 */
 	case BPF_MAP_TYPE_CPUMAP:
-	case BPF_MAP_TYPE_XSKMAP:
 		if (func_id != BPF_FUNC_redirect_map)
 			goto error;
 		break;
+	case BPF_MAP_TYPE_XSKMAP:
+		if (func_id != BPF_FUNC_redirect_map &&
+		    func_id != BPF_FUNC_map_lookup_elem)
+			goto error;
+		break;
 	case BPF_MAP_TYPE_ARRAY_OF_MAPS:
 	case BPF_MAP_TYPE_HASH_OF_MAPS:
 		if (func_id != BPF_FUNC_map_lookup_elem)
diff --git a/kernel/bpf/xskmap.c b/kernel/bpf/xskmap.c
index 686d244e798d..249b22089014 100644
--- a/kernel/bpf/xskmap.c
+++ b/kernel/bpf/xskmap.c
@@ -154,7 +154,9 @@ void __xsk_map_flush(struct bpf_map *map)
 
 static void *xsk_map_lookup_elem(struct bpf_map *map, void *key)
 {
-	return ERR_PTR(-EOPNOTSUPP);
+	struct xdp_sock *xs = __xsk_map_lookup_elem(map, *(u32 *)key);
+
+	return xs ? &xs->queue_id : NULL;
 }
 
 static int xsk_map_update_elem(struct bpf_map *map, void *key, void *value,
diff --git a/tools/testing/selftests/bpf/verifier/prevent_map_lookup.c b/tools/testing/selftests/bpf/verifier/prevent_map_lookup.c
index bbdba990fefb..da7a4b37cb98 100644
--- a/tools/testing/selftests/bpf/verifier/prevent_map_lookup.c
+++ b/tools/testing/selftests/bpf/verifier/prevent_map_lookup.c
@@ -28,21 +28,6 @@
 	.errstr = "cannot pass map_type 18 into func bpf_map_lookup_elem",
 	.prog_type = BPF_PROG_TYPE_SOCK_OPS,
 },
-{
-	"prevent map lookup in xskmap",
-	.insns = {
-	BPF_ST_MEM(BPF_DW, BPF_REG_10, -8, 0),
-	BPF_MOV64_REG(BPF_REG_2, BPF_REG_10),
-	BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8),
-	BPF_LD_MAP_FD(BPF_REG_1, 0),
-	BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, 0, 0, BPF_FUNC_map_lookup_elem),
-	BPF_EXIT_INSN(),
-	},
-	.fixup_map_xskmap = { 3 },
-	.result = REJECT,
-	.errstr = "cannot pass map_type 17 into func bpf_map_lookup_elem",
-	.prog_type = BPF_PROG_TYPE_XDP,
-},
 {
 	"prevent map lookup in stack trace",
 	.insns = {
-- 
2.17.1


^ permalink raw reply related	[flat|nested] 8+ messages in thread

* [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs.
  2019-05-31 18:57 [PATCH v3 bpf-next 0/2] Better handling of xskmap entries Jonathan Lemon
  2019-05-31 18:57 ` [PATCH v3 bpf-next 1/2] bpf: Allow bpf_map_lookup_elem() on an xskmap Jonathan Lemon
@ 2019-05-31 18:57 ` Jonathan Lemon
  2019-06-01 23:05   ` Song Liu
  1 sibling, 1 reply; 8+ messages in thread
From: Jonathan Lemon @ 2019-05-31 18:57 UTC (permalink / raw)
  To: netdev; +Cc: kernel-team, bjorn.topel, magnus.karlsson

Use the recent change to XSKMAP bpf_map_lookup_elem() to test if
there is a xsk present in the map instead of duplicating the work
with qidconf.

Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD
bypass any internal bpf maps, so xsk_socket__{create|delete} works
properly.

Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com>
---
 tools/lib/bpf/xsk.c | 79 +++++++++------------------------------------
 1 file changed, 16 insertions(+), 63 deletions(-)

diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c
index 38667b62f1fe..7ce7494b5b50 100644
--- a/tools/lib/bpf/xsk.c
+++ b/tools/lib/bpf/xsk.c
@@ -60,10 +60,8 @@ struct xsk_socket {
 	struct xsk_umem *umem;
 	struct xsk_socket_config config;
 	int fd;
-	int xsks_map;
 	int ifindex;
 	int prog_fd;
-	int qidconf_map_fd;
 	int xsks_map_fd;
 	__u32 queue_id;
 	char ifname[IFNAMSIZ];
@@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk)
 	/* This is the C-program:
 	 * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx)
 	 * {
-	 *     int *qidconf, index = ctx->rx_queue_index;
+	 *     int index = ctx->rx_queue_index;
 	 *
 	 *     // A set entry here means that the correspnding queue_id
 	 *     // has an active AF_XDP socket bound to it.
-	 *     qidconf = bpf_map_lookup_elem(&qidconf_map, &index);
-	 *     if (!qidconf)
-	 *         return XDP_ABORTED;
-	 *
-	 *     if (*qidconf)
+	 *     if (bpf_map_lookup_elem(&xsks_map, &index))
 	 *         return bpf_redirect_map(&xsks_map, index, 0);
 	 *
 	 *     return XDP_PASS;
@@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk)
 		BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4),
 		BPF_MOV64_REG(BPF_REG_2, BPF_REG_10),
 		BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4),
-		BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd),
+		BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd),
 		BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem),
 		BPF_MOV64_REG(BPF_REG_1, BPF_REG_0),
-		BPF_MOV32_IMM(BPF_REG_0, 0),
-		/* if r1 == 0 goto +8 */
-		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8),
 		BPF_MOV32_IMM(BPF_REG_0, 2),
-		/* r1 = *(u32 *)(r1 + 0) */
-		BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0),
 		/* if r1 == 0 goto +5 */
 		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5),
 		/* r2 = *(u32 *)(r10 - 4) */
@@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk)
 	if (max_queues < 0)
 		return max_queues;
 
-	fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map",
+	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
 				 sizeof(int), sizeof(int), max_queues, 0);
 	if (fd < 0)
 		return fd;
-	xsk->qidconf_map_fd = fd;
 
-	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
-				 sizeof(int), sizeof(int), max_queues, 0);
-	if (fd < 0) {
-		close(xsk->qidconf_map_fd);
-		return fd;
-	}
 	xsk->xsks_map_fd = fd;
 
 	return 0;
@@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk)
 
 static void xsk_delete_bpf_maps(struct xsk_socket *xsk)
 {
-	close(xsk->qidconf_map_fd);
+	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
 	close(xsk->xsks_map_fd);
-	xsk->qidconf_map_fd = -1;
-	xsk->xsks_map_fd = -1;
 }
 
 static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
@@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
 	if (err)
 		goto out_map_ids;
 
-	for (i = 0; i < prog_info.nr_map_ids; i++) {
-		if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1)
-			break;
+	xsk->xsks_map_fd = -1;
 
+	for (i = 0; i < prog_info.nr_map_ids; i++) {
 		fd = bpf_map_get_fd_by_id(map_ids[i]);
 		if (fd < 0)
 			continue;
@@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
 			continue;
 		}
 
-		if (!strcmp(map_info.name, "qidconf_map")) {
-			xsk->qidconf_map_fd = fd;
-			continue;
-		}
-
 		if (!strcmp(map_info.name, "xsks_map")) {
 			xsk->xsks_map_fd = fd;
 			continue;
@@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
 	}
 
 	err = 0;
-	if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) {
+	if (xsk->xsks_map_fd == -1)
 		err = -ENOENT;
-		xsk_delete_bpf_maps(xsk);
-	}
 
 out_map_ids:
 	free(map_ids);
 	return err;
 }
 
-static void xsk_clear_bpf_maps(struct xsk_socket *xsk)
-{
-	int qid = false;
-
-	bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
-	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
-}
-
 static int xsk_set_bpf_maps(struct xsk_socket *xsk)
 {
-	int qid = true, fd = xsk->fd, err;
-
-	err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
-	if (err)
-		goto out;
-
-	err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, 0);
-	if (err)
-		goto out;
-
-	return 0;
-out:
-	xsk_clear_bpf_maps(xsk);
-	return err;
+	return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id,
+				   &xsk->fd, 0);
 }
 
 static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
@@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
 
 out_load:
 	close(xsk->prog_fd);
+	xsk->prog_fd = -1;
 out_maps:
 	xsk_delete_bpf_maps(xsk);
 	return err;
@@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket **xsk_ptr, const char *ifname,
 		goto out_mmap_tx;
 	}
 
-	xsk->qidconf_map_fd = -1;
-	xsk->xsks_map_fd = -1;
-
+	xsk->prog_fd = -1;
 	if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
 		err = xsk_setup_xdp_prog(xsk);
 		if (err)
@@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk)
 	if (!xsk)
 		return;
 
-	xsk_clear_bpf_maps(xsk);
-	xsk_delete_bpf_maps(xsk);
+	if (xsk->prog_fd != -1) {
+		xsk_delete_bpf_maps(xsk);
+		close(xsk->prog_fd);
+	}
 
 	optlen = sizeof(off);
 	err = getsockopt(xsk->fd, SOL_XDP, XDP_MMAP_OFFSETS, &off, &optlen);
-- 
2.17.1


^ permalink raw reply related	[flat|nested] 8+ messages in thread

* Re: [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs.
  2019-05-31 18:57 ` [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs Jonathan Lemon
@ 2019-06-01 23:05   ` Song Liu
  2019-06-02  4:18     ` Jonathan Lemon
  0 siblings, 1 reply; 8+ messages in thread
From: Song Liu @ 2019-06-01 23:05 UTC (permalink / raw)
  To: Jonathan Lemon; +Cc: Networking, Kernel Team, bjorn.topel, magnus.karlsson



> On May 31, 2019, at 11:57 AM, Jonathan Lemon <jonathan.lemon@gmail.com> wrote:
> 
> Use the recent change to XSKMAP bpf_map_lookup_elem() to test if
> there is a xsk present in the map instead of duplicating the work
> with qidconf.
> 
> Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD
> bypass any internal bpf maps, so xsk_socket__{create|delete} works
> properly.
> 
> Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com>
> ---
> tools/lib/bpf/xsk.c | 79 +++++++++------------------------------------
> 1 file changed, 16 insertions(+), 63 deletions(-)
> 
> diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c
> index 38667b62f1fe..7ce7494b5b50 100644
> --- a/tools/lib/bpf/xsk.c
> +++ b/tools/lib/bpf/xsk.c
> @@ -60,10 +60,8 @@ struct xsk_socket {
> 	struct xsk_umem *umem;
> 	struct xsk_socket_config config;
> 	int fd;
> -	int xsks_map;
> 	int ifindex;
> 	int prog_fd;
> -	int qidconf_map_fd;
> 	int xsks_map_fd;
> 	__u32 queue_id;
> 	char ifname[IFNAMSIZ];
> @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk)
> 	/* This is the C-program:
> 	 * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx)
> 	 * {
> -	 *     int *qidconf, index = ctx->rx_queue_index;
> +	 *     int index = ctx->rx_queue_index;
> 	 *
> 	 *     // A set entry here means that the correspnding queue_id
> 	 *     // has an active AF_XDP socket bound to it.
> -	 *     qidconf = bpf_map_lookup_elem(&qidconf_map, &index);
> -	 *     if (!qidconf)
> -	 *         return XDP_ABORTED;
> -	 *
> -	 *     if (*qidconf)
> +	 *     if (bpf_map_lookup_elem(&xsks_map, &index))
> 	 *         return bpf_redirect_map(&xsks_map, index, 0);
> 	 *
> 	 *     return XDP_PASS;
> @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk)
> 		BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4),
> 		BPF_MOV64_REG(BPF_REG_2, BPF_REG_10),
> 		BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4),
> -		BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd),
> +		BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd),
> 		BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem),
> 		BPF_MOV64_REG(BPF_REG_1, BPF_REG_0),
> -		BPF_MOV32_IMM(BPF_REG_0, 0),
> -		/* if r1 == 0 goto +8 */
> -		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8),
> 		BPF_MOV32_IMM(BPF_REG_0, 2),
> -		/* r1 = *(u32 *)(r1 + 0) */
> -		BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0),
> 		/* if r1 == 0 goto +5 */
> 		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5),
> 		/* r2 = *(u32 *)(r10 - 4) */
> @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk)
> 	if (max_queues < 0)
> 		return max_queues;
> 
> -	fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map",
> +	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
> 				 sizeof(int), sizeof(int), max_queues, 0);
> 	if (fd < 0)
> 		return fd;
> -	xsk->qidconf_map_fd = fd;
> 
> -	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
> -				 sizeof(int), sizeof(int), max_queues, 0);
> -	if (fd < 0) {
> -		close(xsk->qidconf_map_fd);
> -		return fd;
> -	}
> 	xsk->xsks_map_fd = fd;
> 
> 	return 0;
> @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk)
> 
> static void xsk_delete_bpf_maps(struct xsk_socket *xsk)
> {
> -	close(xsk->qidconf_map_fd);
> +	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
> 	close(xsk->xsks_map_fd);
> -	xsk->qidconf_map_fd = -1;
> -	xsk->xsks_map_fd = -1;
> }
> 
> static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
> @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
> 	if (err)
> 		goto out_map_ids;
> 
> -	for (i = 0; i < prog_info.nr_map_ids; i++) {
> -		if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1)
> -			break;
> +	xsk->xsks_map_fd = -1;
> 
> +	for (i = 0; i < prog_info.nr_map_ids; i++) {
> 		fd = bpf_map_get_fd_by_id(map_ids[i]);
> 		if (fd < 0)
> 			continue;
> @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
> 			continue;
> 		}
> 
> -		if (!strcmp(map_info.name, "qidconf_map")) {
> -			xsk->qidconf_map_fd = fd;
> -			continue;
> -		}
> -
> 		if (!strcmp(map_info.name, "xsks_map")) {
> 			xsk->xsks_map_fd = fd;
> 			continue;
> @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
> 	}
> 
> 	err = 0;
> -	if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) {
> +	if (xsk->xsks_map_fd == -1)
> 		err = -ENOENT;
> -		xsk_delete_bpf_maps(xsk);
> -	}
> 
> out_map_ids:
> 	free(map_ids);
> 	return err;
> }
> 
> -static void xsk_clear_bpf_maps(struct xsk_socket *xsk)
> -{
> -	int qid = false;
> -
> -	bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
> -	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
> -}
> -
> static int xsk_set_bpf_maps(struct xsk_socket *xsk)
> {
> -	int qid = true, fd = xsk->fd, err;
> -
> -	err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
> -	if (err)
> -		goto out;
> -
> -	err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, 0);
> -	if (err)
> -		goto out;
> -
> -	return 0;
> -out:
> -	xsk_clear_bpf_maps(xsk);
> -	return err;
> +	return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id,
> +				   &xsk->fd, 0);
> }
> 
> static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
> @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
> 
> out_load:
> 	close(xsk->prog_fd);
> +	xsk->prog_fd = -1;

I found xsk->prog_fd confusing. Why do we need to set it here? 

I think we don't need to call xsk_delete_bpf_maps() in out_load path? 

> out_maps:
> 	xsk_delete_bpf_maps(xsk);
> 	return err;
> @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket **xsk_ptr, const char *ifname,
> 		goto out_mmap_tx;
> 	}
> 
> -	xsk->qidconf_map_fd = -1;
> -	xsk->xsks_map_fd = -1;
> -
> +	xsk->prog_fd = -1;
> 	if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
> 		err = xsk_setup_xdp_prog(xsk);
> 		if (err)
> @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk)
> 	if (!xsk)
> 		return;
> 
> -	xsk_clear_bpf_maps(xsk);
> -	xsk_delete_bpf_maps(xsk);
> +	if (xsk->prog_fd != -1) {
> +		xsk_delete_bpf_maps(xsk);
> +		close(xsk->prog_fd);

Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is 
confusing. I looked at the code for quite sometime, but still cannot 
confirm it is correct. 

Thanks,
Song

> +	}
> 
> 	optlen = sizeof(off);
> 	err = getsockopt(xsk->fd, SOL_XDP, XDP_MMAP_OFFSETS, &off, &optlen);
> -- 
> 2.17.1
> 


^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v3 bpf-next 1/2] bpf: Allow bpf_map_lookup_elem() on an xskmap
  2019-05-31 18:57 ` [PATCH v3 bpf-next 1/2] bpf: Allow bpf_map_lookup_elem() on an xskmap Jonathan Lemon
@ 2019-06-01 23:06   ` Song Liu
  0 siblings, 0 replies; 8+ messages in thread
From: Song Liu @ 2019-06-01 23:06 UTC (permalink / raw)
  To: Jonathan Lemon; +Cc: Networking, Kernel Team, bjorn.topel, magnus.karlsson



> On May 31, 2019, at 11:57 AM, Jonathan Lemon <jonathan.lemon@gmail.com> wrote:
> 
> Currently, the AF_XDP code uses a separate map in order to
> determine if an xsk is bound to a queue.  Instead of doing this,
> have bpf_map_lookup_elem() return the queue_id, as a way of
> indicating that there is a valid entry at the map index.
> 
> Rearrange some xdp_sock members to eliminate structure holes.
> 
> Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com>
> Acked-by: Björn Töpel <bjorn.topel@intel.com>

Acked-by: Song Liu <songliubraving@fb.com>

> ---
> include/net/xdp_sock.h                            |  6 +++---
> kernel/bpf/verifier.c                             |  6 +++++-
> kernel/bpf/xskmap.c                               |  4 +++-
> .../selftests/bpf/verifier/prevent_map_lookup.c   | 15 ---------------
> 4 files changed, 11 insertions(+), 20 deletions(-)
> 
> diff --git a/include/net/xdp_sock.h b/include/net/xdp_sock.h
> index d074b6d60f8a..7d84b1da43d2 100644
> --- a/include/net/xdp_sock.h
> +++ b/include/net/xdp_sock.h
> @@ -57,12 +57,12 @@ struct xdp_sock {
> 	struct net_device *dev;
> 	struct xdp_umem *umem;
> 	struct list_head flush_node;
> -	u16 queue_id;
> -	struct xsk_queue *tx ____cacheline_aligned_in_smp;
> -	struct list_head list;
> +	u32 queue_id;
> 	bool zc;
> 	/* Protects multiple processes in the control path */
> 	struct mutex mutex;
> +	struct xsk_queue *tx ____cacheline_aligned_in_smp;
> +	struct list_head list;
> 	/* Mutual exclusion of NAPI TX thread and sendmsg error paths
> 	 * in the SKB destructor callback.
> 	 */
> diff --git a/kernel/bpf/verifier.c b/kernel/bpf/verifier.c
> index 2778417e6e0c..91c730f85e92 100644
> --- a/kernel/bpf/verifier.c
> +++ b/kernel/bpf/verifier.c
> @@ -2905,10 +2905,14 @@ static int check_map_func_compatibility(struct bpf_verifier_env *env,
> 	 * appear.
> 	 */
> 	case BPF_MAP_TYPE_CPUMAP:
> -	case BPF_MAP_TYPE_XSKMAP:
> 		if (func_id != BPF_FUNC_redirect_map)
> 			goto error;
> 		break;
> +	case BPF_MAP_TYPE_XSKMAP:
> +		if (func_id != BPF_FUNC_redirect_map &&
> +		    func_id != BPF_FUNC_map_lookup_elem)
> +			goto error;
> +		break;
> 	case BPF_MAP_TYPE_ARRAY_OF_MAPS:
> 	case BPF_MAP_TYPE_HASH_OF_MAPS:
> 		if (func_id != BPF_FUNC_map_lookup_elem)
> diff --git a/kernel/bpf/xskmap.c b/kernel/bpf/xskmap.c
> index 686d244e798d..249b22089014 100644
> --- a/kernel/bpf/xskmap.c
> +++ b/kernel/bpf/xskmap.c
> @@ -154,7 +154,9 @@ void __xsk_map_flush(struct bpf_map *map)
> 
> static void *xsk_map_lookup_elem(struct bpf_map *map, void *key)
> {
> -	return ERR_PTR(-EOPNOTSUPP);
> +	struct xdp_sock *xs = __xsk_map_lookup_elem(map, *(u32 *)key);
> +
> +	return xs ? &xs->queue_id : NULL;
> }
> 
> static int xsk_map_update_elem(struct bpf_map *map, void *key, void *value,
> diff --git a/tools/testing/selftests/bpf/verifier/prevent_map_lookup.c b/tools/testing/selftests/bpf/verifier/prevent_map_lookup.c
> index bbdba990fefb..da7a4b37cb98 100644
> --- a/tools/testing/selftests/bpf/verifier/prevent_map_lookup.c
> +++ b/tools/testing/selftests/bpf/verifier/prevent_map_lookup.c
> @@ -28,21 +28,6 @@
> 	.errstr = "cannot pass map_type 18 into func bpf_map_lookup_elem",
> 	.prog_type = BPF_PROG_TYPE_SOCK_OPS,
> },
> -{
> -	"prevent map lookup in xskmap",
> -	.insns = {
> -	BPF_ST_MEM(BPF_DW, BPF_REG_10, -8, 0),
> -	BPF_MOV64_REG(BPF_REG_2, BPF_REG_10),
> -	BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8),
> -	BPF_LD_MAP_FD(BPF_REG_1, 0),
> -	BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, 0, 0, BPF_FUNC_map_lookup_elem),
> -	BPF_EXIT_INSN(),
> -	},
> -	.fixup_map_xskmap = { 3 },
> -	.result = REJECT,
> -	.errstr = "cannot pass map_type 17 into func bpf_map_lookup_elem",
> -	.prog_type = BPF_PROG_TYPE_XDP,
> -},
> {
> 	"prevent map lookup in stack trace",
> 	.insns = {
> -- 
> 2.17.1
> 


^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs.
  2019-06-01 23:05   ` Song Liu
@ 2019-06-02  4:18     ` Jonathan Lemon
  2019-06-02  6:55       ` Song Liu
  2019-06-03 11:11       ` Björn Töpel
  0 siblings, 2 replies; 8+ messages in thread
From: Jonathan Lemon @ 2019-06-02  4:18 UTC (permalink / raw)
  To: Song Liu; +Cc: Networking, Kernel Team, bjorn.topel, magnus.karlsson



On 1 Jun 2019, at 16:05, Song Liu wrote:

>> On May 31, 2019, at 11:57 AM, Jonathan Lemon 
>> <jonathan.lemon@gmail.com> wrote:
>>
>> Use the recent change to XSKMAP bpf_map_lookup_elem() to test if
>> there is a xsk present in the map instead of duplicating the work
>> with qidconf.
>>
>> Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD
>> bypass any internal bpf maps, so xsk_socket__{create|delete} works
>> properly.
>>
>> Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com>
>> ---
>> tools/lib/bpf/xsk.c | 79 
>> +++++++++------------------------------------
>> 1 file changed, 16 insertions(+), 63 deletions(-)
>>
>> diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c
>> index 38667b62f1fe..7ce7494b5b50 100644
>> --- a/tools/lib/bpf/xsk.c
>> +++ b/tools/lib/bpf/xsk.c
>> @@ -60,10 +60,8 @@ struct xsk_socket {
>> 	struct xsk_umem *umem;
>> 	struct xsk_socket_config config;
>> 	int fd;
>> -	int xsks_map;
>> 	int ifindex;
>> 	int prog_fd;
>> -	int qidconf_map_fd;
>> 	int xsks_map_fd;
>> 	__u32 queue_id;
>> 	char ifname[IFNAMSIZ];
>> @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket 
>> *xsk)
>> 	/* This is the C-program:
>> 	 * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx)
>> 	 * {
>> -	 *     int *qidconf, index = ctx->rx_queue_index;
>> +	 *     int index = ctx->rx_queue_index;
>> 	 *
>> 	 *     // A set entry here means that the correspnding queue_id
>> 	 *     // has an active AF_XDP socket bound to it.
>> -	 *     qidconf = bpf_map_lookup_elem(&qidconf_map, &index);
>> -	 *     if (!qidconf)
>> -	 *         return XDP_ABORTED;
>> -	 *
>> -	 *     if (*qidconf)
>> +	 *     if (bpf_map_lookup_elem(&xsks_map, &index))
>> 	 *         return bpf_redirect_map(&xsks_map, index, 0);
>> 	 *
>> 	 *     return XDP_PASS;
>> @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket 
>> *xsk)
>> 		BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4),
>> 		BPF_MOV64_REG(BPF_REG_2, BPF_REG_10),
>> 		BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4),
>> -		BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd),
>> +		BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd),
>> 		BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem),
>> 		BPF_MOV64_REG(BPF_REG_1, BPF_REG_0),
>> -		BPF_MOV32_IMM(BPF_REG_0, 0),
>> -		/* if r1 == 0 goto +8 */
>> -		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8),
>> 		BPF_MOV32_IMM(BPF_REG_0, 2),
>> -		/* r1 = *(u32 *)(r1 + 0) */
>> -		BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0),
>> 		/* if r1 == 0 goto +5 */
>> 		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5),
>> 		/* r2 = *(u32 *)(r10 - 4) */
>> @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct 
>> xsk_socket *xsk)
>> 	if (max_queues < 0)
>> 		return max_queues;
>>
>> -	fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map",
>> +	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
>> 				 sizeof(int), sizeof(int), max_queues, 0);
>> 	if (fd < 0)
>> 		return fd;
>> -	xsk->qidconf_map_fd = fd;
>>
>> -	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
>> -				 sizeof(int), sizeof(int), max_queues, 0);
>> -	if (fd < 0) {
>> -		close(xsk->qidconf_map_fd);
>> -		return fd;
>> -	}
>> 	xsk->xsks_map_fd = fd;
>>
>> 	return 0;
>> @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket 
>> *xsk)
>>
>> static void xsk_delete_bpf_maps(struct xsk_socket *xsk)
>> {
>> -	close(xsk->qidconf_map_fd);
>> +	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
>> 	close(xsk->xsks_map_fd);
>> -	xsk->qidconf_map_fd = -1;
>> -	xsk->xsks_map_fd = -1;
>> }
>>
>> static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
>> @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket 
>> *xsk)
>> 	if (err)
>> 		goto out_map_ids;
>>
>> -	for (i = 0; i < prog_info.nr_map_ids; i++) {
>> -		if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1)
>> -			break;
>> +	xsk->xsks_map_fd = -1;
>>
>> +	for (i = 0; i < prog_info.nr_map_ids; i++) {
>> 		fd = bpf_map_get_fd_by_id(map_ids[i]);
>> 		if (fd < 0)
>> 			continue;
>> @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket 
>> *xsk)
>> 			continue;
>> 		}
>>
>> -		if (!strcmp(map_info.name, "qidconf_map")) {
>> -			xsk->qidconf_map_fd = fd;
>> -			continue;
>> -		}
>> -
>> 		if (!strcmp(map_info.name, "xsks_map")) {
>> 			xsk->xsks_map_fd = fd;
>> 			continue;
>> @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct 
>> xsk_socket *xsk)
>> 	}
>>
>> 	err = 0;
>> -	if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) {
>> +	if (xsk->xsks_map_fd == -1)
>> 		err = -ENOENT;
>> -		xsk_delete_bpf_maps(xsk);
>> -	}
>>
>> out_map_ids:
>> 	free(map_ids);
>> 	return err;
>> }
>>
>> -static void xsk_clear_bpf_maps(struct xsk_socket *xsk)
>> -{
>> -	int qid = false;
>> -
>> -	bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
>> -	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
>> -}
>> -
>> static int xsk_set_bpf_maps(struct xsk_socket *xsk)
>> {
>> -	int qid = true, fd = xsk->fd, err;
>> -
>> -	err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, 
>> &qid, 0);
>> -	if (err)
>> -		goto out;
>> -
>> -	err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, 
>> 0);
>> -	if (err)
>> -		goto out;
>> -
>> -	return 0;
>> -out:
>> -	xsk_clear_bpf_maps(xsk);
>> -	return err;
>> +	return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id,
>> +				   &xsk->fd, 0);
>> }
>>
>> static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
>> @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket 
>> *xsk)
>>
>> out_load:
>> 	close(xsk->prog_fd);
>> +	xsk->prog_fd = -1;
>
> I found xsk->prog_fd confusing. Why do we need to set it here?

I suppose this one isn't strictly required - I set it as a guard out of 
habit.
xsk is (currently) immediately freed by the caller, so it can be 
removed.


The main logic is:

         xsk->prog_fd = -1;
         if (!(xsk->config.libbpf_flags & 
XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
                 err = xsk_setup_xdp_prog(xsk);

The user may pass INHIBIT_PROG_LOAD, which bypasses setting up the xdp 
program
(and any maps associated with the program), allowing installation of a 
custom
program.  The cleanup behavior is then gated on prog_fd being -1,

>
> I think we don't need to call xsk_delete_bpf_maps() in out_load path?

Hmm, there's two out_load paths, but only one needs the delete maps 
call.  Let
me redo the error handling so it's a bit more explicit.


>
>> out_maps:
>> 	xsk_delete_bpf_maps(xsk);
>> 	return err;
>> @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket 
>> **xsk_ptr, const char *ifname,
>> 		goto out_mmap_tx;
>> 	}
>>
>> -	xsk->qidconf_map_fd = -1;
>> -	xsk->xsks_map_fd = -1;
>> -
>> +	xsk->prog_fd = -1;
>> 	if (!(xsk->config.libbpf_flags & 
>> XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
>> 		err = xsk_setup_xdp_prog(xsk);
>> 		if (err)
>> @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk)
>> 	if (!xsk)
>> 		return;
>>
>> -	xsk_clear_bpf_maps(xsk);
>> -	xsk_delete_bpf_maps(xsk);
>> +	if (xsk->prog_fd != -1) {
>> +		xsk_delete_bpf_maps(xsk);
>> +		close(xsk->prog_fd);
>
> Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is
> confusing. I looked at the code for quite sometime, but still cannot
> confirm it is correct.

See above reasoning - with INHIBIT_PROG_LOAD, there is no 
library-provided
program or maps, so cleanup actions are skipped.
-- 
Jonathan

^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs.
  2019-06-02  4:18     ` Jonathan Lemon
@ 2019-06-02  6:55       ` Song Liu
  2019-06-03 11:11       ` Björn Töpel
  1 sibling, 0 replies; 8+ messages in thread
From: Song Liu @ 2019-06-02  6:55 UTC (permalink / raw)
  To: Jonathan Lemon; +Cc: Networking, Kernel Team, bjorn.topel, magnus.karlsson



> On Jun 1, 2019, at 9:18 PM, Jonathan Lemon <jonathan.lemon@gmail.com> wrote:
> 
> 
> 
> On 1 Jun 2019, at 16:05, Song Liu wrote:
> 
>>> On May 31, 2019, at 11:57 AM, Jonathan Lemon <jonathan.lemon@gmail.com> wrote:
>>> 
>>> Use the recent change to XSKMAP bpf_map_lookup_elem() to test if
>>> there is a xsk present in the map instead of duplicating the work
>>> with qidconf.
>>> 
>>> Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD
>>> bypass any internal bpf maps, so xsk_socket__{create|delete} works
>>> properly.
>>> 
>>> Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com>
>>> ---
>>> tools/lib/bpf/xsk.c | 79 +++++++++------------------------------------
>>> 1 file changed, 16 insertions(+), 63 deletions(-)
>>> 
>>> diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c
>>> index 38667b62f1fe..7ce7494b5b50 100644
>>> --- a/tools/lib/bpf/xsk.c
>>> +++ b/tools/lib/bpf/xsk.c
>>> @@ -60,10 +60,8 @@ struct xsk_socket {
>>> 	struct xsk_umem *umem;
>>> 	struct xsk_socket_config config;
>>> 	int fd;
>>> -	int xsks_map;
>>> 	int ifindex;
>>> 	int prog_fd;
>>> -	int qidconf_map_fd;
>>> 	int xsks_map_fd;
>>> 	__u32 queue_id;
>>> 	char ifname[IFNAMSIZ];
>>> @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk)
>>> 	/* This is the C-program:
>>> 	 * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx)
>>> 	 * {
>>> -	 *     int *qidconf, index = ctx->rx_queue_index;
>>> +	 *     int index = ctx->rx_queue_index;
>>> 	 *
>>> 	 *     // A set entry here means that the correspnding queue_id
>>> 	 *     // has an active AF_XDP socket bound to it.
>>> -	 *     qidconf = bpf_map_lookup_elem(&qidconf_map, &index);
>>> -	 *     if (!qidconf)
>>> -	 *         return XDP_ABORTED;
>>> -	 *
>>> -	 *     if (*qidconf)
>>> +	 *     if (bpf_map_lookup_elem(&xsks_map, &index))
>>> 	 *         return bpf_redirect_map(&xsks_map, index, 0);
>>> 	 *
>>> 	 *     return XDP_PASS;
>>> @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket *xsk)
>>> 		BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4),
>>> 		BPF_MOV64_REG(BPF_REG_2, BPF_REG_10),
>>> 		BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4),
>>> -		BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd),
>>> +		BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd),
>>> 		BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem),
>>> 		BPF_MOV64_REG(BPF_REG_1, BPF_REG_0),
>>> -		BPF_MOV32_IMM(BPF_REG_0, 0),
>>> -		/* if r1 == 0 goto +8 */
>>> -		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8),
>>> 		BPF_MOV32_IMM(BPF_REG_0, 2),
>>> -		/* r1 = *(u32 *)(r1 + 0) */
>>> -		BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0),
>>> 		/* if r1 == 0 goto +5 */
>>> 		BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5),
>>> 		/* r2 = *(u32 *)(r10 - 4) */
>>> @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk)
>>> 	if (max_queues < 0)
>>> 		return max_queues;
>>> 
>>> -	fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map",
>>> +	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
>>> 				 sizeof(int), sizeof(int), max_queues, 0);
>>> 	if (fd < 0)
>>> 		return fd;
>>> -	xsk->qidconf_map_fd = fd;
>>> 
>>> -	fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
>>> -				 sizeof(int), sizeof(int), max_queues, 0);
>>> -	if (fd < 0) {
>>> -		close(xsk->qidconf_map_fd);
>>> -		return fd;
>>> -	}
>>> 	xsk->xsks_map_fd = fd;
>>> 
>>> 	return 0;
>>> @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket *xsk)
>>> 
>>> static void xsk_delete_bpf_maps(struct xsk_socket *xsk)
>>> {
>>> -	close(xsk->qidconf_map_fd);
>>> +	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
>>> 	close(xsk->xsks_map_fd);
>>> -	xsk->qidconf_map_fd = -1;
>>> -	xsk->xsks_map_fd = -1;
>>> }
>>> 
>>> static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
>>> @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
>>> 	if (err)
>>> 		goto out_map_ids;
>>> 
>>> -	for (i = 0; i < prog_info.nr_map_ids; i++) {
>>> -		if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1)
>>> -			break;
>>> +	xsk->xsks_map_fd = -1;
>>> 
>>> +	for (i = 0; i < prog_info.nr_map_ids; i++) {
>>> 		fd = bpf_map_get_fd_by_id(map_ids[i]);
>>> 		if (fd < 0)
>>> 			continue;
>>> @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
>>> 			continue;
>>> 		}
>>> 
>>> -		if (!strcmp(map_info.name, "qidconf_map")) {
>>> -			xsk->qidconf_map_fd = fd;
>>> -			continue;
>>> -		}
>>> -
>>> 		if (!strcmp(map_info.name, "xsks_map")) {
>>> 			xsk->xsks_map_fd = fd;
>>> 			continue;
>>> @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
>>> 	}
>>> 
>>> 	err = 0;
>>> -	if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) {
>>> +	if (xsk->xsks_map_fd == -1)
>>> 		err = -ENOENT;
>>> -		xsk_delete_bpf_maps(xsk);
>>> -	}
>>> 
>>> out_map_ids:
>>> 	free(map_ids);
>>> 	return err;
>>> }
>>> 
>>> -static void xsk_clear_bpf_maps(struct xsk_socket *xsk)
>>> -{
>>> -	int qid = false;
>>> -
>>> -	bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
>>> -	bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
>>> -}
>>> -
>>> static int xsk_set_bpf_maps(struct xsk_socket *xsk)
>>> {
>>> -	int qid = true, fd = xsk->fd, err;
>>> -
>>> -	err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
>>> -	if (err)
>>> -		goto out;
>>> -
>>> -	err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd, 0);
>>> -	if (err)
>>> -		goto out;
>>> -
>>> -	return 0;
>>> -out:
>>> -	xsk_clear_bpf_maps(xsk);
>>> -	return err;
>>> +	return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id,
>>> +				   &xsk->fd, 0);
>>> }
>>> 
>>> static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
>>> @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
>>> 
>>> out_load:
>>> 	close(xsk->prog_fd);
>>> +	xsk->prog_fd = -1;
>> 
>> I found xsk->prog_fd confusing. Why do we need to set it here?
> 
> I suppose this one isn't strictly required - I set it as a guard out of habit.
> xsk is (currently) immediately freed by the caller, so it can be removed.
> 
> 
> The main logic is:
> 
>        xsk->prog_fd = -1;
>        if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
>                err = xsk_setup_xdp_prog(xsk);
> 
> The user may pass INHIBIT_PROG_LOAD, which bypasses setting up the xdp program
> (and any maps associated with the program), allowing installation of a custom
> program.  The cleanup behavior is then gated on prog_fd being -1,
> 
>> 
>> I think we don't need to call xsk_delete_bpf_maps() in out_load path?
> 
> Hmm, there's two out_load paths, but only one needs the delete maps call.  Let
> me redo the error handling so it's a bit more explicit.
> 
> 
>> 
>>> out_maps:
>>> 	xsk_delete_bpf_maps(xsk);
>>> 	return err;
>>> @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket **xsk_ptr, const char *ifname,
>>> 		goto out_mmap_tx;
>>> 	}
>>> 
>>> -	xsk->qidconf_map_fd = -1;
>>> -	xsk->xsks_map_fd = -1;
>>> -
>>> +	xsk->prog_fd = -1;
>>> 	if (!(xsk->config.libbpf_flags & XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
>>> 		err = xsk_setup_xdp_prog(xsk);
>>> 		if (err)
>>> @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk)
>>> 	if (!xsk)
>>> 		return;
>>> 
>>> -	xsk_clear_bpf_maps(xsk);
>>> -	xsk_delete_bpf_maps(xsk);
>>> +	if (xsk->prog_fd != -1) {
>>> +		xsk_delete_bpf_maps(xsk);
>>> +		close(xsk->prog_fd);
>> 
>> Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is
>> confusing. I looked at the code for quite sometime, but still cannot
>> confirm it is correct.
> 
> See above reasoning - with INHIBIT_PROG_LOAD, there is no library-provided
> program or maps, so cleanup actions are skipped.
> -- 
> Jonathan

Thanks for the explanation. 

Acked-by: Song Liu <songliubraving@fb.com>




^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs.
  2019-06-02  4:18     ` Jonathan Lemon
  2019-06-02  6:55       ` Song Liu
@ 2019-06-03 11:11       ` Björn Töpel
  1 sibling, 0 replies; 8+ messages in thread
From: Björn Töpel @ 2019-06-03 11:11 UTC (permalink / raw)
  To: Jonathan Lemon
  Cc: Song Liu, Networking, Kernel Team, Björn Töpel,
	Karlsson, Magnus

On Sun, 2 Jun 2019 at 06:19, Jonathan Lemon <jonathan.lemon@gmail.com> wrote:
>
>
>
> On 1 Jun 2019, at 16:05, Song Liu wrote:
>
> >> On May 31, 2019, at 11:57 AM, Jonathan Lemon
> >> <jonathan.lemon@gmail.com> wrote:
> >>
> >> Use the recent change to XSKMAP bpf_map_lookup_elem() to test if
> >> there is a xsk present in the map instead of duplicating the work
> >> with qidconf.
> >>
> >> Fix things so callers using XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD
> >> bypass any internal bpf maps, so xsk_socket__{create|delete} works
> >> properly.
> >>
> >> Signed-off-by: Jonathan Lemon <jonathan.lemon@gmail.com>
> >> ---
> >> tools/lib/bpf/xsk.c | 79
> >> +++++++++------------------------------------
> >> 1 file changed, 16 insertions(+), 63 deletions(-)
> >>
> >> diff --git a/tools/lib/bpf/xsk.c b/tools/lib/bpf/xsk.c
> >> index 38667b62f1fe..7ce7494b5b50 100644
> >> --- a/tools/lib/bpf/xsk.c
> >> +++ b/tools/lib/bpf/xsk.c
> >> @@ -60,10 +60,8 @@ struct xsk_socket {
> >>      struct xsk_umem *umem;
> >>      struct xsk_socket_config config;
> >>      int fd;
> >> -    int xsks_map;
> >>      int ifindex;
> >>      int prog_fd;
> >> -    int qidconf_map_fd;
> >>      int xsks_map_fd;
> >>      __u32 queue_id;
> >>      char ifname[IFNAMSIZ];
> >> @@ -265,15 +263,11 @@ static int xsk_load_xdp_prog(struct xsk_socket
> >> *xsk)
> >>      /* This is the C-program:
> >>       * SEC("xdp_sock") int xdp_sock_prog(struct xdp_md *ctx)
> >>       * {
> >> -     *     int *qidconf, index = ctx->rx_queue_index;
> >> +     *     int index = ctx->rx_queue_index;
> >>       *
> >>       *     // A set entry here means that the correspnding queue_id
> >>       *     // has an active AF_XDP socket bound to it.
> >> -     *     qidconf = bpf_map_lookup_elem(&qidconf_map, &index);
> >> -     *     if (!qidconf)
> >> -     *         return XDP_ABORTED;
> >> -     *
> >> -     *     if (*qidconf)
> >> +     *     if (bpf_map_lookup_elem(&xsks_map, &index))
> >>       *         return bpf_redirect_map(&xsks_map, index, 0);
> >>       *
> >>       *     return XDP_PASS;
> >> @@ -286,15 +280,10 @@ static int xsk_load_xdp_prog(struct xsk_socket
> >> *xsk)
> >>              BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_1, -4),
> >>              BPF_MOV64_REG(BPF_REG_2, BPF_REG_10),
> >>              BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4),
> >> -            BPF_LD_MAP_FD(BPF_REG_1, xsk->qidconf_map_fd),
> >> +            BPF_LD_MAP_FD(BPF_REG_1, xsk->xsks_map_fd),
> >>              BPF_EMIT_CALL(BPF_FUNC_map_lookup_elem),
> >>              BPF_MOV64_REG(BPF_REG_1, BPF_REG_0),
> >> -            BPF_MOV32_IMM(BPF_REG_0, 0),
> >> -            /* if r1 == 0 goto +8 */
> >> -            BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 8),
> >>              BPF_MOV32_IMM(BPF_REG_0, 2),
> >> -            /* r1 = *(u32 *)(r1 + 0) */
> >> -            BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_1, 0),
> >>              /* if r1 == 0 goto +5 */
> >>              BPF_JMP_IMM(BPF_JEQ, BPF_REG_1, 0, 5),
> >>              /* r2 = *(u32 *)(r10 - 4) */
> >> @@ -366,18 +355,11 @@ static int xsk_create_bpf_maps(struct
> >> xsk_socket *xsk)
> >>      if (max_queues < 0)
> >>              return max_queues;
> >>
> >> -    fd = bpf_create_map_name(BPF_MAP_TYPE_ARRAY, "qidconf_map",
> >> +    fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
> >>                               sizeof(int), sizeof(int), max_queues, 0);
> >>      if (fd < 0)
> >>              return fd;
> >> -    xsk->qidconf_map_fd = fd;
> >>
> >> -    fd = bpf_create_map_name(BPF_MAP_TYPE_XSKMAP, "xsks_map",
> >> -                             sizeof(int), sizeof(int), max_queues, 0);
> >> -    if (fd < 0) {
> >> -            close(xsk->qidconf_map_fd);
> >> -            return fd;
> >> -    }
> >>      xsk->xsks_map_fd = fd;
> >>
> >>      return 0;
> >> @@ -385,10 +367,8 @@ static int xsk_create_bpf_maps(struct xsk_socket
> >> *xsk)
> >>
> >> static void xsk_delete_bpf_maps(struct xsk_socket *xsk)
> >> {
> >> -    close(xsk->qidconf_map_fd);
> >> +    bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
> >>      close(xsk->xsks_map_fd);
> >> -    xsk->qidconf_map_fd = -1;
> >> -    xsk->xsks_map_fd = -1;
> >> }
> >>
> >> static int xsk_lookup_bpf_maps(struct xsk_socket *xsk)
> >> @@ -417,10 +397,9 @@ static int xsk_lookup_bpf_maps(struct xsk_socket
> >> *xsk)
> >>      if (err)
> >>              goto out_map_ids;
> >>
> >> -    for (i = 0; i < prog_info.nr_map_ids; i++) {
> >> -            if (xsk->qidconf_map_fd != -1 && xsk->xsks_map_fd != -1)
> >> -                    break;
> >> +    xsk->xsks_map_fd = -1;
> >>
> >> +    for (i = 0; i < prog_info.nr_map_ids; i++) {
> >>              fd = bpf_map_get_fd_by_id(map_ids[i]);
> >>              if (fd < 0)
> >>                      continue;
> >> @@ -431,11 +410,6 @@ static int xsk_lookup_bpf_maps(struct xsk_socket
> >> *xsk)
> >>                      continue;
> >>              }
> >>
> >> -            if (!strcmp(map_info.name, "qidconf_map")) {
> >> -                    xsk->qidconf_map_fd = fd;
> >> -                    continue;
> >> -            }
> >> -
> >>              if (!strcmp(map_info.name, "xsks_map")) {
> >>                      xsk->xsks_map_fd = fd;
> >>                      continue;
> >> @@ -445,40 +419,18 @@ static int xsk_lookup_bpf_maps(struct
> >> xsk_socket *xsk)
> >>      }
> >>
> >>      err = 0;
> >> -    if (xsk->qidconf_map_fd < 0 || xsk->xsks_map_fd < 0) {
> >> +    if (xsk->xsks_map_fd == -1)
> >>              err = -ENOENT;
> >> -            xsk_delete_bpf_maps(xsk);
> >> -    }
> >>
> >> out_map_ids:
> >>      free(map_ids);
> >>      return err;
> >> }
> >>
> >> -static void xsk_clear_bpf_maps(struct xsk_socket *xsk)
> >> -{
> >> -    int qid = false;
> >> -
> >> -    bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id, &qid, 0);
> >> -    bpf_map_delete_elem(xsk->xsks_map_fd, &xsk->queue_id);
> >> -}
> >> -
> >> static int xsk_set_bpf_maps(struct xsk_socket *xsk)
> >> {
> >> -    int qid = true, fd = xsk->fd, err;
> >> -
> >> -    err = bpf_map_update_elem(xsk->qidconf_map_fd, &xsk->queue_id,
> >> &qid, 0);
> >> -    if (err)
> >> -            goto out;
> >> -
> >> -    err = bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id, &fd,
> >> 0);
> >> -    if (err)
> >> -            goto out;
> >> -
> >> -    return 0;
> >> -out:
> >> -    xsk_clear_bpf_maps(xsk);
> >> -    return err;
> >> +    return bpf_map_update_elem(xsk->xsks_map_fd, &xsk->queue_id,
> >> +                               &xsk->fd, 0);
> >> }
> >>
> >> static int xsk_setup_xdp_prog(struct xsk_socket *xsk)
> >> @@ -514,6 +466,7 @@ static int xsk_setup_xdp_prog(struct xsk_socket
> >> *xsk)
> >>
> >> out_load:
> >>      close(xsk->prog_fd);
> >> +    xsk->prog_fd = -1;
> >
> > I found xsk->prog_fd confusing. Why do we need to set it here?
>
> I suppose this one isn't strictly required - I set it as a guard out of
> habit.
> xsk is (currently) immediately freed by the caller, so it can be
> removed.
>
>
> The main logic is:
>
>          xsk->prog_fd = -1;
>          if (!(xsk->config.libbpf_flags &
> XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
>                  err = xsk_setup_xdp_prog(xsk);
>
> The user may pass INHIBIT_PROG_LOAD, which bypasses setting up the xdp
> program
> (and any maps associated with the program), allowing installation of a
> custom
> program.  The cleanup behavior is then gated on prog_fd being -1,
>
> >
> > I think we don't need to call xsk_delete_bpf_maps() in out_load path?
>
> Hmm, there's two out_load paths, but only one needs the delete maps
> call.  Let
> me redo the error handling so it's a bit more explicit.
>

You're going for a V4?

I took this series for a spin:
Tested-by: Björn Töpel <bjorn.topel@intel.com>

Very nice cleanup, Jonathan!


Björn


>
> >
> >> out_maps:
> >>      xsk_delete_bpf_maps(xsk);
> >>      return err;
> >> @@ -643,9 +596,7 @@ int xsk_socket__create(struct xsk_socket
> >> **xsk_ptr, const char *ifname,
> >>              goto out_mmap_tx;
> >>      }
> >>
> >> -    xsk->qidconf_map_fd = -1;
> >> -    xsk->xsks_map_fd = -1;
> >> -
> >> +    xsk->prog_fd = -1;
> >>      if (!(xsk->config.libbpf_flags &
> >> XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD)) {
> >>              err = xsk_setup_xdp_prog(xsk);
> >>              if (err)
> >> @@ -708,8 +659,10 @@ void xsk_socket__delete(struct xsk_socket *xsk)
> >>      if (!xsk)
> >>              return;
> >>
> >> -    xsk_clear_bpf_maps(xsk);
> >> -    xsk_delete_bpf_maps(xsk);
> >> +    if (xsk->prog_fd != -1) {
> >> +            xsk_delete_bpf_maps(xsk);
> >> +            close(xsk->prog_fd);
> >
> > Here, we use prog_fd != -1 to gate xsk_delete_bpf_maps(), which is
> > confusing. I looked at the code for quite sometime, but still cannot
> > confirm it is correct.
>
> See above reasoning - with INHIBIT_PROG_LOAD, there is no
> library-provided
> program or maps, so cleanup actions are skipped.
> --
> Jonathan

^ permalink raw reply	[flat|nested] 8+ messages in thread

end of thread, other threads:[~2019-06-03 11:11 UTC | newest]

Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-05-31 18:57 [PATCH v3 bpf-next 0/2] Better handling of xskmap entries Jonathan Lemon
2019-05-31 18:57 ` [PATCH v3 bpf-next 1/2] bpf: Allow bpf_map_lookup_elem() on an xskmap Jonathan Lemon
2019-06-01 23:06   ` Song Liu
2019-05-31 18:57 ` [PATCH v3 bpf-next 2/2] libbpf: remove qidconf and better support external bpf programs Jonathan Lemon
2019-06-01 23:05   ` Song Liu
2019-06-02  4:18     ` Jonathan Lemon
2019-06-02  6:55       ` Song Liu
2019-06-03 11:11       ` Björn Töpel

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.