From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 88482C433EF for ; Thu, 7 Jul 2022 10:48:06 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234834AbiGGKsF (ORCPT ); Thu, 7 Jul 2022 06:48:05 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:40742 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231837AbiGGKsE (ORCPT ); Thu, 7 Jul 2022 06:48:04 -0400 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id 293F84D4CC for ; Thu, 7 Jul 2022 03:48:01 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1657190880; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=T7vhJhQinGVPIg9ZMrkjjwzeezP5PnQmEm5c49UEHOQ=; b=DSskAKZQtXcXewyqMte2gZExsnLxr6dwEgtxGRS9HoApBzAoDW+nysjju0Og52EEM/tPZm gaEQGbj1q9i7QOMz5+X2fo1Bu/JNyG8wxU27RdynKAhqpOQXFHX2csYxI2YAhorDNgHg5L W3EdExXSX7PH72ChgkxZJLiDILq1cJ4= Received: from mail-pf1-f197.google.com (mail-pf1-f197.google.com [209.85.210.197]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-134-MC4UuIF1OvSo_n6CLi9zwQ-1; Thu, 07 Jul 2022 06:47:53 -0400 X-MC-Unique: MC4UuIF1OvSo_n6CLi9zwQ-1 Received: by mail-pf1-f197.google.com with SMTP id bx17-20020a056a00429100b005283c415c8cso6077615pfb.14 for ; Thu, 07 Jul 2022 03:47:53 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=T7vhJhQinGVPIg9ZMrkjjwzeezP5PnQmEm5c49UEHOQ=; b=DZ/ht0y/0TbEthrZ9VCcwIN0aLkrWvWwY6VC2XRlyQt3nFM8YWUvUcz1sQ0T0TkeqM NsXu7rjnOVYZ2FsDa6g5hHBtN0b9ZrgRODTbmFmLOYGUAGvNQfBsINMOm2eou3PCXlAX Cfj+o2hnLpSN0NXnhfd67pPkAk/csSHdCkeDvYec98fRykN47J/g/7izFptici730SNl UMWPX3Oj8l/IiBdRdc57ZRf8RXepmGEJdw6QC28LC5u6Q/nuK7Unaza1gCQgJC2YNXQk jAWtKA/HpxLevCzNeTlhnCggiUxpNqB9jFE+LjPRP6+qSCEk7gJGsT9Vq7um/kpLY+bv mBbA== X-Gm-Message-State: AJIora+7um4hba8jlsGeeDOmI96S9w6RN5Hsctj74TBwsqovr2UxP6K5 oBrcrq9Ehw4KWCL0p644cVhZH/pEhMJZxxUvUyq5PJXBqwHAdNOpgV2BbDfUFwPD7+LVNM05z4R sbwMPiwNphQXIjly1jYPqqr/9a66bc2kHkBd4AedMAm5q X-Received: by 2002:a17:903:2d1:b0:168:e83e:dab0 with SMTP id s17-20020a17090302d100b00168e83edab0mr50827132plk.118.1657190871084; Thu, 07 Jul 2022 03:47:51 -0700 (PDT) X-Google-Smtp-Source: AGRyM1tOwN59xjdWF5rHVM/momWS/uIXAjzlD3SBZ8xnYwUqOl4yVWObxHZyplTcAhOaN76vtjIfnZQcmHxpiWvxbhY= X-Received: by 2002:a17:903:2d1:b0:168:e83e:dab0 with SMTP id s17-20020a17090302d100b00168e83edab0mr50827073plk.118.1657190870449; Thu, 07 Jul 2022 03:47:50 -0700 (PDT) MIME-Version: 1.0 References: <20220615134847.3753567-1-maximmi@nvidia.com> <20220615134847.3753567-5-maximmi@nvidia.com> In-Reply-To: <20220615134847.3753567-5-maximmi@nvidia.com> From: Yauheni Kaliuta Date: Thu, 7 Jul 2022 13:47:34 +0300 Message-ID: Subject: Re: [PATCH bpf-next v10 4/6] selftests/bpf: Add selftests for raw syncookie helpers To: Maxim Mikityanskiy Cc: bpf , Alexei Starovoitov , Daniel Borkmann , Andrii Nakryiko , netdev@vger.kernel.org, Tariq Toukan , Martin KaFai Lau , Song Liu , Yonghong Song , John Fastabend , KP Singh , "David S. Miller" , Jakub Kicinski , Eric Dumazet , Hideaki YOSHIFUJI , David Ahern , Shuah Khan , Jesper Dangaard Brouer , Nathan Chancellor , Nick Desaulniers , Joe Stringer , Florent Revest , "open list:KERNEL SELFTEST FRAMEWORK" , =?UTF-8?B?VG9rZSBIw7hpbGFuZC1Kw7hyZ2Vuc2Vu?= , Kumar Kartikeya Dwivedi , Florian Westphal , Paolo Abeni Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kselftest@vger.kernel.org Hi, Maxim! The test is not buildable for me with CONFIG_NF_CONNTRACK=m due to missing BPF_F_CURRENT_NETNS in vmlinux.h. On Wed, Jun 15, 2022 at 4:49 PM Maxim Mikityanskiy wrote: > > This commit adds selftests for the new BPF helpers: > bpf_tcp_raw_{gen,check}_syncookie_ipv{4,6}. > > xdp_synproxy_kern.c is a BPF program that generates SYN cookies on > allowed TCP ports and sends SYNACKs to clients, accelerating synproxy > iptables module. > > xdp_synproxy.c is a userspace control application that allows to > configure the following options in runtime: list of allowed ports, MSS, > window scale, TTL. > > A selftest is added to prog_tests that leverages the above programs to > test the functionality of the new helpers. > > Signed-off-by: Maxim Mikityanskiy > Reviewed-by: Tariq Toukan > --- > tools/testing/selftests/bpf/.gitignore | 1 + > tools/testing/selftests/bpf/Makefile | 3 +- > .../selftests/bpf/prog_tests/xdp_synproxy.c | 146 ++++ > .../selftests/bpf/progs/xdp_synproxy_kern.c | 763 ++++++++++++++++++ > tools/testing/selftests/bpf/xdp_synproxy.c | 418 ++++++++++ > 5 files changed, 1330 insertions(+), 1 deletion(-) > create mode 100644 tools/testing/selftests/bpf/prog_tests/xdp_synproxy.c > create mode 100644 tools/testing/selftests/bpf/progs/xdp_synproxy_kern.c > create mode 100644 tools/testing/selftests/bpf/xdp_synproxy.c > > diff --git a/tools/testing/selftests/bpf/.gitignore b/tools/testing/selftests/bpf/.gitignore > index 595565eb68c0..ca2f47f45670 100644 > --- a/tools/testing/selftests/bpf/.gitignore > +++ b/tools/testing/selftests/bpf/.gitignore > @@ -43,3 +43,4 @@ test_cpp > *.tmp > xdpxceiver > xdp_redirect_multi > +xdp_synproxy > diff --git a/tools/testing/selftests/bpf/Makefile b/tools/testing/selftests/bpf/Makefile > index 8ad7a733a505..e4ce21f4a474 100644 > --- a/tools/testing/selftests/bpf/Makefile > +++ b/tools/testing/selftests/bpf/Makefile > @@ -82,7 +82,7 @@ TEST_PROGS_EXTENDED := with_addr.sh \ > TEST_GEN_PROGS_EXTENDED = test_sock_addr test_skb_cgroup_id_user \ > flow_dissector_load test_flow_dissector test_tcp_check_syncookie_user \ > test_lirc_mode2_user xdping test_cpp runqslower bench bpf_testmod.ko \ > - xdpxceiver xdp_redirect_multi > + xdpxceiver xdp_redirect_multi xdp_synproxy > > TEST_CUSTOM_PROGS = $(OUTPUT)/urandom_read > > @@ -502,6 +502,7 @@ TRUNNER_EXTRA_SOURCES := test_progs.c cgroup_helpers.c trace_helpers.c \ > cap_helpers.c > TRUNNER_EXTRA_FILES := $(OUTPUT)/urandom_read $(OUTPUT)/bpf_testmod.ko \ > $(OUTPUT)/liburandom_read.so \ > + $(OUTPUT)/xdp_synproxy \ > ima_setup.sh \ > $(wildcard progs/btf_dump_test_case_*.c) > TRUNNER_BPF_BUILD_RULE := CLANG_BPF_BUILD_RULE > diff --git a/tools/testing/selftests/bpf/prog_tests/xdp_synproxy.c b/tools/testing/selftests/bpf/prog_tests/xdp_synproxy.c > new file mode 100644 > index 000000000000..d9ee884c2a2b > --- /dev/null > +++ b/tools/testing/selftests/bpf/prog_tests/xdp_synproxy.c > @@ -0,0 +1,146 @@ > +// SPDX-License-Identifier: LGPL-2.1 OR BSD-2-Clause > +/* Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES. All rights reserved. */ > + > +#include > +#include > +#include > + > +#define CMD_OUT_BUF_SIZE 1023 > + > +#define SYS(cmd) ({ \ > + if (!ASSERT_OK(system(cmd), (cmd))) \ > + goto out; \ > +}) > + > +#define SYS_OUT(cmd) ({ \ > + FILE *f = popen((cmd), "r"); \ > + if (!ASSERT_OK_PTR(f, (cmd))) \ > + goto out; \ > + f; \ > +}) > + > +/* out must be at least `size * 4 + 1` bytes long */ > +static void escape_str(char *out, const char *in, size_t size) > +{ > + static const char *hex = "0123456789ABCDEF"; > + size_t i; > + > + for (i = 0; i < size; i++) { > + if (isprint(in[i]) && in[i] != '\\' && in[i] != '\'') { > + *out++ = in[i]; > + } else { > + *out++ = '\\'; > + *out++ = 'x'; > + *out++ = hex[(in[i] >> 4) & 0xf]; > + *out++ = hex[in[i] & 0xf]; > + } > + } > + *out++ = '\0'; > +} > + > +static bool expect_str(char *buf, size_t size, const char *str, const char *name) > +{ > + static char escbuf_expected[CMD_OUT_BUF_SIZE * 4]; > + static char escbuf_actual[CMD_OUT_BUF_SIZE * 4]; > + static int duration = 0; > + bool ok; > + > + ok = size == strlen(str) && !memcmp(buf, str, size); > + > + if (!ok) { > + escape_str(escbuf_expected, str, strlen(str)); > + escape_str(escbuf_actual, buf, size); > + } > + CHECK(!ok, name, "unexpected %s: actual '%s' != expected '%s'\n", > + name, escbuf_actual, escbuf_expected); > + > + return ok; > +} > + > +void test_xdp_synproxy(void) > +{ > + int server_fd = -1, client_fd = -1, accept_fd = -1; > + struct nstoken *ns = NULL; > + FILE *ctrl_file = NULL; > + char buf[CMD_OUT_BUF_SIZE]; > + size_t size; > + > + SYS("ip netns add synproxy"); > + > + SYS("ip link add tmp0 type veth peer name tmp1"); > + SYS("ip link set tmp1 netns synproxy"); > + SYS("ip link set tmp0 up"); > + SYS("ip addr replace 198.18.0.1/24 dev tmp0"); > + > + /* When checksum offload is enabled, the XDP program sees wrong > + * checksums and drops packets. > + */ > + SYS("ethtool -K tmp0 tx off"); > + /* Workaround required for veth. */ > + SYS("ip link set tmp0 xdp object xdp_dummy.o section xdp 2> /dev/null"); > + > + ns = open_netns("synproxy"); > + if (!ASSERT_OK_PTR(ns, "setns")) > + goto out; > + > + SYS("ip link set lo up"); > + SYS("ip link set tmp1 up"); > + SYS("ip addr replace 198.18.0.2/24 dev tmp1"); > + SYS("sysctl -w net.ipv4.tcp_syncookies=2"); > + SYS("sysctl -w net.ipv4.tcp_timestamps=1"); > + SYS("sysctl -w net.netfilter.nf_conntrack_tcp_loose=0"); > + SYS("iptables -t raw -I PREROUTING \ > + -i tmp1 -p tcp -m tcp --syn --dport 8080 -j CT --notrack"); > + SYS("iptables -t filter -A INPUT \ > + -i tmp1 -p tcp -m tcp --dport 8080 -m state --state INVALID,UNTRACKED \ > + -j SYNPROXY --sack-perm --timestamp --wscale 7 --mss 1460"); > + SYS("iptables -t filter -A INPUT \ > + -i tmp1 -m state --state INVALID -j DROP"); > + > + ctrl_file = SYS_OUT("./xdp_synproxy --iface tmp1 --ports 8080 --single \ > + --mss4 1460 --mss6 1440 --wscale 7 --ttl 64"); > + size = fread(buf, 1, sizeof(buf), ctrl_file); > + pclose(ctrl_file); > + if (!expect_str(buf, size, "Total SYNACKs generated: 0\n", > + "initial SYNACKs")) > + goto out; > + > + server_fd = start_server(AF_INET, SOCK_STREAM, "198.18.0.2", 8080, 0); > + if (!ASSERT_GE(server_fd, 0, "start_server")) > + goto out; > + > + close_netns(ns); > + ns = NULL; > + > + client_fd = connect_to_fd(server_fd, 10000); > + if (!ASSERT_GE(client_fd, 0, "connect_to_fd")) > + goto out; > + > + accept_fd = accept(server_fd, NULL, NULL); > + if (!ASSERT_GE(accept_fd, 0, "accept")) > + goto out; > + > + ns = open_netns("synproxy"); > + if (!ASSERT_OK_PTR(ns, "setns")) > + goto out; > + > + ctrl_file = SYS_OUT("./xdp_synproxy --iface tmp1 --single"); > + size = fread(buf, 1, sizeof(buf), ctrl_file); > + pclose(ctrl_file); > + if (!expect_str(buf, size, "Total SYNACKs generated: 1\n", > + "SYNACKs after connection")) > + goto out; > + > +out: > + if (accept_fd >= 0) > + close(accept_fd); > + if (client_fd >= 0) > + close(client_fd); > + if (server_fd >= 0) > + close(server_fd); > + if (ns) > + close_netns(ns); > + > + system("ip link del tmp0"); > + system("ip netns del synproxy"); > +} > diff --git a/tools/testing/selftests/bpf/progs/xdp_synproxy_kern.c b/tools/testing/selftests/bpf/progs/xdp_synproxy_kern.c > new file mode 100644 > index 000000000000..53b9865276a4 > --- /dev/null > +++ b/tools/testing/selftests/bpf/progs/xdp_synproxy_kern.c > @@ -0,0 +1,763 @@ > +// SPDX-License-Identifier: LGPL-2.1 OR BSD-2-Clause > +/* Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES. All rights reserved. */ > + > +#include "vmlinux.h" > + > +#include > +#include > +#include > + > +#define NSEC_PER_SEC 1000000000L > + > +#define ETH_ALEN 6 > +#define ETH_P_IP 0x0800 > +#define ETH_P_IPV6 0x86DD > + > +#define tcp_flag_word(tp) (((union tcp_word_hdr *)(tp))->words[3]) > + > +#define IP_DF 0x4000 > +#define IP_MF 0x2000 > +#define IP_OFFSET 0x1fff > + > +#define NEXTHDR_TCP 6 > + > +#define TCPOPT_NOP 1 > +#define TCPOPT_EOL 0 > +#define TCPOPT_MSS 2 > +#define TCPOPT_WINDOW 3 > +#define TCPOPT_SACK_PERM 4 > +#define TCPOPT_TIMESTAMP 8 > + > +#define TCPOLEN_MSS 4 > +#define TCPOLEN_WINDOW 3 > +#define TCPOLEN_SACK_PERM 2 > +#define TCPOLEN_TIMESTAMP 10 > + > +#define TCP_TS_HZ 1000 > +#define TS_OPT_WSCALE_MASK 0xf > +#define TS_OPT_SACK (1 << 4) > +#define TS_OPT_ECN (1 << 5) > +#define TSBITS 6 > +#define TSMASK (((__u32)1 << TSBITS) - 1) > +#define TCP_MAX_WSCALE 14U > + > +#define IPV4_MAXLEN 60 > +#define TCP_MAXLEN 60 > + > +#define DEFAULT_MSS4 1460 > +#define DEFAULT_MSS6 1440 > +#define DEFAULT_WSCALE 7 > +#define DEFAULT_TTL 64 > +#define MAX_ALLOWED_PORTS 8 > + > +#define swap(a, b) \ > + do { typeof(a) __tmp = (a); (a) = (b); (b) = __tmp; } while (0) > + > +#define __get_unaligned_t(type, ptr) ({ \ > + const struct { type x; } __attribute__((__packed__)) *__pptr = (typeof(__pptr))(ptr); \ > + __pptr->x; \ > +}) > + > +#define get_unaligned(ptr) __get_unaligned_t(typeof(*(ptr)), (ptr)) > + > +struct { > + __uint(type, BPF_MAP_TYPE_ARRAY); > + __type(key, __u32); > + __type(value, __u64); > + __uint(max_entries, 2); > +} values SEC(".maps"); > + > +struct { > + __uint(type, BPF_MAP_TYPE_ARRAY); > + __type(key, __u32); > + __type(value, __u16); > + __uint(max_entries, MAX_ALLOWED_PORTS); > +} allowed_ports SEC(".maps"); > + > +extern struct nf_conn *bpf_xdp_ct_lookup(struct xdp_md *xdp_ctx, > + struct bpf_sock_tuple *bpf_tuple, > + __u32 len_tuple, > + struct bpf_ct_opts *opts, > + __u32 len_opts) __ksym; > + > +extern void bpf_ct_release(struct nf_conn *ct) __ksym; > + > +static __always_inline void swap_eth_addr(__u8 *a, __u8 *b) > +{ > + __u8 tmp[ETH_ALEN]; > + > + __builtin_memcpy(tmp, a, ETH_ALEN); > + __builtin_memcpy(a, b, ETH_ALEN); > + __builtin_memcpy(b, tmp, ETH_ALEN); > +} > + > +static __always_inline __u16 csum_fold(__u32 csum) > +{ > + csum = (csum & 0xffff) + (csum >> 16); > + csum = (csum & 0xffff) + (csum >> 16); > + return (__u16)~csum; > +} > + > +static __always_inline __u16 csum_tcpudp_magic(__be32 saddr, __be32 daddr, > + __u32 len, __u8 proto, > + __u32 csum) > +{ > + __u64 s = csum; > + > + s += (__u32)saddr; > + s += (__u32)daddr; > +#if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ > + s += proto + len; > +#elif __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ > + s += (proto + len) << 8; > +#else > +#error Unknown endian > +#endif > + s = (s & 0xffffffff) + (s >> 32); > + s = (s & 0xffffffff) + (s >> 32); > + > + return csum_fold((__u32)s); > +} > + > +static __always_inline __u16 csum_ipv6_magic(const struct in6_addr *saddr, > + const struct in6_addr *daddr, > + __u32 len, __u8 proto, __u32 csum) > +{ > + __u64 sum = csum; > + int i; > + > +#pragma unroll > + for (i = 0; i < 4; i++) > + sum += (__u32)saddr->in6_u.u6_addr32[i]; > + > +#pragma unroll > + for (i = 0; i < 4; i++) > + sum += (__u32)daddr->in6_u.u6_addr32[i]; > + > + /* Don't combine additions to avoid 32-bit overflow. */ > + sum += bpf_htonl(len); > + sum += bpf_htonl(proto); > + > + sum = (sum & 0xffffffff) + (sum >> 32); > + sum = (sum & 0xffffffff) + (sum >> 32); > + > + return csum_fold((__u32)sum); > +} > + > +static __always_inline __u64 tcp_clock_ns(void) > +{ > + return bpf_ktime_get_ns(); > +} > + > +static __always_inline __u32 tcp_ns_to_ts(__u64 ns) > +{ > + return ns / (NSEC_PER_SEC / TCP_TS_HZ); > +} > + > +static __always_inline __u32 tcp_time_stamp_raw(void) > +{ > + return tcp_ns_to_ts(tcp_clock_ns()); > +} > + > +struct tcpopt_context { > + __u8 *ptr; > + __u8 *end; > + void *data_end; > + __be32 *tsecr; > + __u8 wscale; > + bool option_timestamp; > + bool option_sack; > +}; > + > +static int tscookie_tcpopt_parse(struct tcpopt_context *ctx) > +{ > + __u8 opcode, opsize; > + > + if (ctx->ptr >= ctx->end) > + return 1; > + if (ctx->ptr >= ctx->data_end) > + return 1; > + > + opcode = ctx->ptr[0]; > + > + if (opcode == TCPOPT_EOL) > + return 1; > + if (opcode == TCPOPT_NOP) { > + ++ctx->ptr; > + return 0; > + } > + > + if (ctx->ptr + 1 >= ctx->end) > + return 1; > + if (ctx->ptr + 1 >= ctx->data_end) > + return 1; > + opsize = ctx->ptr[1]; > + if (opsize < 2) > + return 1; > + > + if (ctx->ptr + opsize > ctx->end) > + return 1; > + > + switch (opcode) { > + case TCPOPT_WINDOW: > + if (opsize == TCPOLEN_WINDOW && ctx->ptr + TCPOLEN_WINDOW <= ctx->data_end) > + ctx->wscale = ctx->ptr[2] < TCP_MAX_WSCALE ? ctx->ptr[2] : TCP_MAX_WSCALE; > + break; > + case TCPOPT_TIMESTAMP: > + if (opsize == TCPOLEN_TIMESTAMP && ctx->ptr + TCPOLEN_TIMESTAMP <= ctx->data_end) { > + ctx->option_timestamp = true; > + /* Client's tsval becomes our tsecr. */ > + *ctx->tsecr = get_unaligned((__be32 *)(ctx->ptr + 2)); > + } > + break; > + case TCPOPT_SACK_PERM: > + if (opsize == TCPOLEN_SACK_PERM) > + ctx->option_sack = true; > + break; > + } > + > + ctx->ptr += opsize; > + > + return 0; > +} > + > +static int tscookie_tcpopt_parse_batch(__u32 index, void *context) > +{ > + int i; > + > + for (i = 0; i < 7; i++) > + if (tscookie_tcpopt_parse(context)) > + return 1; > + return 0; > +} > + > +static __always_inline bool tscookie_init(struct tcphdr *tcp_header, > + __u16 tcp_len, __be32 *tsval, > + __be32 *tsecr, void *data_end) > +{ > + struct tcpopt_context loop_ctx = { > + .ptr = (__u8 *)(tcp_header + 1), > + .end = (__u8 *)tcp_header + tcp_len, > + .data_end = data_end, > + .tsecr = tsecr, > + .wscale = TS_OPT_WSCALE_MASK, > + .option_timestamp = false, > + .option_sack = false, > + }; > + u32 cookie; > + > + bpf_loop(6, tscookie_tcpopt_parse_batch, &loop_ctx, 0); > + > + if (!loop_ctx.option_timestamp) > + return false; > + > + cookie = tcp_time_stamp_raw() & ~TSMASK; > + cookie |= loop_ctx.wscale & TS_OPT_WSCALE_MASK; > + if (loop_ctx.option_sack) > + cookie |= TS_OPT_SACK; > + if (tcp_header->ece && tcp_header->cwr) > + cookie |= TS_OPT_ECN; > + *tsval = bpf_htonl(cookie); > + > + return true; > +} > + > +static __always_inline void values_get_tcpipopts(__u16 *mss, __u8 *wscale, > + __u8 *ttl, bool ipv6) > +{ > + __u32 key = 0; > + __u64 *value; > + > + value = bpf_map_lookup_elem(&values, &key); > + if (value && *value != 0) { > + if (ipv6) > + *mss = (*value >> 32) & 0xffff; > + else > + *mss = *value & 0xffff; > + *wscale = (*value >> 16) & 0xf; > + *ttl = (*value >> 24) & 0xff; > + return; > + } > + > + *mss = ipv6 ? DEFAULT_MSS6 : DEFAULT_MSS4; > + *wscale = DEFAULT_WSCALE; > + *ttl = DEFAULT_TTL; > +} > + > +static __always_inline void values_inc_synacks(void) > +{ > + __u32 key = 1; > + __u32 *value; > + > + value = bpf_map_lookup_elem(&values, &key); > + if (value) > + __sync_fetch_and_add(value, 1); > +} > + > +static __always_inline bool check_port_allowed(__u16 port) > +{ > + __u32 i; > + > + for (i = 0; i < MAX_ALLOWED_PORTS; i++) { > + __u32 key = i; > + __u16 *value; > + > + value = bpf_map_lookup_elem(&allowed_ports, &key); > + > + if (!value) > + break; > + /* 0 is a terminator value. Check it first to avoid matching on > + * a forbidden port == 0 and returning true. > + */ > + if (*value == 0) > + break; > + > + if (*value == port) > + return true; > + } > + > + return false; > +} > + > +struct header_pointers { > + struct ethhdr *eth; > + struct iphdr *ipv4; > + struct ipv6hdr *ipv6; > + struct tcphdr *tcp; > + __u16 tcp_len; > +}; > + > +static __always_inline int tcp_dissect(void *data, void *data_end, > + struct header_pointers *hdr) > +{ > + hdr->eth = data; > + if (hdr->eth + 1 > data_end) > + return XDP_DROP; > + > + switch (bpf_ntohs(hdr->eth->h_proto)) { > + case ETH_P_IP: > + hdr->ipv6 = NULL; > + > + hdr->ipv4 = (void *)hdr->eth + sizeof(*hdr->eth); > + if (hdr->ipv4 + 1 > data_end) > + return XDP_DROP; > + if (hdr->ipv4->ihl * 4 < sizeof(*hdr->ipv4)) > + return XDP_DROP; > + if (hdr->ipv4->version != 4) > + return XDP_DROP; > + > + if (hdr->ipv4->protocol != IPPROTO_TCP) > + return XDP_PASS; > + > + hdr->tcp = (void *)hdr->ipv4 + hdr->ipv4->ihl * 4; > + break; > + case ETH_P_IPV6: > + hdr->ipv4 = NULL; > + > + hdr->ipv6 = (void *)hdr->eth + sizeof(*hdr->eth); > + if (hdr->ipv6 + 1 > data_end) > + return XDP_DROP; > + if (hdr->ipv6->version != 6) > + return XDP_DROP; > + > + /* XXX: Extension headers are not supported and could circumvent > + * XDP SYN flood protection. > + */ > + if (hdr->ipv6->nexthdr != NEXTHDR_TCP) > + return XDP_PASS; > + > + hdr->tcp = (void *)hdr->ipv6 + sizeof(*hdr->ipv6); > + break; > + default: > + /* XXX: VLANs will circumvent XDP SYN flood protection. */ > + return XDP_PASS; > + } > + > + if (hdr->tcp + 1 > data_end) > + return XDP_DROP; > + hdr->tcp_len = hdr->tcp->doff * 4; > + if (hdr->tcp_len < sizeof(*hdr->tcp)) > + return XDP_DROP; > + > + return XDP_TX; > +} > + > +static __always_inline int tcp_lookup(struct xdp_md *ctx, struct header_pointers *hdr) > +{ > + struct bpf_ct_opts ct_lookup_opts = { > + .netns_id = BPF_F_CURRENT_NETNS, > + .l4proto = IPPROTO_TCP, > + }; > + struct bpf_sock_tuple tup = {}; > + struct nf_conn *ct; > + __u32 tup_size; > + > + if (hdr->ipv4) { > + /* TCP doesn't normally use fragments, and XDP can't reassemble > + * them. > + */ > + if ((hdr->ipv4->frag_off & bpf_htons(IP_DF | IP_MF | IP_OFFSET)) != bpf_htons(IP_DF)) > + return XDP_DROP; > + > + tup.ipv4.saddr = hdr->ipv4->saddr; > + tup.ipv4.daddr = hdr->ipv4->daddr; > + tup.ipv4.sport = hdr->tcp->source; > + tup.ipv4.dport = hdr->tcp->dest; > + tup_size = sizeof(tup.ipv4); > + } else if (hdr->ipv6) { > + __builtin_memcpy(tup.ipv6.saddr, &hdr->ipv6->saddr, sizeof(tup.ipv6.saddr)); > + __builtin_memcpy(tup.ipv6.daddr, &hdr->ipv6->daddr, sizeof(tup.ipv6.daddr)); > + tup.ipv6.sport = hdr->tcp->source; > + tup.ipv6.dport = hdr->tcp->dest; > + tup_size = sizeof(tup.ipv6); > + } else { > + /* The verifier can't track that either ipv4 or ipv6 is not > + * NULL. > + */ > + return XDP_ABORTED; > + } > + ct = bpf_xdp_ct_lookup(ctx, &tup, tup_size, &ct_lookup_opts, sizeof(ct_lookup_opts)); > + if (ct) { > + unsigned long status = ct->status; > + > + bpf_ct_release(ct); > + if (status & IPS_CONFIRMED_BIT) > + return XDP_PASS; > + } else if (ct_lookup_opts.error != -ENOENT) { > + return XDP_ABORTED; > + } > + > + /* error == -ENOENT || !(status & IPS_CONFIRMED_BIT) */ > + return XDP_TX; > +} > + > +static __always_inline __u8 tcp_mkoptions(__be32 *buf, __be32 *tsopt, __u16 mss, > + __u8 wscale) > +{ > + __be32 *start = buf; > + > + *buf++ = bpf_htonl((TCPOPT_MSS << 24) | (TCPOLEN_MSS << 16) | mss); > + > + if (!tsopt) > + return buf - start; > + > + if (tsopt[0] & bpf_htonl(1 << 4)) > + *buf++ = bpf_htonl((TCPOPT_SACK_PERM << 24) | > + (TCPOLEN_SACK_PERM << 16) | > + (TCPOPT_TIMESTAMP << 8) | > + TCPOLEN_TIMESTAMP); > + else > + *buf++ = bpf_htonl((TCPOPT_NOP << 24) | > + (TCPOPT_NOP << 16) | > + (TCPOPT_TIMESTAMP << 8) | > + TCPOLEN_TIMESTAMP); > + *buf++ = tsopt[0]; > + *buf++ = tsopt[1]; > + > + if ((tsopt[0] & bpf_htonl(0xf)) != bpf_htonl(0xf)) > + *buf++ = bpf_htonl((TCPOPT_NOP << 24) | > + (TCPOPT_WINDOW << 16) | > + (TCPOLEN_WINDOW << 8) | > + wscale); > + > + return buf - start; > +} > + > +static __always_inline void tcp_gen_synack(struct tcphdr *tcp_header, > + __u32 cookie, __be32 *tsopt, > + __u16 mss, __u8 wscale) > +{ > + void *tcp_options; > + > + tcp_flag_word(tcp_header) = TCP_FLAG_SYN | TCP_FLAG_ACK; > + if (tsopt && (tsopt[0] & bpf_htonl(1 << 5))) > + tcp_flag_word(tcp_header) |= TCP_FLAG_ECE; > + tcp_header->doff = 5; /* doff is part of tcp_flag_word. */ > + swap(tcp_header->source, tcp_header->dest); > + tcp_header->ack_seq = bpf_htonl(bpf_ntohl(tcp_header->seq) + 1); > + tcp_header->seq = bpf_htonl(cookie); > + tcp_header->window = 0; > + tcp_header->urg_ptr = 0; > + tcp_header->check = 0; /* Calculate checksum later. */ > + > + tcp_options = (void *)(tcp_header + 1); > + tcp_header->doff += tcp_mkoptions(tcp_options, tsopt, mss, wscale); > +} > + > +static __always_inline void tcpv4_gen_synack(struct header_pointers *hdr, > + __u32 cookie, __be32 *tsopt) > +{ > + __u8 wscale; > + __u16 mss; > + __u8 ttl; > + > + values_get_tcpipopts(&mss, &wscale, &ttl, false); > + > + swap_eth_addr(hdr->eth->h_source, hdr->eth->h_dest); > + > + swap(hdr->ipv4->saddr, hdr->ipv4->daddr); > + hdr->ipv4->check = 0; /* Calculate checksum later. */ > + hdr->ipv4->tos = 0; > + hdr->ipv4->id = 0; > + hdr->ipv4->ttl = ttl; > + > + tcp_gen_synack(hdr->tcp, cookie, tsopt, mss, wscale); > + > + hdr->tcp_len = hdr->tcp->doff * 4; > + hdr->ipv4->tot_len = bpf_htons(sizeof(*hdr->ipv4) + hdr->tcp_len); > +} > + > +static __always_inline void tcpv6_gen_synack(struct header_pointers *hdr, > + __u32 cookie, __be32 *tsopt) > +{ > + __u8 wscale; > + __u16 mss; > + __u8 ttl; > + > + values_get_tcpipopts(&mss, &wscale, &ttl, true); > + > + swap_eth_addr(hdr->eth->h_source, hdr->eth->h_dest); > + > + swap(hdr->ipv6->saddr, hdr->ipv6->daddr); > + *(__be32 *)hdr->ipv6 = bpf_htonl(0x60000000); > + hdr->ipv6->hop_limit = ttl; > + > + tcp_gen_synack(hdr->tcp, cookie, tsopt, mss, wscale); > + > + hdr->tcp_len = hdr->tcp->doff * 4; > + hdr->ipv6->payload_len = bpf_htons(hdr->tcp_len); > +} > + > +static __always_inline int syncookie_handle_syn(struct header_pointers *hdr, > + struct xdp_md *ctx, > + void *data, void *data_end) > +{ > + __u32 old_pkt_size, new_pkt_size; > + /* Unlike clang 10, clang 11 and 12 generate code that doesn't pass the > + * BPF verifier if tsopt is not volatile. Volatile forces it to store > + * the pointer value and use it directly, otherwise tcp_mkoptions is > + * (mis)compiled like this: > + * if (!tsopt) > + * return buf - start; > + * reg = stored_return_value_of_tscookie_init; > + * if (reg) > + * tsopt = tsopt_buf; > + * else > + * tsopt = NULL; > + * ... > + * *buf++ = tsopt[1]; > + * It creates a dead branch where tsopt is assigned NULL, but the > + * verifier can't prove it's dead and blocks the program. > + */ > + __be32 * volatile tsopt = NULL; > + __be32 tsopt_buf[2] = {}; > + __u16 ip_len; > + __u32 cookie; > + __s64 value; > + > + /* Checksum is not yet verified, but both checksum failure and TCP > + * header checks return XDP_DROP, so the order doesn't matter. > + */ > + if (hdr->tcp->fin || hdr->tcp->rst) > + return XDP_DROP; > + > + /* Issue SYN cookies on allowed ports, drop SYN packets on blocked > + * ports. > + */ > + if (!check_port_allowed(bpf_ntohs(hdr->tcp->dest))) > + return XDP_DROP; > + > + if (hdr->ipv4) { > + /* Check the IPv4 and TCP checksums before creating a SYNACK. */ > + value = bpf_csum_diff(0, 0, (void *)hdr->ipv4, hdr->ipv4->ihl * 4, 0); > + if (value < 0) > + return XDP_ABORTED; > + if (csum_fold(value) != 0) > + return XDP_DROP; /* Bad IPv4 checksum. */ > + > + value = bpf_csum_diff(0, 0, (void *)hdr->tcp, hdr->tcp_len, 0); > + if (value < 0) > + return XDP_ABORTED; > + if (csum_tcpudp_magic(hdr->ipv4->saddr, hdr->ipv4->daddr, > + hdr->tcp_len, IPPROTO_TCP, value) != 0) > + return XDP_DROP; /* Bad TCP checksum. */ > + > + ip_len = sizeof(*hdr->ipv4); > + > + value = bpf_tcp_raw_gen_syncookie_ipv4(hdr->ipv4, hdr->tcp, > + hdr->tcp_len); > + } else if (hdr->ipv6) { > + /* Check the TCP checksum before creating a SYNACK. */ > + value = bpf_csum_diff(0, 0, (void *)hdr->tcp, hdr->tcp_len, 0); > + if (value < 0) > + return XDP_ABORTED; > + if (csum_ipv6_magic(&hdr->ipv6->saddr, &hdr->ipv6->daddr, > + hdr->tcp_len, IPPROTO_TCP, value) != 0) > + return XDP_DROP; /* Bad TCP checksum. */ > + > + ip_len = sizeof(*hdr->ipv6); > + > + value = bpf_tcp_raw_gen_syncookie_ipv6(hdr->ipv6, hdr->tcp, > + hdr->tcp_len); > + } else { > + return XDP_ABORTED; > + } > + > + if (value < 0) > + return XDP_ABORTED; > + cookie = (__u32)value; > + > + if (tscookie_init((void *)hdr->tcp, hdr->tcp_len, > + &tsopt_buf[0], &tsopt_buf[1], data_end)) > + tsopt = tsopt_buf; > + > + /* Check that there is enough space for a SYNACK. It also covers > + * the check that the destination of the __builtin_memmove below > + * doesn't overflow. > + */ > + if (data + sizeof(*hdr->eth) + ip_len + TCP_MAXLEN > data_end) > + return XDP_ABORTED; > + > + if (hdr->ipv4) { > + if (hdr->ipv4->ihl * 4 > sizeof(*hdr->ipv4)) { > + struct tcphdr *new_tcp_header; > + > + new_tcp_header = data + sizeof(*hdr->eth) + sizeof(*hdr->ipv4); > + __builtin_memmove(new_tcp_header, hdr->tcp, sizeof(*hdr->tcp)); > + hdr->tcp = new_tcp_header; > + > + hdr->ipv4->ihl = sizeof(*hdr->ipv4) / 4; > + } > + > + tcpv4_gen_synack(hdr, cookie, tsopt); > + } else if (hdr->ipv6) { > + tcpv6_gen_synack(hdr, cookie, tsopt); > + } else { > + return XDP_ABORTED; > + } > + > + /* Recalculate checksums. */ > + hdr->tcp->check = 0; > + value = bpf_csum_diff(0, 0, (void *)hdr->tcp, hdr->tcp_len, 0); > + if (value < 0) > + return XDP_ABORTED; > + if (hdr->ipv4) { > + hdr->tcp->check = csum_tcpudp_magic(hdr->ipv4->saddr, > + hdr->ipv4->daddr, > + hdr->tcp_len, > + IPPROTO_TCP, > + value); > + > + hdr->ipv4->check = 0; > + value = bpf_csum_diff(0, 0, (void *)hdr->ipv4, sizeof(*hdr->ipv4), 0); > + if (value < 0) > + return XDP_ABORTED; > + hdr->ipv4->check = csum_fold(value); > + } else if (hdr->ipv6) { > + hdr->tcp->check = csum_ipv6_magic(&hdr->ipv6->saddr, > + &hdr->ipv6->daddr, > + hdr->tcp_len, > + IPPROTO_TCP, > + value); > + } else { > + return XDP_ABORTED; > + } > + > + /* Set the new packet size. */ > + old_pkt_size = data_end - data; > + new_pkt_size = sizeof(*hdr->eth) + ip_len + hdr->tcp->doff * 4; > + if (bpf_xdp_adjust_tail(ctx, new_pkt_size - old_pkt_size)) > + return XDP_ABORTED; > + > + values_inc_synacks(); > + > + return XDP_TX; > +} > + > +static __always_inline int syncookie_handle_ack(struct header_pointers *hdr) > +{ > + int err; > + > + if (hdr->tcp->rst) > + return XDP_DROP; > + > + if (hdr->ipv4) > + err = bpf_tcp_raw_check_syncookie_ipv4(hdr->ipv4, hdr->tcp); > + else if (hdr->ipv6) > + err = bpf_tcp_raw_check_syncookie_ipv6(hdr->ipv6, hdr->tcp); > + else > + return XDP_ABORTED; > + if (err) > + return XDP_DROP; > + > + return XDP_PASS; > +} > + > +SEC("xdp") > +int syncookie_xdp(struct xdp_md *ctx) > +{ > + void *data_end = (void *)(long)ctx->data_end; > + void *data = (void *)(long)ctx->data; > + struct header_pointers hdr; > + __s64 value; > + int ret; > + > + struct bpf_ct_opts ct_lookup_opts = { > + .netns_id = BPF_F_CURRENT_NETNS, > + .l4proto = IPPROTO_TCP, > + }; > + > + ret = tcp_dissect(data, data_end, &hdr); > + if (ret != XDP_TX) > + return ret; > + > + ret = tcp_lookup(ctx, &hdr); > + if (ret != XDP_TX) > + return ret; > + > + /* Packet is TCP and doesn't belong to an established connection. */ > + > + if ((hdr.tcp->syn ^ hdr.tcp->ack) != 1) > + return XDP_DROP; > + > + /* Grow the TCP header to TCP_MAXLEN to be able to pass any hdr.tcp_len > + * to bpf_tcp_raw_gen_syncookie_ipv{4,6} and pass the verifier. > + */ > + if (bpf_xdp_adjust_tail(ctx, TCP_MAXLEN - hdr.tcp_len)) > + return XDP_ABORTED; > + > + data_end = (void *)(long)ctx->data_end; > + data = (void *)(long)ctx->data; > + > + if (hdr.ipv4) { > + hdr.eth = data; > + hdr.ipv4 = (void *)hdr.eth + sizeof(*hdr.eth); > + /* IPV4_MAXLEN is needed when calculating checksum. > + * At least sizeof(struct iphdr) is needed here to access ihl. > + */ > + if ((void *)hdr.ipv4 + IPV4_MAXLEN > data_end) > + return XDP_ABORTED; > + hdr.tcp = (void *)hdr.ipv4 + hdr.ipv4->ihl * 4; > + } else if (hdr.ipv6) { > + hdr.eth = data; > + hdr.ipv6 = (void *)hdr.eth + sizeof(*hdr.eth); > + hdr.tcp = (void *)hdr.ipv6 + sizeof(*hdr.ipv6); > + } else { > + return XDP_ABORTED; > + } > + > + if ((void *)hdr.tcp + TCP_MAXLEN > data_end) > + return XDP_ABORTED; > + > + /* We run out of registers, tcp_len gets spilled to the stack, and the > + * verifier forgets its min and max values checked above in tcp_dissect. > + */ > + hdr.tcp_len = hdr.tcp->doff * 4; > + if (hdr.tcp_len < sizeof(*hdr.tcp)) > + return XDP_ABORTED; > + > + return hdr.tcp->syn ? syncookie_handle_syn(&hdr, ctx, data, data_end) : > + syncookie_handle_ack(&hdr); > +} > + > +char _license[] SEC("license") = "GPL"; > diff --git a/tools/testing/selftests/bpf/xdp_synproxy.c b/tools/testing/selftests/bpf/xdp_synproxy.c > new file mode 100644 > index 000000000000..4653d4655b5f > --- /dev/null > +++ b/tools/testing/selftests/bpf/xdp_synproxy.c > @@ -0,0 +1,418 @@ > +// SPDX-License-Identifier: LGPL-2.1 OR BSD-2-Clause > +/* Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES. All rights reserved. */ > + > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > +#include > + > +static unsigned int ifindex; > +static __u32 attached_prog_id; > + > +static void noreturn cleanup(int sig) > +{ > + DECLARE_LIBBPF_OPTS(bpf_xdp_attach_opts, opts); > + int prog_fd; > + int err; > + > + if (attached_prog_id == 0) > + exit(0); > + > + prog_fd = bpf_prog_get_fd_by_id(attached_prog_id); > + if (prog_fd < 0) { > + fprintf(stderr, "Error: bpf_prog_get_fd_by_id: %s\n", strerror(-prog_fd)); > + err = bpf_xdp_attach(ifindex, -1, 0, NULL); > + if (err < 0) { > + fprintf(stderr, "Error: bpf_set_link_xdp_fd: %s\n", strerror(-err)); > + fprintf(stderr, "Failed to detach XDP program\n"); > + exit(1); > + } > + } else { > + opts.old_prog_fd = prog_fd; > + err = bpf_xdp_attach(ifindex, -1, XDP_FLAGS_REPLACE, &opts); > + close(prog_fd); > + if (err < 0) { > + fprintf(stderr, "Error: bpf_set_link_xdp_fd_opts: %s\n", strerror(-err)); > + /* Not an error if already replaced by someone else. */ > + if (err != -EEXIST) { > + fprintf(stderr, "Failed to detach XDP program\n"); > + exit(1); > + } > + } > + } > + exit(0); > +} > + > +static noreturn void usage(const char *progname) > +{ > + fprintf(stderr, "Usage: %s [--iface |--prog ] [--mss4 --mss6 --wscale --ttl ] [--ports ,,...] [--single]\n", > + progname); > + exit(1); > +} > + > +static unsigned long parse_arg_ul(const char *progname, const char *arg, unsigned long limit) > +{ > + unsigned long res; > + char *endptr; > + > + errno = 0; > + res = strtoul(arg, &endptr, 10); > + if (errno != 0 || *endptr != '\0' || arg[0] == '\0' || res > limit) > + usage(progname); > + > + return res; > +} > + > +static void parse_options(int argc, char *argv[], unsigned int *ifindex, __u32 *prog_id, > + __u64 *tcpipopts, char **ports, bool *single) > +{ > + static struct option long_options[] = { > + { "help", no_argument, NULL, 'h' }, > + { "iface", required_argument, NULL, 'i' }, > + { "prog", required_argument, NULL, 'x' }, > + { "mss4", required_argument, NULL, 4 }, > + { "mss6", required_argument, NULL, 6 }, > + { "wscale", required_argument, NULL, 'w' }, > + { "ttl", required_argument, NULL, 't' }, > + { "ports", required_argument, NULL, 'p' }, > + { "single", no_argument, NULL, 's' }, > + { NULL, 0, NULL, 0 }, > + }; > + unsigned long mss4, mss6, wscale, ttl; > + unsigned int tcpipopts_mask = 0; > + > + if (argc < 2) > + usage(argv[0]); > + > + *ifindex = 0; > + *prog_id = 0; > + *tcpipopts = 0; > + *ports = NULL; > + *single = false; > + > + while (true) { > + int opt; > + > + opt = getopt_long(argc, argv, "", long_options, NULL); > + if (opt == -1) > + break; > + > + switch (opt) { > + case 'h': > + usage(argv[0]); > + break; > + case 'i': > + *ifindex = if_nametoindex(optarg); > + if (*ifindex == 0) > + usage(argv[0]); > + break; > + case 'x': > + *prog_id = parse_arg_ul(argv[0], optarg, UINT32_MAX); > + if (*prog_id == 0) > + usage(argv[0]); > + break; > + case 4: > + mss4 = parse_arg_ul(argv[0], optarg, UINT16_MAX); > + tcpipopts_mask |= 1 << 0; > + break; > + case 6: > + mss6 = parse_arg_ul(argv[0], optarg, UINT16_MAX); > + tcpipopts_mask |= 1 << 1; > + break; > + case 'w': > + wscale = parse_arg_ul(argv[0], optarg, 14); > + tcpipopts_mask |= 1 << 2; > + break; > + case 't': > + ttl = parse_arg_ul(argv[0], optarg, UINT8_MAX); > + tcpipopts_mask |= 1 << 3; > + break; > + case 'p': > + *ports = optarg; > + break; > + case 's': > + *single = true; > + break; > + default: > + usage(argv[0]); > + } > + } > + if (optind < argc) > + usage(argv[0]); > + > + if (tcpipopts_mask == 0xf) { > + if (mss4 == 0 || mss6 == 0 || wscale == 0 || ttl == 0) > + usage(argv[0]); > + *tcpipopts = (mss6 << 32) | (ttl << 24) | (wscale << 16) | mss4; > + } else if (tcpipopts_mask != 0) { > + usage(argv[0]); > + } > + > + if (*ifindex != 0 && *prog_id != 0) > + usage(argv[0]); > + if (*ifindex == 0 && *prog_id == 0) > + usage(argv[0]); > +} > + > +static int syncookie_attach(const char *argv0, unsigned int ifindex) > +{ > + struct bpf_prog_info info = {}; > + __u32 info_len = sizeof(info); > + char xdp_filename[PATH_MAX]; > + struct bpf_program *prog; > + struct bpf_object *obj; > + int prog_fd; > + int err; > + > + snprintf(xdp_filename, sizeof(xdp_filename), "%s_kern.o", argv0); > + obj = bpf_object__open_file(xdp_filename, NULL); > + err = libbpf_get_error(obj); > + if (err < 0) { > + fprintf(stderr, "Error: bpf_object__open_file: %s\n", strerror(-err)); > + return err; > + } > + > + err = bpf_object__load(obj); > + if (err < 0) { > + fprintf(stderr, "Error: bpf_object__open_file: %s\n", strerror(-err)); > + return err; > + } > + > + prog = bpf_object__find_program_by_name(obj, "syncookie_xdp"); > + if (!prog) { > + fprintf(stderr, "Error: bpf_object__find_program_by_name: program syncookie_xdp was not found\n"); > + return -ENOENT; > + } > + > + prog_fd = bpf_program__fd(prog); > + > + err = bpf_obj_get_info_by_fd(prog_fd, &info, &info_len); > + if (err < 0) { > + fprintf(stderr, "Error: bpf_obj_get_info_by_fd: %s\n", strerror(-err)); > + goto out; > + } > + attached_prog_id = info.id; > + signal(SIGINT, cleanup); > + signal(SIGTERM, cleanup); > + err = bpf_xdp_attach(ifindex, prog_fd, XDP_FLAGS_UPDATE_IF_NOEXIST, NULL); > + if (err < 0) { > + fprintf(stderr, "Error: bpf_set_link_xdp_fd: %s\n", strerror(-err)); > + signal(SIGINT, SIG_DFL); > + signal(SIGTERM, SIG_DFL); > + attached_prog_id = 0; > + goto out; > + } > + err = 0; > +out: > + bpf_object__close(obj); > + return err; > +} > + > +static int syncookie_open_bpf_maps(__u32 prog_id, int *values_map_fd, int *ports_map_fd) > +{ > + struct bpf_prog_info prog_info; > + __u32 map_ids[8]; > + __u32 info_len; > + int prog_fd; > + int err; > + int i; > + > + *values_map_fd = -1; > + *ports_map_fd = -1; > + > + prog_fd = bpf_prog_get_fd_by_id(prog_id); > + if (prog_fd < 0) { > + fprintf(stderr, "Error: bpf_prog_get_fd_by_id: %s\n", strerror(-prog_fd)); > + return prog_fd; > + } > + > + prog_info = (struct bpf_prog_info) { > + .nr_map_ids = 8, > + .map_ids = (__u64)map_ids, > + }; > + info_len = sizeof(prog_info); > + > + err = bpf_obj_get_info_by_fd(prog_fd, &prog_info, &info_len); > + if (err != 0) { > + fprintf(stderr, "Error: bpf_obj_get_info_by_fd: %s\n", strerror(-err)); > + goto out; > + } > + > + if (prog_info.type != BPF_PROG_TYPE_XDP) { > + fprintf(stderr, "Error: BPF prog type is not BPF_PROG_TYPE_XDP\n"); > + err = -ENOENT; > + goto out; > + } > + if (prog_info.nr_map_ids < 2) { > + fprintf(stderr, "Error: Found %u BPF maps, expected at least 2\n", > + prog_info.nr_map_ids); > + err = -ENOENT; > + goto out; > + } > + > + for (i = 0; i < prog_info.nr_map_ids; i++) { > + struct bpf_map_info map_info = {}; > + int map_fd; > + > + err = bpf_map_get_fd_by_id(map_ids[i]); > + if (err < 0) { > + fprintf(stderr, "Error: bpf_map_get_fd_by_id: %s\n", strerror(-err)); > + goto err_close_map_fds; > + } > + map_fd = err; > + > + info_len = sizeof(map_info); > + err = bpf_obj_get_info_by_fd(map_fd, &map_info, &info_len); > + if (err != 0) { > + fprintf(stderr, "Error: bpf_obj_get_info_by_fd: %s\n", strerror(-err)); > + close(map_fd); > + goto err_close_map_fds; > + } > + if (strcmp(map_info.name, "values") == 0) { > + *values_map_fd = map_fd; > + continue; > + } > + if (strcmp(map_info.name, "allowed_ports") == 0) { > + *ports_map_fd = map_fd; > + continue; > + } > + close(map_fd); > + } > + > + if (*values_map_fd != -1 && *ports_map_fd != -1) { > + err = 0; > + goto out; > + } > + > + err = -ENOENT; > + > +err_close_map_fds: > + if (*values_map_fd != -1) > + close(*values_map_fd); > + if (*ports_map_fd != -1) > + close(*ports_map_fd); > + *values_map_fd = -1; > + *ports_map_fd = -1; > + > +out: > + close(prog_fd); > + return err; > +} > + > +int main(int argc, char *argv[]) > +{ > + int values_map_fd, ports_map_fd; > + __u64 tcpipopts; > + bool firstiter; > + __u64 prevcnt; > + __u32 prog_id; > + char *ports; > + bool single; > + int err = 0; > + > + parse_options(argc, argv, &ifindex, &prog_id, &tcpipopts, &ports, &single); > + > + if (prog_id == 0) { > + err = bpf_xdp_query_id(ifindex, 0, &prog_id); > + if (err < 0) { > + fprintf(stderr, "Error: bpf_get_link_xdp_id: %s\n", strerror(-err)); > + goto out; > + } > + if (prog_id == 0) { > + err = syncookie_attach(argv[0], ifindex); > + if (err < 0) > + goto out; > + prog_id = attached_prog_id; > + } > + } > + > + err = syncookie_open_bpf_maps(prog_id, &values_map_fd, &ports_map_fd); > + if (err < 0) > + goto out; > + > + if (ports) { > + __u16 port_last = 0; > + __u32 port_idx = 0; > + char *p = ports; > + > + fprintf(stderr, "Replacing allowed ports\n"); > + > + while (p && *p != '\0') { > + char *token = strsep(&p, ","); > + __u16 port; > + > + port = parse_arg_ul(argv[0], token, UINT16_MAX); > + err = bpf_map_update_elem(ports_map_fd, &port_idx, &port, BPF_ANY); > + if (err != 0) { > + fprintf(stderr, "Error: bpf_map_update_elem: %s\n", strerror(-err)); > + fprintf(stderr, "Failed to add port %u (index %u)\n", > + port, port_idx); > + goto out_close_maps; > + } > + fprintf(stderr, "Added port %u\n", port); > + port_idx++; > + } > + err = bpf_map_update_elem(ports_map_fd, &port_idx, &port_last, BPF_ANY); > + if (err != 0) { > + fprintf(stderr, "Error: bpf_map_update_elem: %s\n", strerror(-err)); > + fprintf(stderr, "Failed to add the terminator value 0 (index %u)\n", > + port_idx); > + goto out_close_maps; > + } > + } > + > + if (tcpipopts) { > + __u32 key = 0; > + > + fprintf(stderr, "Replacing TCP/IP options\n"); > + > + err = bpf_map_update_elem(values_map_fd, &key, &tcpipopts, BPF_ANY); > + if (err != 0) { > + fprintf(stderr, "Error: bpf_map_update_elem: %s\n", strerror(-err)); > + goto out_close_maps; > + } > + } > + > + if ((ports || tcpipopts) && attached_prog_id == 0 && !single) > + goto out_close_maps; > + > + prevcnt = 0; > + firstiter = true; > + while (true) { > + __u32 key = 1; > + __u64 value; > + > + err = bpf_map_lookup_elem(values_map_fd, &key, &value); > + if (err != 0) { > + fprintf(stderr, "Error: bpf_map_lookup_elem: %s\n", strerror(-err)); > + goto out_close_maps; > + } > + if (firstiter) { > + prevcnt = value; > + firstiter = false; > + } > + if (single) { > + printf("Total SYNACKs generated: %llu\n", value); > + break; > + } > + printf("SYNACKs generated: %llu (total %llu)\n", value - prevcnt, value); > + prevcnt = value; > + sleep(1); > + } > + > +out_close_maps: > + close(values_map_fd); > + close(ports_map_fd); > +out: > + return err == 0 ? 0 : 1; > +} > -- > 2.30.2 > -- WBR, Yauheni