* [MPTCP][PATCH v6 mptcp-next 0/5] MP_FAIL support
@ 2021-07-28 9:35 Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending Geliang Tang
2021-07-28 10:37 ` [MPTCP][PATCH v6 mptcp-next 0/5] MP_FAIL support Paolo Abeni
0 siblings, 2 replies; 11+ messages in thread
From: Geliang Tang @ 2021-07-28 9:35 UTC (permalink / raw)
To: mptcp, geliangtang; +Cc: Geliang Tang
From: Geliang Tang <geliangtang@xiaomi.com>
v6: only rebased patch 1
- move the struct member 'fail_seq' behind 'ext_copy'.
- define OPTION_MPTCP_FAIL to BIT(12), BIT(11) is used by DSS
- move the MP_FAIL writing code at the beginning of mptcp_write_options,
and add the 'unlikely' tag.
- tag: export/20210728T080904
v5:
- patch 1, change "ret = true;" to "return true;"
- patch 3, in the single-subflow case, send MP_FAIL and receive the
echo, then temporarily handled by reset.
v4:
- just deal with the multiple subflows case, put the single subflow
case into the new 'infinite mapping' part.
v3:
- respond with MP_FAIL
- add single subflow check
- add infinite mapping sending and receiving
- export/20210626T054902
v2:
- MP_FAIL logic:
* Peer B send a DSS to peer A, and the data has been modify by the
middleboxes, then peer A detects the bad checksum.
* In the multiple subflows case, peer A sends MP_FAIL+RST back to peer B,
and peer A discards the data following the bad data sequence number. Peer
B receives this MP_FAIL+RST, and close this subflow.
* In the single subflow case, using the simple implementation, peer A
sends MP_FAIL back to peer B, and peer A fallback to a regular TCP. Peer
B receives this MP_FAIL, and fallback to a regular TCP.
Closes: https://github.com/multipath-tcp/mptcp_net-next/issues/52
Geliang Tang (5):
mptcp: MP_FAIL suboption sending
mptcp: MP_FAIL suboption receiving
mptcp: send out MP_FAIL when data checksum fails
mptcp: add the mibs for MP_FAIL
selftests: mptcp: add MP_FAIL mibs check
include/net/mptcp.h | 5 +-
net/mptcp/mib.c | 2 +
net/mptcp/mib.h | 2 +
net/mptcp/options.c | 78 ++++++++++++++++++-
net/mptcp/pm.c | 20 +++++
net/mptcp/protocol.h | 20 +++++
net/mptcp/subflow.c | 18 +++++
.../testing/selftests/net/mptcp/mptcp_join.sh | 38 +++++++++
8 files changed, 178 insertions(+), 5 deletions(-)
--
2.31.1
^ permalink raw reply [flat|nested] 11+ messages in thread
* [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending
2021-07-28 9:35 [MPTCP][PATCH v6 mptcp-next 0/5] MP_FAIL support Geliang Tang
@ 2021-07-28 9:35 ` Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Geliang Tang
2021-07-28 10:31 ` [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending Paolo Abeni
2021-07-28 10:37 ` [MPTCP][PATCH v6 mptcp-next 0/5] MP_FAIL support Paolo Abeni
1 sibling, 2 replies; 11+ messages in thread
From: Geliang Tang @ 2021-07-28 9:35 UTC (permalink / raw)
To: mptcp, geliangtang; +Cc: Geliang Tang
From: Geliang Tang <geliangtang@xiaomi.com>
This patch added the MP_FAIL suboption sending support.
Add a new flag named send_mp_fail in struct mptcp_subflow_context. If
this flag is set, send out MP_FAIL suboption.
Add a new member fail_seq in struct mptcp_out_options to save the data
sequence number to put into the MP_FAIL suboption.
An MP_FAIL option could be included in a RST or on the subflow-level
ACK.
Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
---
include/net/mptcp.h | 5 +++-
net/mptcp/options.c | 61 +++++++++++++++++++++++++++++++++++++++++---
net/mptcp/protocol.h | 3 +++
3 files changed, 64 insertions(+), 5 deletions(-)
diff --git a/include/net/mptcp.h b/include/net/mptcp.h
index 3236010afa29..6026bbefbffd 100644
--- a/include/net/mptcp.h
+++ b/include/net/mptcp.h
@@ -74,7 +74,10 @@ struct mptcp_out_options {
struct mptcp_addr_info addr;
u64 ahmac;
};
- struct mptcp_ext ext_copy;
+ struct {
+ struct mptcp_ext ext_copy;
+ u64 fail_seq;
+ };
struct {
u32 nonce;
u32 token;
diff --git a/net/mptcp/options.c b/net/mptcp/options.c
index 753d6ac43bff..2b15063c8009 100644
--- a/net/mptcp/options.c
+++ b/net/mptcp/options.c
@@ -763,7 +763,7 @@ static bool mptcp_established_options_mp_prio(struct sock *sk,
return true;
}
-static noinline void mptcp_established_options_rst(struct sock *sk, struct sk_buff *skb,
+static noinline bool mptcp_established_options_rst(struct sock *sk, struct sk_buff *skb,
unsigned int *size,
unsigned int remaining,
struct mptcp_out_options *opts)
@@ -771,12 +771,36 @@ static noinline void mptcp_established_options_rst(struct sock *sk, struct sk_bu
const struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
if (remaining < TCPOLEN_MPTCP_RST)
- return;
+ return false;
*size = TCPOLEN_MPTCP_RST;
opts->suboptions |= OPTION_MPTCP_RST;
opts->reset_transient = subflow->reset_transient;
opts->reset_reason = subflow->reset_reason;
+
+ return true;
+}
+
+static bool mptcp_established_options_mp_fail(struct sock *sk,
+ unsigned int *size,
+ unsigned int remaining,
+ struct mptcp_out_options *opts)
+{
+ struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
+
+ if (!subflow->send_mp_fail)
+ return false;
+
+ if (remaining < TCPOLEN_MPTCP_FAIL)
+ return false;
+
+ *size = TCPOLEN_MPTCP_FAIL;
+ opts->suboptions |= OPTION_MPTCP_FAIL;
+ opts->fail_seq = subflow->map_seq;
+
+ pr_debug("MP_FAIL fail_seq=%llu", opts->fail_seq);
+
+ return true;
}
bool mptcp_established_options(struct sock *sk, struct sk_buff *skb,
@@ -795,15 +819,30 @@ bool mptcp_established_options(struct sock *sk, struct sk_buff *skb,
return false;
if (unlikely(skb && TCP_SKB_CB(skb)->tcp_flags & TCPHDR_RST)) {
- mptcp_established_options_rst(sk, skb, size, remaining, opts);
+ if (mptcp_established_options_mp_fail(sk, &opt_size, remaining, opts)) {
+ *size += opt_size;
+ remaining -= opt_size;
+ }
+ if (mptcp_established_options_rst(sk, skb, &opt_size, remaining, opts)) {
+ *size += opt_size;
+ remaining -= opt_size;
+ }
return true;
}
snd_data_fin = mptcp_data_fin_enabled(msk);
if (mptcp_established_options_mp(sk, skb, snd_data_fin, &opt_size, remaining, opts))
ret = true;
- else if (mptcp_established_options_dss(sk, skb, snd_data_fin, &opt_size, remaining, opts))
+ else if (mptcp_established_options_dss(sk, skb, snd_data_fin, &opt_size, remaining, opts)) {
ret = true;
+ if (opts->ext_copy.use_ack) {
+ if (mptcp_established_options_mp_fail(sk, &opt_size, remaining, opts)) {
+ *size += opt_size;
+ remaining -= opt_size;
+ return true;
+ }
+ }
+ }
/* we reserved enough space for the above options, and exceeding the
* TCP option space would be fatal
@@ -1210,6 +1249,20 @@ static u16 mptcp_make_csum(const struct mptcp_ext *mpext)
void mptcp_write_options(__be32 *ptr, const struct tcp_sock *tp,
struct mptcp_out_options *opts)
{
+ if (unlikely(OPTION_MPTCP_FAIL & opts->suboptions)) {
+ const struct sock *ssk = (const struct sock *)tp;
+ struct mptcp_subflow_context *subflow;
+
+ subflow = mptcp_subflow_ctx(ssk);
+ subflow->send_mp_fail = 0;
+
+ *ptr++ = mptcp_option(MPTCPOPT_MP_FAIL,
+ TCPOLEN_MPTCP_FAIL,
+ 0, 0);
+ put_unaligned_be64(opts->fail_seq, ptr);
+ ptr += 2;
+ }
+
/* RST is mutually exclusive with everything else */
if (unlikely(OPTION_MPTCP_RST & opts->suboptions)) {
*ptr++ = mptcp_option(MPTCPOPT_RST,
diff --git a/net/mptcp/protocol.h b/net/mptcp/protocol.h
index e8a36ff52af6..b389fec18c89 100644
--- a/net/mptcp/protocol.h
+++ b/net/mptcp/protocol.h
@@ -27,6 +27,7 @@
#define OPTION_MPTCP_PRIO BIT(9)
#define OPTION_MPTCP_RST BIT(10)
#define OPTION_MPTCP_DSS BIT(11)
+#define OPTION_MPTCP_FAIL BIT(12)
/* MPTCP option subtypes */
#define MPTCPOPT_MP_CAPABLE 0
@@ -68,6 +69,7 @@
#define TCPOLEN_MPTCP_PRIO_ALIGN 4
#define TCPOLEN_MPTCP_FASTCLOSE 12
#define TCPOLEN_MPTCP_RST 4
+#define TCPOLEN_MPTCP_FAIL 12
#define TCPOLEN_MPTCP_MPC_ACK_DATA_CSUM (TCPOLEN_MPTCP_DSS_CHECKSUM + TCPOLEN_MPTCP_MPC_ACK_DATA)
@@ -429,6 +431,7 @@ struct mptcp_subflow_context {
mpc_map : 1,
backup : 1,
send_mp_prio : 1,
+ send_mp_fail : 1,
rx_eof : 1,
can_ack : 1, /* only after processing the remote a key */
disposable : 1, /* ctx can be free at ulp release time */
--
2.31.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending Geliang Tang
@ 2021-07-28 9:35 ` Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails Geliang Tang
2021-07-28 10:36 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Paolo Abeni
2021-07-28 10:31 ` [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending Paolo Abeni
1 sibling, 2 replies; 11+ messages in thread
From: Geliang Tang @ 2021-07-28 9:35 UTC (permalink / raw)
To: mptcp, geliangtang; +Cc: Geliang Tang
From: Geliang Tang <geliangtang@xiaomi.com>
This patch added handling for receiving MP_FAIL suboption.
Add a new members mp_fail and fail_seq in struct mptcp_options_received.
When MP_FAIL suboption is received, set mp_fail to 1 and save the sequence
number to fail_seq.
Then invoke mptcp_pm_mp_fail_received to deal with the MP_FAIL suboption.
Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
---
net/mptcp/options.c | 16 ++++++++++++++++
net/mptcp/pm.c | 5 +++++
net/mptcp/protocol.h | 3 +++
3 files changed, 24 insertions(+)
diff --git a/net/mptcp/options.c b/net/mptcp/options.c
index 2b15063c8009..cd9ec4acf127 100644
--- a/net/mptcp/options.c
+++ b/net/mptcp/options.c
@@ -336,6 +336,16 @@ static void mptcp_parse_option(const struct sk_buff *skb,
mp_opt->reset_reason = *ptr;
break;
+ case MPTCPOPT_MP_FAIL:
+ if (opsize != TCPOLEN_MPTCP_FAIL)
+ break;
+
+ ptr += 2;
+ mp_opt->mp_fail = 1;
+ mp_opt->fail_seq = get_unaligned_be64(ptr);
+ pr_debug("MP_FAIL: data_seq=%llu", mp_opt->fail_seq);
+ break;
+
default:
break;
}
@@ -364,6 +374,7 @@ void mptcp_get_options(const struct sock *sk,
mp_opt->reset = 0;
mp_opt->csum_reqd = READ_ONCE(msk->csum_enabled);
mp_opt->deny_join_id0 = 0;
+ mp_opt->mp_fail = 0;
length = (th->doff * 4) - sizeof(struct tcphdr);
ptr = (const unsigned char *)(th + 1);
@@ -1147,6 +1158,11 @@ bool mptcp_incoming_options(struct sock *sk, struct sk_buff *skb)
mp_opt.mp_prio = 0;
}
+ if (mp_opt.mp_fail) {
+ mptcp_pm_mp_fail_received(sk, mp_opt.fail_seq);
+ mp_opt.mp_fail = 0;
+ }
+
if (mp_opt.reset) {
subflow->reset_seen = 1;
subflow->reset_reason = mp_opt.reset_reason;
diff --git a/net/mptcp/pm.c b/net/mptcp/pm.c
index da0c4c925350..6ab386ff3294 100644
--- a/net/mptcp/pm.c
+++ b/net/mptcp/pm.c
@@ -249,6 +249,11 @@ void mptcp_pm_mp_prio_received(struct sock *sk, u8 bkup)
mptcp_event(MPTCP_EVENT_SUB_PRIORITY, mptcp_sk(subflow->conn), sk, GFP_ATOMIC);
}
+void mptcp_pm_mp_fail_received(struct sock *sk, u64 fail_seq)
+{
+ pr_debug("fail_seq=%llu", fail_seq);
+}
+
/* path manager helpers */
bool mptcp_pm_add_addr_signal(struct mptcp_sock *msk, struct sk_buff *skb,
diff --git a/net/mptcp/protocol.h b/net/mptcp/protocol.h
index b389fec18c89..09d0e9406ea9 100644
--- a/net/mptcp/protocol.h
+++ b/net/mptcp/protocol.h
@@ -140,6 +140,7 @@ struct mptcp_options_received {
add_addr : 1,
rm_addr : 1,
mp_prio : 1,
+ mp_fail : 1,
echo : 1,
csum_reqd : 1,
backup : 1,
@@ -161,6 +162,7 @@ struct mptcp_options_received {
u64 ahmac;
u8 reset_reason:4;
u8 reset_transient:1;
+ u64 fail_seq;
};
static inline __be32 mptcp_option(u8 subopt, u8 len, u8 nib, u8 field)
@@ -727,6 +729,7 @@ void mptcp_pm_mp_prio_received(struct sock *sk, u8 bkup);
int mptcp_pm_nl_mp_prio_send_ack(struct mptcp_sock *msk,
struct mptcp_addr_info *addr,
u8 bkup);
+void mptcp_pm_mp_fail_received(struct sock *sk, u64 fail_seq);
void mptcp_pm_free_anno_list(struct mptcp_sock *msk);
bool mptcp_pm_sport_in_anno_list(struct mptcp_sock *msk, const struct sock *sk);
struct mptcp_pm_add_entry *
--
2.31.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Geliang Tang
@ 2021-07-28 9:35 ` Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 4/5] mptcp: add the mibs for MP_FAIL Geliang Tang
2021-07-28 23:31 ` [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails Mat Martineau
2021-07-28 10:36 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Paolo Abeni
1 sibling, 2 replies; 11+ messages in thread
From: Geliang Tang @ 2021-07-28 9:35 UTC (permalink / raw)
To: mptcp, geliangtang; +Cc: Geliang Tang
From: Geliang Tang <geliangtang@xiaomi.com>
When a bad checksum is detected, set the send_mp_fail flag to send out
the MP_FAIL option.
Add a new function mptcp_has_another_subflow() to check whether there's
only a single subflow.
When multiple subflows are in use, close the affected subflow with a RST
that includes an MP_FAIL option and discard the data with the bad
checksum.
Set the sk_state of the subsocket to TCP_CLOSE, then the flag
MPTCP_WORK_CLOSE_SUBFLOW will be set in subflow_sched_work_if_closed,
and the subflow will be closed.
When a single subfow is in use, send back an MP_FAIL option on the
subflow-level ACK. And the receiver of this MP_FAIL respond with an
MP_FAIL in the reverse direction.
Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
---
net/mptcp/pm.c | 14 ++++++++++++++
net/mptcp/protocol.h | 14 ++++++++++++++
net/mptcp/subflow.c | 17 +++++++++++++++++
3 files changed, 45 insertions(+)
diff --git a/net/mptcp/pm.c b/net/mptcp/pm.c
index 6ab386ff3294..c2df5cc28ba1 100644
--- a/net/mptcp/pm.c
+++ b/net/mptcp/pm.c
@@ -251,7 +251,21 @@ void mptcp_pm_mp_prio_received(struct sock *sk, u8 bkup)
void mptcp_pm_mp_fail_received(struct sock *sk, u64 fail_seq)
{
+ struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
+
pr_debug("fail_seq=%llu", fail_seq);
+
+ if (!mptcp_has_another_subflow(sk)) {
+ if (!subflow->mp_fail_expect_echo) {
+ subflow->send_mp_fail = 1;
+ } else {
+ subflow->mp_fail_expect_echo = 0;
+ /* TODO the single-subflow case is temporarily
+ * handled by reset.
+ */
+ mptcp_subflow_reset(sk);
+ }
+ }
}
/* path manager helpers */
diff --git a/net/mptcp/protocol.h b/net/mptcp/protocol.h
index 09d0e9406ea9..c46011318f65 100644
--- a/net/mptcp/protocol.h
+++ b/net/mptcp/protocol.h
@@ -434,6 +434,7 @@ struct mptcp_subflow_context {
backup : 1,
send_mp_prio : 1,
send_mp_fail : 1,
+ mp_fail_expect_echo : 1,
rx_eof : 1,
can_ack : 1, /* only after processing the remote a key */
disposable : 1, /* ctx can be free at ulp release time */
@@ -615,6 +616,19 @@ static inline void mptcp_subflow_tcp_fallback(struct sock *sk,
inet_csk(sk)->icsk_af_ops = ctx->icsk_af_ops;
}
+static inline bool mptcp_has_another_subflow(struct sock *ssk)
+{
+ struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk), *tmp;
+ struct mptcp_sock *msk = mptcp_sk(subflow->conn);
+
+ mptcp_for_each_subflow(msk, tmp) {
+ if (tmp != subflow)
+ return true;
+ }
+
+ return false;
+}
+
void __init mptcp_proto_init(void);
#if IS_ENABLED(CONFIG_MPTCP_IPV6)
int __init mptcp_proto_v6_init(void);
diff --git a/net/mptcp/subflow.c b/net/mptcp/subflow.c
index 1151926d335b..a69839520472 100644
--- a/net/mptcp/subflow.c
+++ b/net/mptcp/subflow.c
@@ -910,6 +910,7 @@ static enum mapping_status validate_data_csum(struct sock *ssk, struct sk_buff *
csum = csum_partial(&header, sizeof(header), subflow->map_data_csum);
if (unlikely(csum_fold(csum))) {
MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DATACSUMERR);
+ subflow->send_mp_fail = 1;
return subflow->mp_join ? MAPPING_INVALID : MAPPING_DUMMY;
}
@@ -1157,6 +1158,22 @@ static bool subflow_check_data_avail(struct sock *ssk)
fallback:
/* RFC 8684 section 3.7. */
+ if (subflow->send_mp_fail) {
+ if (mptcp_has_another_subflow(ssk)) {
+ ssk->sk_err = EBADMSG;
+ tcp_set_state(ssk, TCP_CLOSE);
+ subflow->reset_transient = 0;
+ subflow->reset_reason = MPTCP_RST_EMIDDLEBOX;
+ tcp_send_active_reset(ssk, GFP_ATOMIC);
+ while ((skb = skb_peek(&ssk->sk_receive_queue)))
+ sk_eat_skb(ssk, skb);
+ } else {
+ subflow->mp_fail_expect_echo = 1;
+ }
+ WRITE_ONCE(subflow->data_avail, 0);
+ return true;
+ }
+
if (subflow->mp_join || subflow->fully_established) {
/* fatal protocol error, close the socket.
* subflow_error_report() will introduce the appropriate barriers
--
2.31.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* [MPTCP][PATCH v6 mptcp-next 4/5] mptcp: add the mibs for MP_FAIL
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails Geliang Tang
@ 2021-07-28 9:35 ` Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 5/5] selftests: mptcp: add MP_FAIL mibs check Geliang Tang
2021-07-28 23:31 ` [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails Mat Martineau
1 sibling, 1 reply; 11+ messages in thread
From: Geliang Tang @ 2021-07-28 9:35 UTC (permalink / raw)
To: mptcp, geliangtang; +Cc: Geliang Tang
From: Geliang Tang <geliangtang@xiaomi.com>
This patch added the mibs for MP_FAIL: MPTCP_MIB_MPFAILTX and
MPTCP_MIB_MPFAILRX.
Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
---
net/mptcp/mib.c | 2 ++
net/mptcp/mib.h | 2 ++
net/mptcp/options.c | 1 +
net/mptcp/pm.c | 1 +
net/mptcp/subflow.c | 1 +
5 files changed, 7 insertions(+)
diff --git a/net/mptcp/mib.c b/net/mptcp/mib.c
index 3a7c4e7b2d79..b21ff9be04c6 100644
--- a/net/mptcp/mib.c
+++ b/net/mptcp/mib.c
@@ -44,6 +44,8 @@ static const struct snmp_mib mptcp_snmp_list[] = {
SNMP_MIB_ITEM("RmSubflow", MPTCP_MIB_RMSUBFLOW),
SNMP_MIB_ITEM("MPPrioTx", MPTCP_MIB_MPPRIOTX),
SNMP_MIB_ITEM("MPPrioRx", MPTCP_MIB_MPPRIORX),
+ SNMP_MIB_ITEM("MPFailTx", MPTCP_MIB_MPFAILTX),
+ SNMP_MIB_ITEM("MPFailRx", MPTCP_MIB_MPFAILRX),
SNMP_MIB_ITEM("RcvPruned", MPTCP_MIB_RCVPRUNED),
SNMP_MIB_ITEM("SubflowStale", MPTCP_MIB_SUBFLOWSTALE),
SNMP_MIB_ITEM("SubflowRecover", MPTCP_MIB_SUBFLOWRECOVER),
diff --git a/net/mptcp/mib.h b/net/mptcp/mib.h
index 8ec16c991aac..ecd3d8b117e0 100644
--- a/net/mptcp/mib.h
+++ b/net/mptcp/mib.h
@@ -37,6 +37,8 @@ enum linux_mptcp_mib_field {
MPTCP_MIB_RMSUBFLOW, /* Remove a subflow */
MPTCP_MIB_MPPRIOTX, /* Transmit a MP_PRIO */
MPTCP_MIB_MPPRIORX, /* Received a MP_PRIO */
+ MPTCP_MIB_MPFAILTX, /* Transmit a MP_FAIL */
+ MPTCP_MIB_MPFAILRX, /* Received a MP_FAIL */
MPTCP_MIB_RCVPRUNED, /* Incoming packet dropped due to memory limit */
MPTCP_MIB_SUBFLOWSTALE, /* Subflows entered 'stale' status */
MPTCP_MIB_SUBFLOWRECOVER, /* Subflows returned to active status after being stale */
diff --git a/net/mptcp/options.c b/net/mptcp/options.c
index cd9ec4acf127..8b899c308b83 100644
--- a/net/mptcp/options.c
+++ b/net/mptcp/options.c
@@ -1160,6 +1160,7 @@ bool mptcp_incoming_options(struct sock *sk, struct sk_buff *skb)
if (mp_opt.mp_fail) {
mptcp_pm_mp_fail_received(sk, mp_opt.fail_seq);
+ MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPFAILRX);
mp_opt.mp_fail = 0;
}
diff --git a/net/mptcp/pm.c b/net/mptcp/pm.c
index c2df5cc28ba1..43530d3a78e9 100644
--- a/net/mptcp/pm.c
+++ b/net/mptcp/pm.c
@@ -258,6 +258,7 @@ void mptcp_pm_mp_fail_received(struct sock *sk, u64 fail_seq)
if (!mptcp_has_another_subflow(sk)) {
if (!subflow->mp_fail_expect_echo) {
subflow->send_mp_fail = 1;
+ MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPFAILTX);
} else {
subflow->mp_fail_expect_echo = 0;
/* TODO the single-subflow case is temporarily
diff --git a/net/mptcp/subflow.c b/net/mptcp/subflow.c
index a69839520472..c25b1d961206 100644
--- a/net/mptcp/subflow.c
+++ b/net/mptcp/subflow.c
@@ -911,6 +911,7 @@ static enum mapping_status validate_data_csum(struct sock *ssk, struct sk_buff *
if (unlikely(csum_fold(csum))) {
MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DATACSUMERR);
subflow->send_mp_fail = 1;
+ MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_MPFAILTX);
return subflow->mp_join ? MAPPING_INVALID : MAPPING_DUMMY;
}
--
2.31.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* [MPTCP][PATCH v6 mptcp-next 5/5] selftests: mptcp: add MP_FAIL mibs check
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 4/5] mptcp: add the mibs for MP_FAIL Geliang Tang
@ 2021-07-28 9:35 ` Geliang Tang
0 siblings, 0 replies; 11+ messages in thread
From: Geliang Tang @ 2021-07-28 9:35 UTC (permalink / raw)
To: mptcp, geliangtang; +Cc: Geliang Tang
From: Geliang Tang <geliangtang@xiaomi.com>
This patch added a function chk_fail_nr to check the mibs for MP_FAIL.
Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
---
.../testing/selftests/net/mptcp/mptcp_join.sh | 38 +++++++++++++++++++
1 file changed, 38 insertions(+)
diff --git a/tools/testing/selftests/net/mptcp/mptcp_join.sh b/tools/testing/selftests/net/mptcp/mptcp_join.sh
index 937e861e9490..551fcce7c2f2 100755
--- a/tools/testing/selftests/net/mptcp/mptcp_join.sh
+++ b/tools/testing/selftests/net/mptcp/mptcp_join.sh
@@ -566,6 +566,43 @@ chk_csum_nr()
fi
}
+chk_fail_nr()
+{
+ local mp_fail_nr_tx=$1
+ local mp_fail_nr_rx=$2
+ local count
+ local dump_stats
+
+ printf "%-39s %s" " " "ftx"
+ count=`ip netns exec $ns1 nstat -as | grep MPTcpExtMPFailTx | awk '{print $2}'`
+ [ -z "$count" ] && count=0
+ if [ "$count" != "$mp_fail_nr_tx" ]; then
+ echo "[fail] got $count MP_FAIL[s] TX expected $mp_fail_nr_tx"
+ ret=1
+ dump_stats=1
+ else
+ echo -n "[ ok ]"
+ fi
+
+ echo -n " - frx "
+ count=`ip netns exec $ns2 nstat -as | grep MPTcpExtMPFailRx | awk '{print $2}'`
+ [ -z "$count" ] && count=0
+ if [ "$count" != "$mp_fail_nr_rx" ]; then
+ echo "[fail] got $count MP_FAIL[s] RX expected $mp_fail_nr_rx"
+ ret=1
+ dump_stats=1
+ else
+ echo "[ ok ]"
+ fi
+
+ if [ "${dump_stats}" = 1 ]; then
+ echo Server ns stats
+ ip netns exec $ns1 nstat -as | grep MPTcp
+ echo Client ns stats
+ ip netns exec $ns2 nstat -as | grep MPTcp
+ fi
+}
+
chk_join_nr()
{
local msg="$1"
@@ -615,6 +652,7 @@ chk_join_nr()
fi
if [ $checksum -eq 1 ]; then
chk_csum_nr
+ chk_fail_nr 0 0
fi
}
--
2.31.1
^ permalink raw reply related [flat|nested] 11+ messages in thread
* Re: [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Geliang Tang
@ 2021-07-28 10:31 ` Paolo Abeni
1 sibling, 0 replies; 11+ messages in thread
From: Paolo Abeni @ 2021-07-28 10:31 UTC (permalink / raw)
To: Geliang Tang, mptcp; +Cc: Geliang Tang
Hello,
On Wed, 2021-07-28 at 17:35 +0800, Geliang Tang wrote:
> From: Geliang Tang <geliangtang@xiaomi.com>
>
> This patch added the MP_FAIL suboption sending support.
>
> Add a new flag named send_mp_fail in struct mptcp_subflow_context. If
> this flag is set, send out MP_FAIL suboption.
>
> Add a new member fail_seq in struct mptcp_out_options to save the data
> sequence number to put into the MP_FAIL suboption.
>
> An MP_FAIL option could be included in a RST or on the subflow-level
> ACK.
>
> Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
> ---
> include/net/mptcp.h | 5 +++-
> net/mptcp/options.c | 61 +++++++++++++++++++++++++++++++++++++++++---
> net/mptcp/protocol.h | 3 +++
> 3 files changed, 64 insertions(+), 5 deletions(-)
>
> diff --git a/include/net/mptcp.h b/include/net/mptcp.h
> index 3236010afa29..6026bbefbffd 100644
> --- a/include/net/mptcp.h
> +++ b/include/net/mptcp.h
> @@ -74,7 +74,10 @@ struct mptcp_out_options {
> struct mptcp_addr_info addr;
> u64 ahmac;
> };
> - struct mptcp_ext ext_copy;
> + struct {
> + struct mptcp_ext ext_copy;
> + u64 fail_seq;
> + };
> struct {
> u32 nonce;
> u32 token;
> diff --git a/net/mptcp/options.c b/net/mptcp/options.c
> index 753d6ac43bff..2b15063c8009 100644
> --- a/net/mptcp/options.c
> +++ b/net/mptcp/options.c
> @@ -763,7 +763,7 @@ static bool mptcp_established_options_mp_prio(struct sock *sk,
> return true;
> }
>
> -static noinline void mptcp_established_options_rst(struct sock *sk, struct sk_buff *skb,
> +static noinline bool mptcp_established_options_rst(struct sock *sk, struct sk_buff *skb,
> unsigned int *size,
> unsigned int remaining,
> struct mptcp_out_options *opts)
> @@ -771,12 +771,36 @@ static noinline void mptcp_established_options_rst(struct sock *sk, struct sk_bu
> const struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
>
> if (remaining < TCPOLEN_MPTCP_RST)
> - return;
> + return false;
>
> *size = TCPOLEN_MPTCP_RST;
> opts->suboptions |= OPTION_MPTCP_RST;
> opts->reset_transient = subflow->reset_transient;
> opts->reset_reason = subflow->reset_reason;
> +
> + return true;
> +}
> +
> +static bool mptcp_established_options_mp_fail(struct sock *sk,
> + unsigned int *size,
> + unsigned int remaining,
> + struct mptcp_out_options *opts)
> +{
> + struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
> +
> + if (!subflow->send_mp_fail)
> + return false;
> +
> + if (remaining < TCPOLEN_MPTCP_FAIL)
> + return false;
> +
> + *size = TCPOLEN_MPTCP_FAIL;
> + opts->suboptions |= OPTION_MPTCP_FAIL;
> + opts->fail_seq = subflow->map_seq;
> +
> + pr_debug("MP_FAIL fail_seq=%llu", opts->fail_seq);
> +
> + return true;
> }
>
> bool mptcp_established_options(struct sock *sk, struct sk_buff *skb,
> @@ -795,15 +819,30 @@ bool mptcp_established_options(struct sock *sk, struct sk_buff *skb,
> return false;
>
> if (unlikely(skb && TCP_SKB_CB(skb)->tcp_flags & TCPHDR_RST)) {
> - mptcp_established_options_rst(sk, skb, size, remaining, opts);
> + if (mptcp_established_options_mp_fail(sk, &opt_size, remaining, opts)) {
> + *size += opt_size;
> + remaining -= opt_size;
> + }
> + if (mptcp_established_options_rst(sk, skb, &opt_size, remaining, opts)) {
> + *size += opt_size;
> + remaining -= opt_size;
> + }
> return true;
> }
>
> snd_data_fin = mptcp_data_fin_enabled(msk);
> if (mptcp_established_options_mp(sk, skb, snd_data_fin, &opt_size, remaining, opts))
> ret = true;
> - else if (mptcp_established_options_dss(sk, skb, snd_data_fin, &opt_size, remaining, opts))
> + else if (mptcp_established_options_dss(sk, skb, snd_data_fin, &opt_size, remaining, opts)) {
> ret = true;
> + if (opts->ext_copy.use_ack) {
I *think* we could drop this check as the RFC says:
"""
it will send back an MP_FAIL option on
the subflow-level ACK,
"""
And to me subflow-level ACK really means TCP ack, that is any packet on
a TCP-established subflow will do.
Anyhow I think we can adjust the above with a squash-to patch after
this series is merged, as it already went through several iterations
and overall it LGTM.
While at that a few 'unlikely()' annotation will help ;)
/P
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails Geliang Tang
@ 2021-07-28 10:36 ` Paolo Abeni
2021-08-12 7:09 ` Geliang Tang
1 sibling, 1 reply; 11+ messages in thread
From: Paolo Abeni @ 2021-07-28 10:36 UTC (permalink / raw)
To: Geliang Tang, mptcp; +Cc: Geliang Tang
On Wed, 2021-07-28 at 17:35 +0800, Geliang Tang wrote:
> From: Geliang Tang <geliangtang@xiaomi.com>
>
> This patch added handling for receiving MP_FAIL suboption.
>
> Add a new members mp_fail and fail_seq in struct mptcp_options_received.
> When MP_FAIL suboption is received, set mp_fail to 1 and save the sequence
> number to fail_seq.
>
> Then invoke mptcp_pm_mp_fail_received to deal with the MP_FAIL suboption.
>
> Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
> ---
> net/mptcp/options.c | 16 ++++++++++++++++
> net/mptcp/pm.c | 5 +++++
> net/mptcp/protocol.h | 3 +++
> 3 files changed, 24 insertions(+)
>
> diff --git a/net/mptcp/options.c b/net/mptcp/options.c
> index 2b15063c8009..cd9ec4acf127 100644
> --- a/net/mptcp/options.c
> +++ b/net/mptcp/options.c
> @@ -336,6 +336,16 @@ static void mptcp_parse_option(const struct sk_buff *skb,
> mp_opt->reset_reason = *ptr;
> break;
>
> + case MPTCPOPT_MP_FAIL:
> + if (opsize != TCPOLEN_MPTCP_FAIL)
> + break;
> +
> + ptr += 2;
> + mp_opt->mp_fail = 1;
> + mp_opt->fail_seq = get_unaligned_be64(ptr);
> + pr_debug("MP_FAIL: data_seq=%llu", mp_opt->fail_seq);
> + break;
> +
> default:
> break;
> }
> @@ -364,6 +374,7 @@ void mptcp_get_options(const struct sock *sk,
> mp_opt->reset = 0;
> mp_opt->csum_reqd = READ_ONCE(msk->csum_enabled);
> mp_opt->deny_join_id0 = 0;
> + mp_opt->mp_fail = 0;
>
> length = (th->doff * 4) - sizeof(struct tcphdr);
> ptr = (const unsigned char *)(th + 1);
> @@ -1147,6 +1158,11 @@ bool mptcp_incoming_options(struct sock *sk, struct sk_buff *skb)
> mp_opt.mp_prio = 0;
> }
>
> + if (mp_opt.mp_fail) {
> + mptcp_pm_mp_fail_received(sk, mp_opt.fail_seq);
> + mp_opt.mp_fail = 0;
> + }
> +
Side note not specifically related to this patch: usually we get a
single MPTCP subopt per packet: a DSS. So we could optimize this code
path with something alike:
if (unlikely(any subopt other than dss is present))
// go checking all of them individually
To do the above we likely need to wrap all the 'mp_capable',
'fastclose', 'rm_addr' flags in a single bitmask. e.v. using a union.
/P
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [MPTCP][PATCH v6 mptcp-next 0/5] MP_FAIL support
2021-07-28 9:35 [MPTCP][PATCH v6 mptcp-next 0/5] MP_FAIL support Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending Geliang Tang
@ 2021-07-28 10:37 ` Paolo Abeni
1 sibling, 0 replies; 11+ messages in thread
From: Paolo Abeni @ 2021-07-28 10:37 UTC (permalink / raw)
To: Geliang Tang, mptcp; +Cc: Geliang Tang
On Wed, 2021-07-28 at 17:35 +0800, Geliang Tang wrote:
> From: Geliang Tang <geliangtang@xiaomi.com>
>
> v6: only rebased patch 1
> - move the struct member 'fail_seq' behind 'ext_copy'.
> - define OPTION_MPTCP_FAIL to BIT(12), BIT(11) is used by DSS
> - move the MP_FAIL writing code at the beginning of mptcp_write_options,
> and add the 'unlikely' tag.
> - tag: export/20210728T080904
>
> v5:
> - patch 1, change "ret = true;" to "return true;"
> - patch 3, in the single-subflow case, send MP_FAIL and receive the
> echo, then temporarily handled by reset.
>
> v4:
> - just deal with the multiple subflows case, put the single subflow
> case into the new 'infinite mapping' part.
>
> v3:
> - respond with MP_FAIL
> - add single subflow check
> - add infinite mapping sending and receiving
> - export/20210626T054902
>
> v2:
> - MP_FAIL logic:
> * Peer B send a DSS to peer A, and the data has been modify by the
> middleboxes, then peer A detects the bad checksum.
> * In the multiple subflows case, peer A sends MP_FAIL+RST back to peer B,
> and peer A discards the data following the bad data sequence number. Peer
> B receives this MP_FAIL+RST, and close this subflow.
> * In the single subflow case, using the simple implementation, peer A
> sends MP_FAIL back to peer B, and peer A fallback to a regular TCP. Peer
> B receives this MP_FAIL, and fallback to a regular TCP.
>
> Closes: https://github.com/multipath-tcp/mptcp_net-next/issues/52
>
> Geliang Tang (5):
> mptcp: MP_FAIL suboption sending
> mptcp: MP_FAIL suboption receiving
> mptcp: send out MP_FAIL when data checksum fails
> mptcp: add the mibs for MP_FAIL
> selftests: mptcp: add MP_FAIL mibs check
>
> include/net/mptcp.h | 5 +-
> net/mptcp/mib.c | 2 +
> net/mptcp/mib.h | 2 +
> net/mptcp/options.c | 78 ++++++++++++++++++-
> net/mptcp/pm.c | 20 +++++
> net/mptcp/protocol.h | 20 +++++
> net/mptcp/subflow.c | 18 +++++
> .../testing/selftests/net/mptcp/mptcp_join.sh | 38 +++++++++
> 8 files changed, 178 insertions(+), 5 deletions(-)
I had a couple of minor comments, but they could be addressed with
squash-to or even completely unrelated changes.
Overall LGTM, thanks!
/P
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 4/5] mptcp: add the mibs for MP_FAIL Geliang Tang
@ 2021-07-28 23:31 ` Mat Martineau
1 sibling, 0 replies; 11+ messages in thread
From: Mat Martineau @ 2021-07-28 23:31 UTC (permalink / raw)
To: Geliang Tang; +Cc: mptcp, Geliang Tang
On Wed, 28 Jul 2021, Geliang Tang wrote:
> From: Geliang Tang <geliangtang@xiaomi.com>
>
> When a bad checksum is detected, set the send_mp_fail flag to send out
> the MP_FAIL option.
>
> Add a new function mptcp_has_another_subflow() to check whether there's
> only a single subflow.
>
> When multiple subflows are in use, close the affected subflow with a RST
> that includes an MP_FAIL option and discard the data with the bad
> checksum.
>
Thanks for the test code! I do see in wireshark that the multiple subflow
case sends a TCP RST with both MP_FAIL and MP_TCPRST options when the
checksum fails on one subflow.
> Set the sk_state of the subsocket to TCP_CLOSE, then the flag
> MPTCP_WORK_CLOSE_SUBFLOW will be set in subflow_sched_work_if_closed,
> and the subflow will be closed.
>
> When a single subfow is in use, send back an MP_FAIL option on the
> subflow-level ACK. And the receiver of this MP_FAIL respond with an
> MP_FAIL in the reverse direction.
>
The single subflow case has some unexpected behavior:
1. The checksum failure is detected, a packet is sent with MP_FAIL.
However, the packet also has data payload and no DSS option.
2. The peer receives MP_FAIL, and echoes back. But it sends two TCP RST +
MP_FAIL packets back-to-back.
I'll upload a pcap to
https://github.com/multipath-tcp/mptcp_net-next/issues/52
I think the right temporary behavior (before implementing infinite
mappings) for single subflow checksum failure is to do what the RFC says
for non-contiguous data: "In the rare case that the data is not contiguous
(which could happen when there is only one subflow but it is
retransmitting data from a subflow that has recently been uncleanly
closed), the receiver MUST close the subflow with a RST with MP_FAIL." So,
in step 1 above the peer that detected the bad checksum would still send
MP_FAIL but with the RST flag. And then the echo would not be needed
because the path would already be disconnected by the RST.
What do you think?
Mat
> Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
> ---
> net/mptcp/pm.c | 14 ++++++++++++++
> net/mptcp/protocol.h | 14 ++++++++++++++
> net/mptcp/subflow.c | 17 +++++++++++++++++
> 3 files changed, 45 insertions(+)
>
> diff --git a/net/mptcp/pm.c b/net/mptcp/pm.c
> index 6ab386ff3294..c2df5cc28ba1 100644
> --- a/net/mptcp/pm.c
> +++ b/net/mptcp/pm.c
> @@ -251,7 +251,21 @@ void mptcp_pm_mp_prio_received(struct sock *sk, u8 bkup)
>
> void mptcp_pm_mp_fail_received(struct sock *sk, u64 fail_seq)
> {
> + struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(sk);
> +
> pr_debug("fail_seq=%llu", fail_seq);
> +
> + if (!mptcp_has_another_subflow(sk)) {
> + if (!subflow->mp_fail_expect_echo) {
> + subflow->send_mp_fail = 1;
> + } else {
> + subflow->mp_fail_expect_echo = 0;
> + /* TODO the single-subflow case is temporarily
> + * handled by reset.
> + */
> + mptcp_subflow_reset(sk);
> + }
> + }
> }
>
> /* path manager helpers */
> diff --git a/net/mptcp/protocol.h b/net/mptcp/protocol.h
> index 09d0e9406ea9..c46011318f65 100644
> --- a/net/mptcp/protocol.h
> +++ b/net/mptcp/protocol.h
> @@ -434,6 +434,7 @@ struct mptcp_subflow_context {
> backup : 1,
> send_mp_prio : 1,
> send_mp_fail : 1,
> + mp_fail_expect_echo : 1,
> rx_eof : 1,
> can_ack : 1, /* only after processing the remote a key */
> disposable : 1, /* ctx can be free at ulp release time */
> @@ -615,6 +616,19 @@ static inline void mptcp_subflow_tcp_fallback(struct sock *sk,
> inet_csk(sk)->icsk_af_ops = ctx->icsk_af_ops;
> }
>
> +static inline bool mptcp_has_another_subflow(struct sock *ssk)
> +{
> + struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk), *tmp;
> + struct mptcp_sock *msk = mptcp_sk(subflow->conn);
> +
> + mptcp_for_each_subflow(msk, tmp) {
> + if (tmp != subflow)
> + return true;
> + }
> +
> + return false;
> +}
> +
> void __init mptcp_proto_init(void);
> #if IS_ENABLED(CONFIG_MPTCP_IPV6)
> int __init mptcp_proto_v6_init(void);
> diff --git a/net/mptcp/subflow.c b/net/mptcp/subflow.c
> index 1151926d335b..a69839520472 100644
> --- a/net/mptcp/subflow.c
> +++ b/net/mptcp/subflow.c
> @@ -910,6 +910,7 @@ static enum mapping_status validate_data_csum(struct sock *ssk, struct sk_buff *
> csum = csum_partial(&header, sizeof(header), subflow->map_data_csum);
> if (unlikely(csum_fold(csum))) {
> MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DATACSUMERR);
> + subflow->send_mp_fail = 1;
> return subflow->mp_join ? MAPPING_INVALID : MAPPING_DUMMY;
> }
>
> @@ -1157,6 +1158,22 @@ static bool subflow_check_data_avail(struct sock *ssk)
>
> fallback:
> /* RFC 8684 section 3.7. */
> + if (subflow->send_mp_fail) {
> + if (mptcp_has_another_subflow(ssk)) {
> + ssk->sk_err = EBADMSG;
> + tcp_set_state(ssk, TCP_CLOSE);
> + subflow->reset_transient = 0;
> + subflow->reset_reason = MPTCP_RST_EMIDDLEBOX;
> + tcp_send_active_reset(ssk, GFP_ATOMIC);
> + while ((skb = skb_peek(&ssk->sk_receive_queue)))
> + sk_eat_skb(ssk, skb);
> + } else {
> + subflow->mp_fail_expect_echo = 1;
> + }
> + WRITE_ONCE(subflow->data_avail, 0);
> + return true;
> + }
> +
> if (subflow->mp_join || subflow->fully_established) {
> /* fatal protocol error, close the socket.
> * subflow_error_report() will introduce the appropriate barriers
> --
> 2.31.1
>
>
>
--
Mat Martineau
Intel
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving
2021-07-28 10:36 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Paolo Abeni
@ 2021-08-12 7:09 ` Geliang Tang
0 siblings, 0 replies; 11+ messages in thread
From: Geliang Tang @ 2021-08-12 7:09 UTC (permalink / raw)
To: Paolo Abeni; +Cc: MPTCP Upstream, Geliang Tang
Hi Paolo,
Paolo Abeni <pabeni@redhat.com> 于2021年7月28日周三 下午6:36写道:
>
> On Wed, 2021-07-28 at 17:35 +0800, Geliang Tang wrote:
> > From: Geliang Tang <geliangtang@xiaomi.com>
> >
> > This patch added handling for receiving MP_FAIL suboption.
> >
> > Add a new members mp_fail and fail_seq in struct mptcp_options_received.
> > When MP_FAIL suboption is received, set mp_fail to 1 and save the sequence
> > number to fail_seq.
> >
> > Then invoke mptcp_pm_mp_fail_received to deal with the MP_FAIL suboption.
> >
> > Signed-off-by: Geliang Tang <geliangtang@xiaomi.com>
> > ---
> > net/mptcp/options.c | 16 ++++++++++++++++
> > net/mptcp/pm.c | 5 +++++
> > net/mptcp/protocol.h | 3 +++
> > 3 files changed, 24 insertions(+)
> >
> > diff --git a/net/mptcp/options.c b/net/mptcp/options.c
> > index 2b15063c8009..cd9ec4acf127 100644
> > --- a/net/mptcp/options.c
> > +++ b/net/mptcp/options.c
> > @@ -336,6 +336,16 @@ static void mptcp_parse_option(const struct sk_buff *skb,
> > mp_opt->reset_reason = *ptr;
> > break;
> >
> > + case MPTCPOPT_MP_FAIL:
> > + if (opsize != TCPOLEN_MPTCP_FAIL)
> > + break;
> > +
> > + ptr += 2;
> > + mp_opt->mp_fail = 1;
> > + mp_opt->fail_seq = get_unaligned_be64(ptr);
> > + pr_debug("MP_FAIL: data_seq=%llu", mp_opt->fail_seq);
> > + break;
> > +
> > default:
> > break;
> > }
> > @@ -364,6 +374,7 @@ void mptcp_get_options(const struct sock *sk,
> > mp_opt->reset = 0;
> > mp_opt->csum_reqd = READ_ONCE(msk->csum_enabled);
> > mp_opt->deny_join_id0 = 0;
> > + mp_opt->mp_fail = 0;
> >
> > length = (th->doff * 4) - sizeof(struct tcphdr);
> > ptr = (const unsigned char *)(th + 1);
> > @@ -1147,6 +1158,11 @@ bool mptcp_incoming_options(struct sock *sk, struct sk_buff *skb)
> > mp_opt.mp_prio = 0;
> > }
> >
> > + if (mp_opt.mp_fail) {
> > + mptcp_pm_mp_fail_received(sk, mp_opt.fail_seq);
> > + mp_opt.mp_fail = 0;
> > + }
> > +
>
> Side note not specifically related to this patch: usually we get a
> single MPTCP subopt per packet: a DSS. So we could optimize this code
> path with something alike:
>
> if (unlikely(any subopt other than dss is present))
> // go checking all of them individually
How about simply doing it like this:
if (unlikely(mp_opt.fastclose || mp_opt.add_addr || mp_opt.rm_addr ||
mp_opt.mp_prio || mp_opt.mp_fail || mp_opt.reset)) {
// go checking all of them individually
}
I just sent out a patch for this named "mptcp: use unlikely for non-DSS
suboptions" to ML. Please review it.
Thanks,
-Geliang
>
> To do the above we likely need to wrap all the 'mp_capable',
> 'fastclose', 'rm_addr' flags in a single bitmask. e.v. using a union.
>
> /P
>
^ permalink raw reply [flat|nested] 11+ messages in thread
end of thread, other threads:[~2021-08-12 7:09 UTC | newest]
Thread overview: 11+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-07-28 9:35 [MPTCP][PATCH v6 mptcp-next 0/5] MP_FAIL support Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 4/5] mptcp: add the mibs for MP_FAIL Geliang Tang
2021-07-28 9:35 ` [MPTCP][PATCH v6 mptcp-next 5/5] selftests: mptcp: add MP_FAIL mibs check Geliang Tang
2021-07-28 23:31 ` [MPTCP][PATCH v6 mptcp-next 3/5] mptcp: send out MP_FAIL when data checksum fails Mat Martineau
2021-07-28 10:36 ` [MPTCP][PATCH v6 mptcp-next 2/5] mptcp: MP_FAIL suboption receiving Paolo Abeni
2021-08-12 7:09 ` Geliang Tang
2021-07-28 10:31 ` [MPTCP][PATCH v6 mptcp-next 1/5] mptcp: MP_FAIL suboption sending Paolo Abeni
2021-07-28 10:37 ` [MPTCP][PATCH v6 mptcp-next 0/5] MP_FAIL support Paolo Abeni
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).