* [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
@ 2022-08-23 13:37 Liu Jian
2022-08-23 13:37 ` [PATCH bpf-next v2 1/2] net: " Liu Jian
` (3 more replies)
0 siblings, 4 replies; 9+ messages in thread
From: Liu Jian @ 2022-08-23 13:37 UTC (permalink / raw)
To: john.fastabend, jakub, edumazet, davem, yoshfuji, dsahern, kuba,
pabeni, andrii, mykolal, ast, daniel, martin.lau, song, yhs,
kpsingh, sdf, haoluo, jolsa, shuah, bpf
Cc: liujian56
If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
v1->v2:
As Jakub's suggested, check sock's DEAD flag before accessing
the wait queue.
Liu Jian (2):
net: If the sock is dead, do not access sock's sk_wq in
sk_stream_wait_memory
selftests/bpf: Add wait send memory test for sockmap redirect
net/core/stream.c | 3 +-
tools/testing/selftests/bpf/test_sockmap.c | 42 ++++++++++++++++++++++
2 files changed, 44 insertions(+), 1 deletion(-)
--
2.17.1
^ permalink raw reply [flat|nested] 9+ messages in thread
* [PATCH bpf-next v2 1/2] net: If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
2022-08-23 13:37 [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory Liu Jian
@ 2022-08-23 13:37 ` Liu Jian
2022-08-30 2:33 ` John Fastabend
2022-08-23 13:37 ` [PATCH bpf-next v2 2/2] selftests/bpf: Add wait send memory test for sockmap redirect Liu Jian
` (2 subsequent siblings)
3 siblings, 1 reply; 9+ messages in thread
From: Liu Jian @ 2022-08-23 13:37 UTC (permalink / raw)
To: john.fastabend, jakub, edumazet, davem, yoshfuji, dsahern, kuba,
pabeni, andrii, mykolal, ast, daniel, martin.lau, song, yhs,
kpsingh, sdf, haoluo, jolsa, shuah, bpf
Cc: liujian56
Fix the below NULL pointer dereference:
[ 14.471200] Call Trace:
[ 14.471562] <TASK>
[ 14.471882] lock_acquire+0x245/0x2e0
[ 14.472416] ? remove_wait_queue+0x12/0x50
[ 14.473014] ? _raw_spin_lock_irqsave+0x17/0x50
[ 14.473681] _raw_spin_lock_irqsave+0x3d/0x50
[ 14.474318] ? remove_wait_queue+0x12/0x50
[ 14.474907] remove_wait_queue+0x12/0x50
[ 14.475480] sk_stream_wait_memory+0x20d/0x340
[ 14.476127] ? do_wait_intr_irq+0x80/0x80
[ 14.476704] do_tcp_sendpages+0x287/0x600
[ 14.477283] tcp_bpf_push+0xab/0x260
[ 14.477817] tcp_bpf_sendmsg_redir+0x297/0x500
[ 14.478461] ? __local_bh_enable_ip+0x77/0xe0
[ 14.479096] tcp_bpf_send_verdict+0x105/0x470
[ 14.479729] tcp_bpf_sendmsg+0x318/0x4f0
[ 14.480311] sock_sendmsg+0x2d/0x40
[ 14.480822] ____sys_sendmsg+0x1b4/0x1c0
[ 14.481390] ? copy_msghdr_from_user+0x62/0x80
[ 14.482048] ___sys_sendmsg+0x78/0xb0
[ 14.482580] ? vmf_insert_pfn_prot+0x91/0x150
[ 14.483215] ? __do_fault+0x2a/0x1a0
[ 14.483738] ? do_fault+0x15e/0x5d0
[ 14.484246] ? __handle_mm_fault+0x56b/0x1040
[ 14.484874] ? lock_is_held_type+0xdf/0x130
[ 14.485474] ? find_held_lock+0x2d/0x90
[ 14.486046] ? __sys_sendmsg+0x41/0x70
[ 14.486587] __sys_sendmsg+0x41/0x70
[ 14.487105] ? intel_pmu_drain_pebs_core+0x350/0x350
[ 14.487822] do_syscall_64+0x34/0x80
[ 14.488345] entry_SYSCALL_64_after_hwframe+0x63/0xcd
The test scene as following flow:
thread1 thread2
----------- ---------------
tcp_bpf_sendmsg
tcp_bpf_send_verdict
tcp_bpf_sendmsg_redir sock_close
tcp_bpf_push_locked __sock_release
tcp_bpf_push //inet_release
do_tcp_sendpages sock->ops->release
sk_stream_wait_memory // tcp_close
sk_wait_event sk->sk_prot->close
release_sock(__sk);
***
lock_sock(sk);
__tcp_close
sock_orphan(sk)
sk->sk_wq = NULL
release_sock
****
lock_sock(__sk);
remove_wait_queue(sk_sleep(sk), &wait);
sk_sleep(sk)
//NULL pointer dereference
&rcu_dereference_raw(sk->sk_wq)->wait
While waiting for memory in thread1, the socket is released with its wait
queue because thread2 has closed it. This caused by tcp_bpf_send_verdict
didn't increase the f_count of psock->sk_redir->sk_socket->file in thread1.
We should check if SOCK_DEAD flag is set on wakeup in
sk_stream_wait_memory, before accessing the wait queue.
Suggested-by: Jakub Sitnicki <jakub@cloudflare.com>
Signed-off-by: Liu Jian <liujian56@huawei.com>
---
v1->v2:
As Jakub's suggested, check sock's DEAD flag before accessing
the wait queue.
net/core/stream.c | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
diff --git a/net/core/stream.c b/net/core/stream.c
index ccc083cdef23..1105057ce00a 100644
--- a/net/core/stream.c
+++ b/net/core/stream.c
@@ -159,7 +159,8 @@ int sk_stream_wait_memory(struct sock *sk, long *timeo_p)
*timeo_p = current_timeo;
}
out:
- remove_wait_queue(sk_sleep(sk), &wait);
+ if (!sock_flag(sk, SOCK_DEAD))
+ remove_wait_queue(sk_sleep(sk), &wait);
return err;
do_error:
--
2.17.1
^ permalink raw reply related [flat|nested] 9+ messages in thread
* [PATCH bpf-next v2 2/2] selftests/bpf: Add wait send memory test for sockmap redirect
2022-08-23 13:37 [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory Liu Jian
2022-08-23 13:37 ` [PATCH bpf-next v2 1/2] net: " Liu Jian
@ 2022-08-23 13:37 ` Liu Jian
2022-09-26 16:00 ` [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory patchwork-bot+netdevbpf
2022-10-27 10:36 ` Jakub Sitnicki
3 siblings, 0 replies; 9+ messages in thread
From: Liu Jian @ 2022-08-23 13:37 UTC (permalink / raw)
To: john.fastabend, jakub, edumazet, davem, yoshfuji, dsahern, kuba,
pabeni, andrii, mykolal, ast, daniel, martin.lau, song, yhs,
kpsingh, sdf, haoluo, jolsa, shuah, bpf
Cc: liujian56
Add one test for wait redirect sock's send memory test for sockmap.
Signed-off-by: Liu Jian <liujian56@huawei.com>
---
tools/testing/selftests/bpf/test_sockmap.c | 42 ++++++++++++++++++++++
1 file changed, 42 insertions(+)
diff --git a/tools/testing/selftests/bpf/test_sockmap.c b/tools/testing/selftests/bpf/test_sockmap.c
index 0fbaccdc8861..95b9b45ad028 100644
--- a/tools/testing/selftests/bpf/test_sockmap.c
+++ b/tools/testing/selftests/bpf/test_sockmap.c
@@ -138,6 +138,7 @@ struct sockmap_options {
bool data_test;
bool drop_expected;
bool check_recved_len;
+ bool tx_wait_mem;
int iov_count;
int iov_length;
int rate;
@@ -578,6 +579,10 @@ static int msg_loop(int fd, int iov_count, int iov_length, int cnt,
sent = sendmsg(fd, &msg, flags);
if (!drop && sent < 0) {
+ if (opt->tx_wait_mem && errno == EACCES) {
+ errno = 0;
+ goto out_errno;
+ }
perror("sendmsg loop error");
goto out_errno;
} else if (drop && sent >= 0) {
@@ -644,6 +649,15 @@ static int msg_loop(int fd, int iov_count, int iov_length, int cnt,
goto out_errno;
}
+ if (opt->tx_wait_mem) {
+ FD_ZERO(&w);
+ FD_SET(fd, &w);
+ slct = select(max_fd + 1, NULL, NULL, &w, &timeout);
+ errno = 0;
+ close(fd);
+ goto out_errno;
+ }
+
errno = 0;
if (peek_flag) {
flags |= MSG_PEEK;
@@ -752,6 +766,22 @@ static int sendmsg_test(struct sockmap_options *opt)
return err;
}
+ if (opt->tx_wait_mem) {
+ struct timeval timeout;
+ int rxtx_buf_len = 1024;
+
+ timeout.tv_sec = 3;
+ timeout.tv_usec = 0;
+
+ err = setsockopt(c2, SOL_SOCKET, SO_SNDTIMEO, &timeout, sizeof(struct timeval));
+ err |= setsockopt(c2, SOL_SOCKET, SO_SNDBUFFORCE, &rxtx_buf_len, sizeof(int));
+ err |= setsockopt(p2, SOL_SOCKET, SO_RCVBUFFORCE, &rxtx_buf_len, sizeof(int));
+ if (err) {
+ perror("setsockopt failed()");
+ return errno;
+ }
+ }
+
rxpid = fork();
if (rxpid == 0) {
if (txmsg_pop || txmsg_start_pop)
@@ -788,6 +818,9 @@ static int sendmsg_test(struct sockmap_options *opt)
return errno;
}
+ if (opt->tx_wait_mem)
+ close(c2);
+
txpid = fork();
if (txpid == 0) {
if (opt->sendpage)
@@ -1452,6 +1485,14 @@ static void test_txmsg_redir(int cgrp, struct sockmap_options *opt)
test_send(opt, cgrp);
}
+static void test_txmsg_redir_wait_sndmem(int cgrp, struct sockmap_options *opt)
+{
+ txmsg_redir = 1;
+ opt->tx_wait_mem = true;
+ test_send_large(opt, cgrp);
+ opt->tx_wait_mem = false;
+}
+
static void test_txmsg_drop(int cgrp, struct sockmap_options *opt)
{
txmsg_drop = 1;
@@ -1800,6 +1841,7 @@ static int populate_progs(char *bpf_file)
struct _test test[] = {
{"txmsg test passthrough", test_txmsg_pass},
{"txmsg test redirect", test_txmsg_redir},
+ {"txmsg test redirect wait send mem", test_txmsg_redir_wait_sndmem},
{"txmsg test drop", test_txmsg_drop},
{"txmsg test ingress redirect", test_txmsg_ingress_redir},
{"txmsg test skb", test_txmsg_skb},
--
2.17.1
^ permalink raw reply related [flat|nested] 9+ messages in thread
* RE: [PATCH bpf-next v2 1/2] net: If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
2022-08-23 13:37 ` [PATCH bpf-next v2 1/2] net: " Liu Jian
@ 2022-08-30 2:33 ` John Fastabend
0 siblings, 0 replies; 9+ messages in thread
From: John Fastabend @ 2022-08-30 2:33 UTC (permalink / raw)
To: Liu Jian, john.fastabend, jakub, edumazet, davem, yoshfuji,
dsahern, kuba, pabeni, andrii, mykolal, ast, daniel, martin.lau,
song, yhs, kpsingh, sdf, haoluo, jolsa, shuah, bpf
Cc: liujian56
Liu Jian wrote:
> Fix the below NULL pointer dereference:
>
> [ 14.471200] Call Trace:
> [ 14.471562] <TASK>
> [ 14.471882] lock_acquire+0x245/0x2e0
> [ 14.472416] ? remove_wait_queue+0x12/0x50
> [ 14.473014] ? _raw_spin_lock_irqsave+0x17/0x50
> [ 14.473681] _raw_spin_lock_irqsave+0x3d/0x50
> [ 14.474318] ? remove_wait_queue+0x12/0x50
> [ 14.474907] remove_wait_queue+0x12/0x50
> [ 14.475480] sk_stream_wait_memory+0x20d/0x340
> [ 14.476127] ? do_wait_intr_irq+0x80/0x80
> [ 14.476704] do_tcp_sendpages+0x287/0x600
> [ 14.477283] tcp_bpf_push+0xab/0x260
> [ 14.477817] tcp_bpf_sendmsg_redir+0x297/0x500
> [ 14.478461] ? __local_bh_enable_ip+0x77/0xe0
> [ 14.479096] tcp_bpf_send_verdict+0x105/0x470
> [ 14.479729] tcp_bpf_sendmsg+0x318/0x4f0
> [ 14.480311] sock_sendmsg+0x2d/0x40
> [ 14.480822] ____sys_sendmsg+0x1b4/0x1c0
> [ 14.481390] ? copy_msghdr_from_user+0x62/0x80
> [ 14.482048] ___sys_sendmsg+0x78/0xb0
> [ 14.482580] ? vmf_insert_pfn_prot+0x91/0x150
> [ 14.483215] ? __do_fault+0x2a/0x1a0
> [ 14.483738] ? do_fault+0x15e/0x5d0
> [ 14.484246] ? __handle_mm_fault+0x56b/0x1040
> [ 14.484874] ? lock_is_held_type+0xdf/0x130
> [ 14.485474] ? find_held_lock+0x2d/0x90
> [ 14.486046] ? __sys_sendmsg+0x41/0x70
> [ 14.486587] __sys_sendmsg+0x41/0x70
> [ 14.487105] ? intel_pmu_drain_pebs_core+0x350/0x350
> [ 14.487822] do_syscall_64+0x34/0x80
> [ 14.488345] entry_SYSCALL_64_after_hwframe+0x63/0xcd
>
> The test scene as following flow:
> thread1 thread2
> ----------- ---------------
> tcp_bpf_sendmsg
> tcp_bpf_send_verdict
> tcp_bpf_sendmsg_redir sock_close
> tcp_bpf_push_locked __sock_release
> tcp_bpf_push //inet_release
> do_tcp_sendpages sock->ops->release
> sk_stream_wait_memory // tcp_close
> sk_wait_event sk->sk_prot->close
> release_sock(__sk);
> ***
>
> lock_sock(sk);
> __tcp_close
> sock_orphan(sk)
> sk->sk_wq = NULL
> release_sock
> ****
> lock_sock(__sk);
> remove_wait_queue(sk_sleep(sk), &wait);
> sk_sleep(sk)
> //NULL pointer dereference
> &rcu_dereference_raw(sk->sk_wq)->wait
>
> While waiting for memory in thread1, the socket is released with its wait
> queue because thread2 has closed it. This caused by tcp_bpf_send_verdict
> didn't increase the f_count of psock->sk_redir->sk_socket->file in thread1.
>
> We should check if SOCK_DEAD flag is set on wakeup in
> sk_stream_wait_memory, before accessing the wait queue.
>
> Suggested-by: Jakub Sitnicki <jakub@cloudflare.com>
> Signed-off-by: Liu Jian <liujian56@huawei.com>
> ---
This LGTM but would be great if Eric could ACK as well from TCP side.
Acked-by: John Fastabend <john.fastabend@gmail.com>
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
2022-08-23 13:37 [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory Liu Jian
2022-08-23 13:37 ` [PATCH bpf-next v2 1/2] net: " Liu Jian
2022-08-23 13:37 ` [PATCH bpf-next v2 2/2] selftests/bpf: Add wait send memory test for sockmap redirect Liu Jian
@ 2022-09-26 16:00 ` patchwork-bot+netdevbpf
2022-10-27 10:36 ` Jakub Sitnicki
3 siblings, 0 replies; 9+ messages in thread
From: patchwork-bot+netdevbpf @ 2022-09-26 16:00 UTC (permalink / raw)
To: Liu Jian
Cc: john.fastabend, jakub, edumazet, davem, yoshfuji, dsahern, kuba,
pabeni, andrii, mykolal, ast, daniel, martin.lau, song, yhs,
kpsingh, sdf, haoluo, jolsa, shuah, bpf
Hello:
This series was applied to bpf/bpf-next.git (master)
by Daniel Borkmann <daniel@iogearbox.net>:
On Tue, 23 Aug 2022 21:37:53 +0800 you wrote:
> If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
>
> v1->v2:
> As Jakub's suggested, check sock's DEAD flag before accessing
> the wait queue.
>
> Liu Jian (2):
> net: If the sock is dead, do not access sock's sk_wq in
> sk_stream_wait_memory
> selftests/bpf: Add wait send memory test for sockmap redirect
>
> [...]
Here is the summary with links:
- [bpf-next,v2,1/2] net: If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
https://git.kernel.org/bpf/bpf-next/c/3f8ef65af927
- [bpf-next,v2,2/2] selftests/bpf: Add wait send memory test for sockmap redirect
https://git.kernel.org/bpf/bpf-next/c/043a7356dbd0
You are awesome, thank you!
--
Deet-doot-dot, I am a bot.
https://korg.docs.kernel.org/patchwork/pwbot.html
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
2022-08-23 13:37 [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory Liu Jian
` (2 preceding siblings ...)
2022-09-26 16:00 ` [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory patchwork-bot+netdevbpf
@ 2022-10-27 10:36 ` Jakub Sitnicki
2022-10-27 15:30 ` Jakub Sitnicki
3 siblings, 1 reply; 9+ messages in thread
From: Jakub Sitnicki @ 2022-10-27 10:36 UTC (permalink / raw)
To: Liu Jian, john.fastabend
Cc: edumazet, davem, yoshfuji, dsahern, kuba, pabeni, andrii,
mykolal, ast, daniel, martin.lau, song, yhs, kpsingh, sdf,
haoluo, jolsa, shuah, bpf
Liu, John,
On Tue, Aug 23, 2022 at 09:37 PM +08, Liu Jian wrote:
> If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
>
> v1->v2:
> As Jakub's suggested, check sock's DEAD flag before accessing
> the wait queue.
>
> Liu Jian (2):
> net: If the sock is dead, do not access sock's sk_wq in
> sk_stream_wait_memory
> selftests/bpf: Add wait send memory test for sockmap redirect
>
> net/core/stream.c | 3 +-
> tools/testing/selftests/bpf/test_sockmap.c | 42 ++++++++++++++++++++++
> 2 files changed, 44 insertions(+), 1 deletion(-)
While testing Cong's fix for the dead lock in sk_psock_backlog [1], I've
noticed that this change introduces a memory accounting issue. See
warnings below.
So what I've proposed is not completely sound. We will need to revisit
it.
[1] https://lore.kernel.org/bpf/Y0xJUc%2FLRu8K%2FAf8@pop-os.localdomain/
--8<--
bash-5.1# uname -r
6.0.0-rc3-00892-g3f8ef65af927
bash-5.1# ./test_sockmap
# 1/ 6 sockmap::txmsg test passthrough:OK
# 2/ 6 sockmap::txmsg test redirect:OK
# 3/ 1 sockmap::txmsg test redirect wait send mem:OK
# 4/ 6 sockmap::txmsg test drop:OK
# 5/ 6 sockmap::txmsg test ingress redirect:OK
# 6/ 7 sockmap::txmsg test skb:OK
# 7/ 8 sockmap::txmsg test apply:OK
# 8/12 sockmap::txmsg test cork:OK
[ 46.324023] ------------[ cut here ]------------
[ 46.325114] WARNING: CPU: 3 PID: 199 at net/core/stream.c:206 sk_stream_kill_queues+0xd6/0xf0
[ 46.326573] Modules linked in:
[ 46.327105] CPU: 3 PID: 199 Comm: test_sockmap Not tainted 6.0.0-rc3-00892-g3f8ef65af927 #36
[ 46.328406] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.15.0-1.fc35 04/01/2014
[ 46.330000] RIP: 0010:sk_stream_kill_queues+0xd6/0xf0
[ 46.330816] Code: 29 5b 5d 31 c0 89 c2 89 c6 89 c7 c3 48 89 df e8 10 14 ff ff 8b 83 70 02 00 00 8b b3 28 02 00 00 85 c0 74 d9 0f 0b 85 f6 74 d7 <0f> 0b 5b 5d 31 c0 89 c2 89 c6 89 c7 c3 0f 0b eb 92 66 0f 1f 84 00
[ 46.331889] RSP: 0018:ffffc90000bc7d48 EFLAGS: 00010206
[ 46.332186] RAX: 0000000000000000 RBX: ffff88810567dc00 RCX: 0000000000000000
[ 46.332583] RDX: 0000000000000000 RSI: 0000000000000fc0 RDI: ffff88810567ddb8
[ 46.332991] RBP: ffff88810567ddb8 R08: 0000000000000000 R09: 0000000000000000
[ 46.333321] R10: 0000000000000000 R11: 0000000000000000 R12: ffff88810567dc00
[ 46.333661] R13: ffff888103894500 R14: ffff888101fdf8e0 R15: ffff88810567dd30
[ 46.334074] FS: 00007f420a4d8b80(0000) GS:ffff88813bd80000(0000) knlGS:0000000000000000
[ 46.334532] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 46.334881] CR2: 00007f420a4d8af8 CR3: 0000000102e17001 CR4: 0000000000370ee0
[ 46.335300] Call Trace:
[ 46.335444] <TASK>
[ 46.335573] inet_csk_destroy_sock+0x4f/0x110
[ 46.335832] tcp_rcv_state_process+0xdcf/0x1140
[ 46.336076] ? tcp_v4_do_rcv+0x77/0x2a0
[ 46.336299] tcp_v4_do_rcv+0x77/0x2a0
[ 46.336512] __release_sock+0x58/0xb0
[ 46.336721] __tcp_close+0x186/0x450
[ 46.336883] tcp_close+0x20/0x70
[ 46.337026] inet_release+0x39/0x80
[ 46.337177] __sock_release+0x37/0xa0
[ 46.337341] sock_close+0x14/0x20
[ 46.337486] __fput+0xa2/0x260
[ 46.337622] task_work_run+0x59/0xa0
[ 46.337785] exit_to_user_mode_prepare+0x185/0x190
[ 46.337992] syscall_exit_to_user_mode+0x19/0x40
[ 46.338189] do_syscall_64+0x42/0x90
[ 46.338350] entry_SYSCALL_64_after_hwframe+0x46/0xb0
[ 46.338565] RIP: 0033:0x7f420a618eb7
[ 46.338741] Code: ff e8 7d e2 01 00 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 00 f3 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 03 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 41 c3 48 83 ec 18 89 7c 24 0c e8 43 cd f5 ff
[ 46.339596] RSP: 002b:00007ffd54748df8 EFLAGS: 00000246 ORIG_RAX: 0000000000000003
[ 46.339944] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 00007f420a618eb7
[ 46.340349] RDX: 0000000000000018 RSI: 00007ffd54748d50 RDI: 0000000000000019
[ 46.340753] RBP: 00007ffd54748e40 R08: 00007ffd54748d50 R09: 00007ffd54748d50
[ 46.341162] R10: 00007ffd54748d50 R11: 0000000000000246 R12: 00007ffd54749118
[ 46.341574] R13: 000000000040d0ad R14: 0000000000474df8 R15: 00007f420a769000
[ 46.341927] </TASK>
[ 46.342025] irq event stamp: 206385
[ 46.342175] hardirqs last enabled at (206393): [<ffffffff810f1f82>] __up_console_sem+0x52/0x60
[ 46.342546] hardirqs last disabled at (206400): [<ffffffff810f1f67>] __up_console_sem+0x37/0x60
[ 46.342965] softirqs last enabled at (206414): [<ffffffff8107bcc5>] __irq_exit_rcu+0xc5/0x120
[ 46.343454] softirqs last disabled at (206409): [<ffffffff8107bcc5>] __irq_exit_rcu+0xc5/0x120
[ 46.343951] ---[ end trace 0000000000000000 ]---
[ 46.344199] ------------[ cut here ]------------
[ 46.344421] WARNING: CPU: 3 PID: 199 at net/ipv4/af_inet.c:154 inet_sock_destruct+0x1a0/0x1d0
[ 46.344897] Modules linked in:
[ 46.345074] CPU: 3 PID: 199 Comm: test_sockmap Tainted: G W 6.0.0-rc3-00892-g3f8ef65af927 #36
[ 46.345549] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.15.0-1.fc35 04/01/2014
[ 46.346028] RIP: 0010:inet_sock_destruct+0x1a0/0x1d0
[ 46.346313] Code: ff 49 8b bc 24 60 02 00 00 e8 cc 1e e9 ff 49 8b bc 24 88 00 00 00 5b 41 5c e9 bc 1e e9 ff 41 8b 84 24 28 02 00 00 85 c0 74 ca <0f> 0b eb c6 4c 89 e7 e8 a4 2d e6 ff e9 50 ff ff ff 0f 0b 41 8b 84
[ 46.347370] RSP: 0018:ffffc90000bc7e40 EFLAGS: 00010206
[ 46.347670] RAX: 0000000000000fc0 RBX: ffff88810567dd60 RCX: 0000000000000000
[ 46.348090] RDX: 0000000000000303 RSI: 0000000000000fc0 RDI: ffff88810567dd60
[ 46.348495] RBP: ffff88810567dc00 R08: 0000000000000000 R09: 0000000000000000
[ 46.348921] R10: 0000000000000000 R11: 0000000000000000 R12: ffff88810567dc00
[ 46.349306] R13: ffff8881001d48e0 R14: ffff88810257a3a8 R15: 0000000000000000
[ 46.349717] FS: 00007f420a4d8b80(0000) GS:ffff88813bd80000(0000) knlGS:0000000000000000
[ 46.350191] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 46.350531] CR2: 00007f420a4d8af8 CR3: 0000000102e17001 CR4: 0000000000370ee0
[ 46.350947] Call Trace:
[ 46.351098] <TASK>
[ 46.351246] __sk_destruct+0x23/0x250
[ 46.351599] inet_release+0x39/0x80
[ 46.351835] __sock_release+0x37/0xa0
[ 46.352059] sock_close+0x14/0x20
[ 46.352206] __fput+0xa2/0x260
[ 46.352347] task_work_run+0x59/0xa0
[ 46.352515] exit_to_user_mode_prepare+0x185/0x190
[ 46.352728] syscall_exit_to_user_mode+0x19/0x40
[ 46.352954] do_syscall_64+0x42/0x90
[ 46.353119] entry_SYSCALL_64_after_hwframe+0x46/0xb0
[ 46.353342] RIP: 0033:0x7f420a618eb7
[ 46.353498] Code: ff e8 7d e2 01 00 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 00 f3 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 03 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 41 c3 48 83 ec 18 89 7c 24 0c e8 43 cd f5 ff
[ 46.354284] RSP: 002b:00007ffd54748df8 EFLAGS: 00000246 ORIG_RAX: 0000000000000003
[ 46.354601] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 00007f420a618eb7
[ 46.354908] RDX: 0000000000000018 RSI: 00007ffd54748d50 RDI: 0000000000000019
[ 46.355308] RBP: 00007ffd54748e40 R08: 00007ffd54748d50 R09: 00007ffd54748d50
[ 46.355710] R10: 00007ffd54748d50 R11: 0000000000000246 R12: 00007ffd54749118
[ 46.356117] R13: 000000000040d0ad R14: 0000000000474df8 R15: 00007f420a769000
[ 46.356532] </TASK>
[ 46.356664] irq event stamp: 206837
[ 46.356873] hardirqs last enabled at (206847): [<ffffffff810f1f82>] __up_console_sem+0x52/0x60
[ 46.357369] hardirqs last disabled at (206854): [<ffffffff810f1f67>] __up_console_sem+0x37/0x60
[ 46.357864] softirqs last enabled at (206604): [<ffffffff8107bcc5>] __irq_exit_rcu+0xc5/0x120
[ 46.358358] softirqs last disabled at (206591): [<ffffffff8107bcc5>] __irq_exit_rcu+0xc5/0x120
[ 46.358865] ---[ end trace 0000000000000000 ]---
# 9/ 3 sockmap::txmsg test hanging corks:OK
#10/11 sockmap::txmsg test push_data:OK
#11/17 sockmap::txmsg test pull-data:OK
#12/ 9 sockmap::txmsg test pop-data:OK
#13/ 1 sockmap::txmsg test push/pop data:OK
#14/ 1 sockmap::txmsg test ingress parser:OK
#15/ 1 sockmap::txmsg test ingress parser2:OK
#16/ 6 sockhash::txmsg test passthrough:OK
#17/ 6 sockhash::txmsg test redirect:OK
#18/ 1 sockhash::txmsg test redirect wait send mem:OK
#19/ 6 sockhash::txmsg test drop:OK
#20/ 6 sockhash::txmsg test ingress redirect:OK
#21/ 7 sockhash::txmsg test skb:OK
#22/ 8 sockhash::txmsg test apply:OK
#23/12 sockhash::txmsg test cork:OK
#24/ 3 sockhash::txmsg test hanging corks:OK
#25/11 sockhash::txmsg test push_data:OK
#26/17 sockhash::txmsg test pull-data:OK
#27/ 9 sockhash::txmsg test pop-data:OK
#28/ 1 sockhash::txmsg test push/pop data:OK
#29/ 1 sockhash::txmsg test ingress parser:OK
#30/ 1 sockhash::txmsg test ingress parser2:OK
#31/ 6 sockhash:ktls:txmsg test passthrough:OK
#32/ 6 sockhash:ktls:txmsg test redirect:OK
#33/ 1 sockhash:ktls:txmsg test redirect wait send mem:OK
#34/ 6 sockhash:ktls:txmsg test drop:OK
#35/ 6 sockhash:ktls:txmsg test ingress redirect:OK
#36/ 7 sockhash:ktls:txmsg test skb:OK
#37/ 8 sockhash:ktls:txmsg test apply:OK
#38/12 sockhash:ktls:txmsg test cork:OK
#39/ 3 sockhash:ktls:txmsg test hanging corks:OK
#40/11 sockhash:ktls:txmsg test push_data:OK
#41/17 sockhash:ktls:txmsg test pull-data:OK
#42/ 9 sockhash:ktls:txmsg test pop-data:OK
#43/ 1 sockhash:ktls:txmsg test push/pop data:OK
#44/ 1 sockhash:ktls:txmsg test ingress parser:OK
#45/ 0 sockhash:ktls:txmsg test ingress parser2:OK
Pass: 45 Fail: 0
bash-5.1# [ 61.124444] ------------[ cut here ]------------
[ 61.125177] page_counter underflow: -15 nr_pages=33
[ 61.125905] WARNING: CPU: 2 PID: 50 at mm/page_counter.c:56 page_counter_uncharge+0x6b/0x80
[ 61.127071] Modules linked in:
[ 61.127508] CPU: 2 PID: 50 Comm: kworker/2:1 Tainted: G W 6.0.0-rc3-00892-g3f8ef65af927 #36
[ 61.128638] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.15.0-1.fc35 04/01/2014
[ 61.129582] Workqueue: events bpf_prog_free_deferred
[ 61.130148] RIP: 0010:page_counter_uncharge+0x6b/0x80
[ 61.130736] Code: 5c 31 d2 89 d6 89 d7 41 89 d0 c3 80 3d 84 14 d7 01 00 75 18 48 89 ea 48 c7 c7 98 4f 2c 82 c6 05 71 14 d7 01 01 e8 dd 6f 74 00 <0f> 0b 48 c7 03 00 00 00 00 45 31 c0 eb b1 0f 1f 80 00 00 00 00 0f
[ 61.132430] RSP: 0018:ffffc900002cfd60 EFLAGS: 00010046
[ 61.132896] RAX: 0000000000000000 RBX: ffff888100280120 RCX: 0000000000000000
[ 61.133439] RDX: 0000000000000000 RSI: 0000000000000000 RDI: 0000000000000000
[ 61.133898] RBP: 0000000000000021 R08: 0000000000000000 R09: 0000000000000000
[ 61.134393] R10: 0000000000000000 R11: 0000000000000000 R12: ffffffffffffffdf
[ 61.134890] R13: ffff888100280000 R14: 0000000000000001 R15: ffff888101d8a140
[ 61.135382] FS: 0000000000000000(0000) GS:ffff88813bd00000(0000) knlGS:0000000000000000
[ 61.135939] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 61.136337] CR2: 000055e7f27926f0 CR3: 0000000002835006 CR4: 0000000000370ee0
[ 61.136721] Call Trace:
[ 61.136880] <TASK>
[ 61.137043] drain_stock+0x3b/0x70
[ 61.137285] refill_stock+0x89/0x150
[ 61.137547] refill_obj_stock+0x220/0x340
[ 61.137845] ? __bpf_prog_free+0x44/0x60
[ 61.138134] kfree+0x11e/0x540
[ 61.138304] __bpf_prog_free+0x44/0x60
[ 61.138533] process_one_work+0x238/0x570
[ 61.138837] ? process_one_work+0x570/0x570
[ 61.139133] worker_thread+0x55/0x3c0
[ 61.139394] ? process_one_work+0x570/0x570
[ 61.139635] kthread+0xea/0x110
[ 61.139821] ? kthread_complete_and_exit+0x20/0x20
[ 61.140098] ret_from_fork+0x1f/0x30
[ 61.140315] </TASK>
[ 61.140440] irq event stamp: 241492
[ 61.140616] hardirqs last enabled at (241491): [<ffffffff8130a1b6>] memcg_account_kmem+0x46/0x70
[ 61.141067] hardirqs last disabled at (241492): [<ffffffff81309c0a>] refill_stock+0xea/0x150
[ 61.141558] softirqs last enabled at (241024): [<ffffffff811497b6>] css_release_work_fn+0xc6/0x2a0
[ 61.142054] softirqs last disabled at (241022): [<ffffffff81149799>] css_release_work_fn+0xa9/0x2a0
[ 61.142551] ---[ end trace 0000000000000000 ]---
bash-5.1#
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
2022-10-27 10:36 ` Jakub Sitnicki
@ 2022-10-27 15:30 ` Jakub Sitnicki
2022-10-27 21:24 ` Jakub Sitnicki
0 siblings, 1 reply; 9+ messages in thread
From: Jakub Sitnicki @ 2022-10-27 15:30 UTC (permalink / raw)
To: Liu Jian, john.fastabend
Cc: edumazet, davem, yoshfuji, dsahern, kuba, pabeni, andrii,
mykolal, ast, daniel, martin.lau, song, yhs, kpsingh, sdf,
haoluo, jolsa, shuah, bpf
On Thu, Oct 27, 2022 at 12:36 PM +02, Jakub Sitnicki wrote:
[...]
> While testing Cong's fix for the dead lock in sk_psock_backlog [1], I've
> noticed that this change introduces a memory accounting issue. See
> warnings below.
>
> So what I've proposed is not completely sound. We will need to revisit
> it.
>
> [1] https://lore.kernel.org/bpf/Y0xJUc%2FLRu8K%2FAf8@pop-os.localdomain/
>
> --8<--
> bash-5.1# uname -r
> 6.0.0-rc3-00892-g3f8ef65af927
> bash-5.1# ./test_sockmap
> # 1/ 6 sockmap::txmsg test passthrough:OK
> # 2/ 6 sockmap::txmsg test redirect:OK
> # 3/ 1 sockmap::txmsg test redirect wait send mem:OK
> # 4/ 6 sockmap::txmsg test drop:OK
> # 5/ 6 sockmap::txmsg test ingress redirect:OK
> # 6/ 7 sockmap::txmsg test skb:OK
> # 7/ 8 sockmap::txmsg test apply:OK
> # 8/12 sockmap::txmsg test cork:OK
> [ 46.324023] ------------[ cut here ]------------
> [ 46.325114] WARNING: CPU: 3 PID: 199 at net/core/stream.c:206
> sk_stream_kill_queues+0xd6/0xf0
> [ 46.326573] Modules linked in:
> [ 46.327105] CPU: 3 PID: 199 Comm: test_sockmap Not tainted
> 6.0.0-rc3-00892-g3f8ef65af927 #36
> [ 46.328406] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS
> 1.15.0-1.fc35 04/01/2014
> [ 46.330000] RIP: 0010:sk_stream_kill_queues+0xd6/0xf0
> [ 46.330816] Code: 29 5b 5d 31 c0 89 c2 89 c6 89 c7 c3 48 89 df e8 10 14 ff ff
> 8b 83 70 02 00 00 8b b3 28 02 00 00 85 c0 74 d9 0f 0b 85 f6 74 d7 <0f> 0b 5b 5d
> 31 c0 89 c2 89 c6 89 c7 c3 0f 0b eb 92 66 0f 1f 84 00
> [ 46.331889] RSP: 0018:ffffc90000bc7d48 EFLAGS: 00010206
> [ 46.332186] RAX: 0000000000000000 RBX: ffff88810567dc00 RCX: 0000000000000000
> [ 46.332583] RDX: 0000000000000000 RSI: 0000000000000fc0 RDI: ffff88810567ddb8
> [ 46.332991] RBP: ffff88810567ddb8 R08: 0000000000000000 R09: 0000000000000000
> [ 46.333321] R10: 0000000000000000 R11: 0000000000000000 R12: ffff88810567dc00
> [ 46.333661] R13: ffff888103894500 R14: ffff888101fdf8e0 R15: ffff88810567dd30
> [ 46.334074] FS: 00007f420a4d8b80(0000) GS:ffff88813bd80000(0000) knlGS:0000000000000000
> [ 46.334532] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> [ 46.334881] CR2: 00007f420a4d8af8 CR3: 0000000102e17001 CR4: 0000000000370ee0
> [ 46.335300] Call Trace:
> [ 46.335444] <TASK>
> [ 46.335573] inet_csk_destroy_sock+0x4f/0x110
> [ 46.335832] tcp_rcv_state_process+0xdcf/0x1140
> [ 46.336076] ? tcp_v4_do_rcv+0x77/0x2a0
> [ 46.336299] tcp_v4_do_rcv+0x77/0x2a0
> [ 46.336512] __release_sock+0x58/0xb0
> [ 46.336721] __tcp_close+0x186/0x450
> [ 46.336883] tcp_close+0x20/0x70
> [ 46.337026] inet_release+0x39/0x80
> [ 46.337177] __sock_release+0x37/0xa0
> [ 46.337341] sock_close+0x14/0x20
> [ 46.337486] __fput+0xa2/0x260
> [ 46.337622] task_work_run+0x59/0xa0
> [ 46.337785] exit_to_user_mode_prepare+0x185/0x190
> [ 46.337992] syscall_exit_to_user_mode+0x19/0x40
> [ 46.338189] do_syscall_64+0x42/0x90
> [ 46.338350] entry_SYSCALL_64_after_hwframe+0x46/0xb0
> [ 46.338565] RIP: 0033:0x7f420a618eb7
> [ 46.338741] Code: ff e8 7d e2 01 00 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 00 f3
> 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 03 00 00 00 0f 05 <48> 3d 00 f0
> ff ff 77 41 c3 48 83 ec 18 89 7c 24 0c e8 43 cd f5 ff
> [ 46.339596] RSP: 002b:00007ffd54748df8 EFLAGS: 00000246 ORIG_RAX: 0000000000000003
> [ 46.339944] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 00007f420a618eb7
> [ 46.340349] RDX: 0000000000000018 RSI: 00007ffd54748d50 RDI: 0000000000000019
> [ 46.340753] RBP: 00007ffd54748e40 R08: 00007ffd54748d50 R09: 00007ffd54748d50
> [ 46.341162] R10: 00007ffd54748d50 R11: 0000000000000246 R12: 00007ffd54749118
> [ 46.341574] R13: 000000000040d0ad R14: 0000000000474df8 R15: 00007f420a769000
> [ 46.341927] </TASK>
> [ 46.342025] irq event stamp: 206385
> [ 46.342175] hardirqs last enabled at (206393): [<ffffffff810f1f82>]
> __up_console_sem+0x52/0x60
> [ 46.342546] hardirqs last disabled at (206400): [<ffffffff810f1f67>]
> __up_console_sem+0x37/0x60
> [ 46.342965] softirqs last enabled at (206414): [<ffffffff8107bcc5>]
> __irq_exit_rcu+0xc5/0x120
> [ 46.343454] softirqs last disabled at (206409): [<ffffffff8107bcc5>]
> __irq_exit_rcu+0xc5/0x120
> [ 46.343951] ---[ end trace 0000000000000000 ]---
> [ 46.344199] ------------[ cut here ]------------
> [ 46.344421] WARNING: CPU: 3 PID: 199 at net/ipv4/af_inet.c:154
> inet_sock_destruct+0x1a0/0x1d0
> [ 46.344897] Modules linked in:
> [ 46.345074] CPU: 3 PID: 199 Comm: test_sockmap Tainted: G W
> 6.0.0-rc3-00892-g3f8ef65af927 #36
> [ 46.345549] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS
> 1.15.0-1.fc35 04/01/2014
> [ 46.346028] RIP: 0010:inet_sock_destruct+0x1a0/0x1d0
> [ 46.346313] Code: ff 49 8b bc 24 60 02 00 00 e8 cc 1e e9 ff 49 8b bc 24 88 00
> 00 00 5b 41 5c e9 bc 1e e9 ff 41 8b 84 24 28 02 00 00 85 c0 74 ca <0f> 0b eb c6
> 4c 89 e7 e8 a4 2d e6 ff e9 50 ff ff ff 0f 0b 41 8b 84
> [ 46.347370] RSP: 0018:ffffc90000bc7e40 EFLAGS: 00010206
> [ 46.347670] RAX: 0000000000000fc0 RBX: ffff88810567dd60 RCX: 0000000000000000
> [ 46.348090] RDX: 0000000000000303 RSI: 0000000000000fc0 RDI: ffff88810567dd60
> [ 46.348495] RBP: ffff88810567dc00 R08: 0000000000000000 R09: 0000000000000000
> [ 46.348921] R10: 0000000000000000 R11: 0000000000000000 R12: ffff88810567dc00
> [ 46.349306] R13: ffff8881001d48e0 R14: ffff88810257a3a8 R15: 0000000000000000
> [ 46.349717] FS: 00007f420a4d8b80(0000) GS:ffff88813bd80000(0000) knlGS:0000000000000000
> [ 46.350191] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> [ 46.350531] CR2: 00007f420a4d8af8 CR3: 0000000102e17001 CR4: 0000000000370ee0
> [ 46.350947] Call Trace:
> [ 46.351098] <TASK>
> [ 46.351246] __sk_destruct+0x23/0x250
> [ 46.351599] inet_release+0x39/0x80
> [ 46.351835] __sock_release+0x37/0xa0
> [ 46.352059] sock_close+0x14/0x20
> [ 46.352206] __fput+0xa2/0x260
> [ 46.352347] task_work_run+0x59/0xa0
> [ 46.352515] exit_to_user_mode_prepare+0x185/0x190
> [ 46.352728] syscall_exit_to_user_mode+0x19/0x40
> [ 46.352954] do_syscall_64+0x42/0x90
> [ 46.353119] entry_SYSCALL_64_after_hwframe+0x46/0xb0
> [ 46.353342] RIP: 0033:0x7f420a618eb7
> [ 46.353498] Code: ff e8 7d e2 01 00 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 00 f3
> 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 03 00 00 00 0f 05 <48> 3d 00 f0
> ff ff 77 41 c3 48 83 ec 18 89 7c 24 0c e8 43 cd f5 ff
> [ 46.354284] RSP: 002b:00007ffd54748df8 EFLAGS: 00000246 ORIG_RAX: 0000000000000003
> [ 46.354601] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 00007f420a618eb7
> [ 46.354908] RDX: 0000000000000018 RSI: 00007ffd54748d50 RDI: 0000000000000019
> [ 46.355308] RBP: 00007ffd54748e40 R08: 00007ffd54748d50 R09: 00007ffd54748d50
> [ 46.355710] R10: 00007ffd54748d50 R11: 0000000000000246 R12: 00007ffd54749118
> [ 46.356117] R13: 000000000040d0ad R14: 0000000000474df8 R15: 00007f420a769000
> [ 46.356532] </TASK>
> [ 46.356664] irq event stamp: 206837
> [ 46.356873] hardirqs last enabled at (206847): [<ffffffff810f1f82>]
> __up_console_sem+0x52/0x60
> [ 46.357369] hardirqs last disabled at (206854): [<ffffffff810f1f67>]
> __up_console_sem+0x37/0x60
> [ 46.357864] softirqs last enabled at (206604): [<ffffffff8107bcc5>]
> __irq_exit_rcu+0xc5/0x120
> [ 46.358358] softirqs last disabled at (206591): [<ffffffff8107bcc5>]
> __irq_exit_rcu+0xc5/0x120
> [ 46.358865] ---[ end trace 0000000000000000 ]---
> # 9/ 3 sockmap::txmsg test hanging corks:OK
[...]
Actually, we had a fix for these warnings in 9c34e38c4a87 ("bpf,
sockmap: Fix memleak in tcp_bpf_sendmsg while sk msg is full") [1].
But they reappear for me in v5.18-rc1. Trying to bisect what went
wrong...
[1] https://lore.kernel.org/bpf/20220304081145.2037182-3-wangyufen@huawei.com
^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
2022-10-27 15:30 ` Jakub Sitnicki
@ 2022-10-27 21:24 ` Jakub Sitnicki
2022-10-28 2:56 ` 答复: " liujian (CE)
0 siblings, 1 reply; 9+ messages in thread
From: Jakub Sitnicki @ 2022-10-27 21:24 UTC (permalink / raw)
To: Liu Jian, john.fastabend
Cc: edumazet, davem, yoshfuji, dsahern, kuba, pabeni, andrii,
mykolal, ast, daniel, martin.lau, song, yhs, kpsingh, sdf,
haoluo, jolsa, shuah, bpf
On Thu, Oct 27, 2022 at 05:30 PM +02, Jakub Sitnicki wrote:
> On Thu, Oct 27, 2022 at 12:36 PM +02, Jakub Sitnicki wrote:
>
> [...]
>
>> While testing Cong's fix for the dead lock in sk_psock_backlog [1], I've
>> noticed that this change introduces a memory accounting issue. See
>> warnings below.
>>
>> So what I've proposed is not completely sound. We will need to revisit
>> it.
>>
>> [1] https://lore.kernel.org/bpf/Y0xJUc%2FLRu8K%2FAf8@pop-os.localdomain/
>>
>> --8<--
>> bash-5.1# uname -r
>> 6.0.0-rc3-00892-g3f8ef65af927
>> bash-5.1# ./test_sockmap
>> # 1/ 6 sockmap::txmsg test passthrough:OK
>> # 2/ 6 sockmap::txmsg test redirect:OK
>> # 3/ 1 sockmap::txmsg test redirect wait send mem:OK
>> # 4/ 6 sockmap::txmsg test drop:OK
>> # 5/ 6 sockmap::txmsg test ingress redirect:OK
>> # 6/ 7 sockmap::txmsg test skb:OK
>> # 7/ 8 sockmap::txmsg test apply:OK
>> # 8/12 sockmap::txmsg test cork:OK
>> [ 46.324023] ------------[ cut here ]------------
>> [ 46.325114] WARNING: CPU: 3 PID: 199 at net/core/stream.c:206
>> sk_stream_kill_queues+0xd6/0xf0
>> [ 46.326573] Modules linked in:
>> [ 46.327105] CPU: 3 PID: 199 Comm: test_sockmap Not tainted
>> 6.0.0-rc3-00892-g3f8ef65af927 #36
>> [ 46.328406] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS
>> 1.15.0-1.fc35 04/01/2014
>> [ 46.330000] RIP: 0010:sk_stream_kill_queues+0xd6/0xf0
>> [ 46.330816] Code: 29 5b 5d 31 c0 89 c2 89 c6 89 c7 c3 48 89 df e8 10 14 ff ff
>> 8b 83 70 02 00 00 8b b3 28 02 00 00 85 c0 74 d9 0f 0b 85 f6 74 d7 <0f> 0b 5b 5d
>> 31 c0 89 c2 89 c6 89 c7 c3 0f 0b eb 92 66 0f 1f 84 00
>> [ 46.331889] RSP: 0018:ffffc90000bc7d48 EFLAGS: 00010206
>> [ 46.332186] RAX: 0000000000000000 RBX: ffff88810567dc00 RCX: 0000000000000000
>> [ 46.332583] RDX: 0000000000000000 RSI: 0000000000000fc0 RDI: ffff88810567ddb8
>> [ 46.332991] RBP: ffff88810567ddb8 R08: 0000000000000000 R09: 0000000000000000
>> [ 46.333321] R10: 0000000000000000 R11: 0000000000000000 R12: ffff88810567dc00
>> [ 46.333661] R13: ffff888103894500 R14: ffff888101fdf8e0 R15: ffff88810567dd30
>> [ 46.334074] FS: 00007f420a4d8b80(0000) GS:ffff88813bd80000(0000) knlGS:0000000000000000
>> [ 46.334532] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
>> [ 46.334881] CR2: 00007f420a4d8af8 CR3: 0000000102e17001 CR4: 0000000000370ee0
>> [ 46.335300] Call Trace:
>> [ 46.335444] <TASK>
>> [ 46.335573] inet_csk_destroy_sock+0x4f/0x110
>> [ 46.335832] tcp_rcv_state_process+0xdcf/0x1140
>> [ 46.336076] ? tcp_v4_do_rcv+0x77/0x2a0
>> [ 46.336299] tcp_v4_do_rcv+0x77/0x2a0
>> [ 46.336512] __release_sock+0x58/0xb0
>> [ 46.336721] __tcp_close+0x186/0x450
>> [ 46.336883] tcp_close+0x20/0x70
>> [ 46.337026] inet_release+0x39/0x80
>> [ 46.337177] __sock_release+0x37/0xa0
>> [ 46.337341] sock_close+0x14/0x20
>> [ 46.337486] __fput+0xa2/0x260
>> [ 46.337622] task_work_run+0x59/0xa0
>> [ 46.337785] exit_to_user_mode_prepare+0x185/0x190
>> [ 46.337992] syscall_exit_to_user_mode+0x19/0x40
>> [ 46.338189] do_syscall_64+0x42/0x90
>> [ 46.338350] entry_SYSCALL_64_after_hwframe+0x46/0xb0
>> [ 46.338565] RIP: 0033:0x7f420a618eb7
>> [ 46.338741] Code: ff e8 7d e2 01 00 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 00 f3
>> 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 03 00 00 00 0f 05 <48> 3d 00 f0
>> ff ff 77 41 c3 48 83 ec 18 89 7c 24 0c e8 43 cd f5 ff
>> [ 46.339596] RSP: 002b:00007ffd54748df8 EFLAGS: 00000246 ORIG_RAX: 0000000000000003
>> [ 46.339944] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 00007f420a618eb7
>> [ 46.340349] RDX: 0000000000000018 RSI: 00007ffd54748d50 RDI: 0000000000000019
>> [ 46.340753] RBP: 00007ffd54748e40 R08: 00007ffd54748d50 R09: 00007ffd54748d50
>> [ 46.341162] R10: 00007ffd54748d50 R11: 0000000000000246 R12: 00007ffd54749118
>> [ 46.341574] R13: 000000000040d0ad R14: 0000000000474df8 R15: 00007f420a769000
>> [ 46.341927] </TASK>
>> [ 46.342025] irq event stamp: 206385
>> [ 46.342175] hardirqs last enabled at (206393): [<ffffffff810f1f82>]
>> __up_console_sem+0x52/0x60
>> [ 46.342546] hardirqs last disabled at (206400): [<ffffffff810f1f67>]
>> __up_console_sem+0x37/0x60
>> [ 46.342965] softirqs last enabled at (206414): [<ffffffff8107bcc5>]
>> __irq_exit_rcu+0xc5/0x120
>> [ 46.343454] softirqs last disabled at (206409): [<ffffffff8107bcc5>]
>> __irq_exit_rcu+0xc5/0x120
>> [ 46.343951] ---[ end trace 0000000000000000 ]---
>> [ 46.344199] ------------[ cut here ]------------
>> [ 46.344421] WARNING: CPU: 3 PID: 199 at net/ipv4/af_inet.c:154
>> inet_sock_destruct+0x1a0/0x1d0
>> [ 46.344897] Modules linked in:
>> [ 46.345074] CPU: 3 PID: 199 Comm: test_sockmap Tainted: G W
>> 6.0.0-rc3-00892-g3f8ef65af927 #36
>> [ 46.345549] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS
>> 1.15.0-1.fc35 04/01/2014
>> [ 46.346028] RIP: 0010:inet_sock_destruct+0x1a0/0x1d0
>> [ 46.346313] Code: ff 49 8b bc 24 60 02 00 00 e8 cc 1e e9 ff 49 8b bc 24 88 00
>> 00 00 5b 41 5c e9 bc 1e e9 ff 41 8b 84 24 28 02 00 00 85 c0 74 ca <0f> 0b eb c6
>> 4c 89 e7 e8 a4 2d e6 ff e9 50 ff ff ff 0f 0b 41 8b 84
>> [ 46.347370] RSP: 0018:ffffc90000bc7e40 EFLAGS: 00010206
>> [ 46.347670] RAX: 0000000000000fc0 RBX: ffff88810567dd60 RCX: 0000000000000000
>> [ 46.348090] RDX: 0000000000000303 RSI: 0000000000000fc0 RDI: ffff88810567dd60
>> [ 46.348495] RBP: ffff88810567dc00 R08: 0000000000000000 R09: 0000000000000000
>> [ 46.348921] R10: 0000000000000000 R11: 0000000000000000 R12: ffff88810567dc00
>> [ 46.349306] R13: ffff8881001d48e0 R14: ffff88810257a3a8 R15: 0000000000000000
>> [ 46.349717] FS: 00007f420a4d8b80(0000) GS:ffff88813bd80000(0000) knlGS:0000000000000000
>> [ 46.350191] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
>> [ 46.350531] CR2: 00007f420a4d8af8 CR3: 0000000102e17001 CR4: 0000000000370ee0
>> [ 46.350947] Call Trace:
>> [ 46.351098] <TASK>
>> [ 46.351246] __sk_destruct+0x23/0x250
>> [ 46.351599] inet_release+0x39/0x80
>> [ 46.351835] __sock_release+0x37/0xa0
>> [ 46.352059] sock_close+0x14/0x20
>> [ 46.352206] __fput+0xa2/0x260
>> [ 46.352347] task_work_run+0x59/0xa0
>> [ 46.352515] exit_to_user_mode_prepare+0x185/0x190
>> [ 46.352728] syscall_exit_to_user_mode+0x19/0x40
>> [ 46.352954] do_syscall_64+0x42/0x90
>> [ 46.353119] entry_SYSCALL_64_after_hwframe+0x46/0xb0
>> [ 46.353342] RIP: 0033:0x7f420a618eb7
>> [ 46.353498] Code: ff e8 7d e2 01 00 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 00 f3
>> 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 03 00 00 00 0f 05 <48> 3d 00 f0
>> ff ff 77 41 c3 48 83 ec 18 89 7c 24 0c e8 43 cd f5 ff
>> [ 46.354284] RSP: 002b:00007ffd54748df8 EFLAGS: 00000246 ORIG_RAX: 0000000000000003
>> [ 46.354601] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 00007f420a618eb7
>> [ 46.354908] RDX: 0000000000000018 RSI: 00007ffd54748d50 RDI: 0000000000000019
>> [ 46.355308] RBP: 00007ffd54748e40 R08: 00007ffd54748d50 R09: 00007ffd54748d50
>> [ 46.355710] R10: 00007ffd54748d50 R11: 0000000000000246 R12: 00007ffd54749118
>> [ 46.356117] R13: 000000000040d0ad R14: 0000000000474df8 R15: 00007f420a769000
>> [ 46.356532] </TASK>
>> [ 46.356664] irq event stamp: 206837
>> [ 46.356873] hardirqs last enabled at (206847): [<ffffffff810f1f82>]
>> __up_console_sem+0x52/0x60
>> [ 46.357369] hardirqs last disabled at (206854): [<ffffffff810f1f67>]
>> __up_console_sem+0x37/0x60
>> [ 46.357864] softirqs last enabled at (206604): [<ffffffff8107bcc5>]
>> __irq_exit_rcu+0xc5/0x120
>> [ 46.358358] softirqs last disabled at (206591): [<ffffffff8107bcc5>]
>> __irq_exit_rcu+0xc5/0x120
>> [ 46.358865] ---[ end trace 0000000000000000 ]---
>> # 9/ 3 sockmap::txmsg test hanging corks:OK
>
> [...]
>
> Actually, we had a fix for these warnings in 9c34e38c4a87 ("bpf,
> sockmap: Fix memleak in tcp_bpf_sendmsg while sk msg is full") [1].
>
> But they reappear for me in v5.18-rc1. Trying to bisect what went
> wrong...
>
> [1] https://lore.kernel.org/bpf/20220304081145.2037182-3-wangyufen@huawei.com
Traced it down to 84472b436e76 ("bpf, sockmap: Fix more uncharged while
msg has more_data"), which was the next commit after the one I thought
contained the fix, that is 9c34e38c4a87.
I started a discussion in the thread for the patch set [1], as the
warnings look completely unrelated to this change.
[1] https://lore.kernel.org/netdev/87v8o5gdw2.fsf@cloudflare.com/
^ permalink raw reply [flat|nested] 9+ messages in thread
* 答复: [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory
2022-10-27 21:24 ` Jakub Sitnicki
@ 2022-10-28 2:56 ` liujian (CE)
0 siblings, 0 replies; 9+ messages in thread
From: liujian (CE) @ 2022-10-28 2:56 UTC (permalink / raw)
To: Jakub Sitnicki, john.fastabend
Cc: edumazet, davem, yoshfuji, dsahern, kuba, pabeni, andrii,
mykolal, ast, daniel, martin.lau, song, yhs, kpsingh, sdf,
haoluo, jolsa, shuah, bpf
> -----邮件原件-----
> 发件人: Jakub Sitnicki <jakub@cloudflare.com>
> 发送时间: 2022年10月28日 5:24
> 收件人: liujian (CE) <liujian56@huawei.com>; john.fastabend@gmail.com
> 抄送: edumazet@google.com; davem@davemloft.net;
> yoshfuji@linux-ipv6.org; dsahern@kernel.org; kuba@kernel.org;
> pabeni@redhat.com; andrii@kernel.org; mykolal@fb.com; ast@kernel.org;
> daniel@iogearbox.net; martin.lau@linux.dev; song@kernel.org; yhs@fb.com;
> kpsingh@kernel.org; sdf@google.com; haoluo@google.com; jolsa@kernel.org;
> shuah@kernel.org; bpf@vger.kernel.org
> 主题: Re: [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's
> sk_wq in sk_stream_wait_memory
>
> On Thu, Oct 27, 2022 at 05:30 PM +02, Jakub Sitnicki wrote:
> > On Thu, Oct 27, 2022 at 12:36 PM +02, Jakub Sitnicki wrote:
> >
> > [...]
> >
> >> While testing Cong's fix for the dead lock in sk_psock_backlog [1],
> >> I've noticed that this change introduces a memory accounting issue.
> >> See warnings below.
> >>
> >> So what I've proposed is not completely sound. We will need to
> >> revisit it.
> >>
> >> [1]
> >>
> https://lore.kernel.org/bpf/Y0xJUc%2FLRu8K%2FAf8@pop-os.localdomain/
> >>
> >> --8<--
> >> bash-5.1# uname -r
> >> 6.0.0-rc3-00892-g3f8ef65af927
> >> bash-5.1# ./test_sockmap
> >> # 1/ 6 sockmap::txmsg test passthrough:OK # 2/ 6 sockmap::txmsg
> >> test redirect:OK # 3/ 1 sockmap::txmsg test redirect wait send
> >> mem:OK # 4/ 6 sockmap::txmsg test drop:OK # 5/ 6 sockmap::txmsg
> >> test ingress redirect:OK # 6/ 7 sockmap::txmsg test skb:OK # 7/ 8
> >> sockmap::txmsg test apply:OK # 8/12 sockmap::txmsg test cork:OK
> >> [ 46.324023] ------------[ cut here ]------------
> >> [ 46.325114] WARNING: CPU: 3 PID: 199 at net/core/stream.c:206
> >> sk_stream_kill_queues+0xd6/0xf0
> >> [ 46.326573] Modules linked in:
> >> [ 46.327105] CPU: 3 PID: 199 Comm: test_sockmap Not tainted
> >> 6.0.0-rc3-00892-g3f8ef65af927 #36
> >> [ 46.328406] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996),
> >> BIOS
> >> 1.15.0-1.fc35 04/01/2014
> >> [ 46.330000] RIP: 0010:sk_stream_kill_queues+0xd6/0xf0
> >> [ 46.330816] Code: 29 5b 5d 31 c0 89 c2 89 c6 89 c7 c3 48 89 df e8 10
> >> 14 ff ff 8b 83 70 02 00 00 8b b3 28 02 00 00 85 c0 74 d9 0f 0b 85 f6
> >> 74 d7 <0f> 0b 5b 5d
> >> 31 c0 89 c2 89 c6 89 c7 c3 0f 0b eb 92 66 0f 1f 84 00
> >> [ 46.331889] RSP: 0018:ffffc90000bc7d48 EFLAGS: 00010206
> >> [ 46.332186] RAX: 0000000000000000 RBX: ffff88810567dc00 RCX:
> 0000000000000000
> >> [ 46.332583] RDX: 0000000000000000 RSI: 0000000000000fc0 RDI:
> ffff88810567ddb8
> >> [ 46.332991] RBP: ffff88810567ddb8 R08: 0000000000000000 R09:
> 0000000000000000
> >> [ 46.333321] R10: 0000000000000000 R11: 0000000000000000 R12:
> ffff88810567dc00
> >> [ 46.333661] R13: ffff888103894500 R14: ffff888101fdf8e0 R15:
> ffff88810567dd30
> >> [ 46.334074] FS: 00007f420a4d8b80(0000)
> GS:ffff88813bd80000(0000) knlGS:0000000000000000
> >> [ 46.334532] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> >> [ 46.334881] CR2: 00007f420a4d8af8 CR3: 0000000102e17001 CR4:
> 0000000000370ee0
> >> [ 46.335300] Call Trace:
> >> [ 46.335444] <TASK>
> >> [ 46.335573] inet_csk_destroy_sock+0x4f/0x110
> >> [ 46.335832] tcp_rcv_state_process+0xdcf/0x1140
> >> [ 46.336076] ? tcp_v4_do_rcv+0x77/0x2a0
> >> [ 46.336299] tcp_v4_do_rcv+0x77/0x2a0
> >> [ 46.336512] __release_sock+0x58/0xb0
> >> [ 46.336721] __tcp_close+0x186/0x450
> >> [ 46.336883] tcp_close+0x20/0x70
> >> [ 46.337026] inet_release+0x39/0x80
> >> [ 46.337177] __sock_release+0x37/0xa0
> >> [ 46.337341] sock_close+0x14/0x20
> >> [ 46.337486] __fput+0xa2/0x260
> >> [ 46.337622] task_work_run+0x59/0xa0
> >> [ 46.337785] exit_to_user_mode_prepare+0x185/0x190
> >> [ 46.337992] syscall_exit_to_user_mode+0x19/0x40
> >> [ 46.338189] do_syscall_64+0x42/0x90
> >> [ 46.338350] entry_SYSCALL_64_after_hwframe+0x46/0xb0
> >> [ 46.338565] RIP: 0033:0x7f420a618eb7
> >> [ 46.338741] Code: ff e8 7d e2 01 00 66 2e 0f 1f 84 00 00 00 00 00 0f
> >> 1f 00 f3 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 03 00 00 00
> >> 0f 05 <48> 3d 00 f0 ff ff 77 41 c3 48 83 ec 18 89 7c 24 0c e8 43 cd f5 ff
> >> [ 46.339596] RSP: 002b:00007ffd54748df8 EFLAGS: 00000246
> ORIG_RAX: 0000000000000003
> >> [ 46.339944] RAX: 0000000000000000 RBX: 0000000000000000 RCX:
> 00007f420a618eb7
> >> [ 46.340349] RDX: 0000000000000018 RSI: 00007ffd54748d50 RDI:
> 0000000000000019
> >> [ 46.340753] RBP: 00007ffd54748e40 R08: 00007ffd54748d50 R09:
> 00007ffd54748d50
> >> [ 46.341162] R10: 00007ffd54748d50 R11: 0000000000000246 R12:
> 00007ffd54749118
> >> [ 46.341574] R13: 000000000040d0ad R14: 0000000000474df8 R15:
> 00007f420a769000
> >> [ 46.341927] </TASK>
> >> [ 46.342025] irq event stamp: 206385
> >> [ 46.342175] hardirqs last enabled at (206393): [<ffffffff810f1f82>]
> >> __up_console_sem+0x52/0x60
> >> [ 46.342546] hardirqs last disabled at (206400): [<ffffffff810f1f67>]
> >> __up_console_sem+0x37/0x60
> >> [ 46.342965] softirqs last enabled at (206414): [<ffffffff8107bcc5>]
> >> __irq_exit_rcu+0xc5/0x120
> >> [ 46.343454] softirqs last disabled at (206409): [<ffffffff8107bcc5>]
> >> __irq_exit_rcu+0xc5/0x120
> >> [ 46.343951] ---[ end trace 0000000000000000 ]---
> >> [ 46.344199] ------------[ cut here ]------------
> >> [ 46.344421] WARNING: CPU: 3 PID: 199 at net/ipv4/af_inet.c:154
> >> inet_sock_destruct+0x1a0/0x1d0
> >> [ 46.344897] Modules linked in:
> >> [ 46.345074] CPU: 3 PID: 199 Comm: test_sockmap Tainted: G W
> >> 6.0.0-rc3-00892-g3f8ef65af927 #36
> >> [ 46.345549] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996),
> >> BIOS
> >> 1.15.0-1.fc35 04/01/2014
> >> [ 46.346028] RIP: 0010:inet_sock_destruct+0x1a0/0x1d0
> >> [ 46.346313] Code: ff 49 8b bc 24 60 02 00 00 e8 cc 1e e9 ff 49 8b bc
> >> 24 88 00
> >> 00 00 5b 41 5c e9 bc 1e e9 ff 41 8b 84 24 28 02 00 00 85 c0 74 ca
> >> <0f> 0b eb c6 4c 89 e7 e8 a4 2d e6 ff e9 50 ff ff ff 0f 0b 41 8b 84
> >> [ 46.347370] RSP: 0018:ffffc90000bc7e40 EFLAGS: 00010206
> >> [ 46.347670] RAX: 0000000000000fc0 RBX: ffff88810567dd60 RCX:
> 0000000000000000
> >> [ 46.348090] RDX: 0000000000000303 RSI: 0000000000000fc0 RDI:
> ffff88810567dd60
> >> [ 46.348495] RBP: ffff88810567dc00 R08: 0000000000000000 R09:
> 0000000000000000
> >> [ 46.348921] R10: 0000000000000000 R11: 0000000000000000 R12:
> ffff88810567dc00
> >> [ 46.349306] R13: ffff8881001d48e0 R14: ffff88810257a3a8 R15:
> 0000000000000000
> >> [ 46.349717] FS: 00007f420a4d8b80(0000)
> GS:ffff88813bd80000(0000) knlGS:0000000000000000
> >> [ 46.350191] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> >> [ 46.350531] CR2: 00007f420a4d8af8 CR3: 0000000102e17001 CR4:
> 0000000000370ee0
> >> [ 46.350947] Call Trace:
> >> [ 46.351098] <TASK>
> >> [ 46.351246] __sk_destruct+0x23/0x250
> >> [ 46.351599] inet_release+0x39/0x80
> >> [ 46.351835] __sock_release+0x37/0xa0
> >> [ 46.352059] sock_close+0x14/0x20
> >> [ 46.352206] __fput+0xa2/0x260
> >> [ 46.352347] task_work_run+0x59/0xa0
> >> [ 46.352515] exit_to_user_mode_prepare+0x185/0x190
> >> [ 46.352728] syscall_exit_to_user_mode+0x19/0x40
> >> [ 46.352954] do_syscall_64+0x42/0x90
> >> [ 46.353119] entry_SYSCALL_64_after_hwframe+0x46/0xb0
> >> [ 46.353342] RIP: 0033:0x7f420a618eb7
> >> [ 46.353498] Code: ff e8 7d e2 01 00 66 2e 0f 1f 84 00 00 00 00 00 0f
> >> 1f 00 f3 0f 1e fa 64 8b 04 25 18 00 00 00 85 c0 75 10 b8 03 00 00 00
> >> 0f 05 <48> 3d 00 f0 ff ff 77 41 c3 48 83 ec 18 89 7c 24 0c e8 43 cd f5 ff
> >> [ 46.354284] RSP: 002b:00007ffd54748df8 EFLAGS: 00000246
> ORIG_RAX: 0000000000000003
> >> [ 46.354601] RAX: 0000000000000000 RBX: 0000000000000000 RCX:
> 00007f420a618eb7
> >> [ 46.354908] RDX: 0000000000000018 RSI: 00007ffd54748d50 RDI:
> 0000000000000019
> >> [ 46.355308] RBP: 00007ffd54748e40 R08: 00007ffd54748d50 R09:
> 00007ffd54748d50
> >> [ 46.355710] R10: 00007ffd54748d50 R11: 0000000000000246 R12:
> 00007ffd54749118
> >> [ 46.356117] R13: 000000000040d0ad R14: 0000000000474df8 R15:
> 00007f420a769000
> >> [ 46.356532] </TASK>
> >> [ 46.356664] irq event stamp: 206837
> >> [ 46.356873] hardirqs last enabled at (206847): [<ffffffff810f1f82>]
> >> __up_console_sem+0x52/0x60
> >> [ 46.357369] hardirqs last disabled at (206854): [<ffffffff810f1f67>]
> >> __up_console_sem+0x37/0x60
> >> [ 46.357864] softirqs last enabled at (206604): [<ffffffff8107bcc5>]
> >> __irq_exit_rcu+0xc5/0x120
> >> [ 46.358358] softirqs last disabled at (206591): [<ffffffff8107bcc5>]
> >> __irq_exit_rcu+0xc5/0x120
> >> [ 46.358865] ---[ end trace 0000000000000000 ]---
> >> # 9/ 3 sockmap::txmsg test hanging corks:OK
> >
> > [...]
> >
> > Actually, we had a fix for these warnings in 9c34e38c4a87 ("bpf,
> > sockmap: Fix memleak in tcp_bpf_sendmsg while sk msg is full") [1].
> >
> > But they reappear for me in v5.18-rc1. Trying to bisect what went
> > wrong...
> >
> > [1]
> >
> https://lore.kernel.org/bpf/20220304081145.2037182-3-wangyufen@huawe
> i.
> > com
>
> Traced it down to 84472b436e76 ("bpf, sockmap: Fix more uncharged while
> msg has more_data"), which was the next commit after the one I thought
> contained the fix, that is 9c34e38c4a87.
>
> I started a discussion in the thread for the patch set [1], as the warnings look
> completely unrelated to this change.
>
Okay, thank you for the clarification.
> [1] https://lore.kernel.org/netdev/87v8o5gdw2.fsf@cloudflare.com/
^ permalink raw reply [flat|nested] 9+ messages in thread
end of thread, other threads:[~2022-10-28 2:56 UTC | newest]
Thread overview: 9+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-08-23 13:37 [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory Liu Jian
2022-08-23 13:37 ` [PATCH bpf-next v2 1/2] net: " Liu Jian
2022-08-30 2:33 ` John Fastabend
2022-08-23 13:37 ` [PATCH bpf-next v2 2/2] selftests/bpf: Add wait send memory test for sockmap redirect Liu Jian
2022-09-26 16:00 ` [PATCH bpf-next v2 0/2] If the sock is dead, do not access sock's sk_wq in sk_stream_wait_memory patchwork-bot+netdevbpf
2022-10-27 10:36 ` Jakub Sitnicki
2022-10-27 15:30 ` Jakub Sitnicki
2022-10-27 21:24 ` Jakub Sitnicki
2022-10-28 2:56 ` 答复: " liujian (CE)
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).