* [PATCH next RFC] net: sched: mirred: move xmit to tasklet
@ 2017-04-18 13:03 Florian Westphal
2017-04-18 16:18 ` Eric Dumazet
2017-04-18 17:15 ` David Miller
0 siblings, 2 replies; 3+ messages in thread
From: Florian Westphal @ 2017-04-18 13:03 UTC (permalink / raw)
To: netdev; +Cc: Florian Westphal
mirred is prone to deadlocks as it invokes dev_queue_xmit while
holding one or more qdisc locks.
Avoid lock recursions by moving tx context to a tasklet.
Signed-off-by: Florian Westphal <fw@strlen.de>
---
This a stab at removing the lock recursions discussed during netconf.
Taking the cost of the tasklet appears to be the only solution;
i tried to use a percpu 'history' instead but its not clear to
me that this avoids all corner cases.
While this patch doesn't avoid loops we don't hang the kernel
anymore and removing the 'looping' filter makes things calm
down again (there are also other ways to create such loops anyway,
including use of a cable... )
diff --git a/net/sched/act_mirred.c b/net/sched/act_mirred.c
index 1b5549ababd4..3dd61eba4741 100644
--- a/net/sched/act_mirred.c
+++ b/net/sched/act_mirred.c
@@ -17,6 +17,7 @@
#include <linux/string.h>
#include <linux/errno.h>
#include <linux/skbuff.h>
+#include <linux/skb_array.h>
#include <linux/rtnetlink.h>
#include <linux/module.h>
#include <linux/init.h>
@@ -25,10 +26,19 @@
#include <net/net_namespace.h>
#include <net/netlink.h>
#include <net/pkt_sched.h>
+#include <net/dst.h>
#include <linux/tc_act/tc_mirred.h>
#include <net/tc_act/tc_mirred.h>
+#define MIRRED_TXLEN 512
#define MIRRED_TAB_MASK 7
+
+struct mirred_tx_data {
+ struct tasklet_struct mirred_tasklet;
+ struct skb_array skb_array;
+};
+
+static DEFINE_PER_CPU(struct mirred_tx_data, mirred_tx_data);
static LIST_HEAD(mirred_list);
static DEFINE_SPINLOCK(mirred_list_lock);
@@ -158,6 +168,44 @@ static int tcf_mirred_init(struct net *net, struct nlattr *nla,
return ret;
}
+static void mirred_tasklet(unsigned long data)
+{
+ struct mirred_tx_data *d = (void *)data;
+ struct sk_buff *skb;
+ int i;
+
+ for (i = 0; i < MIRRED_TXLEN; i++) {
+ struct net_device *dev;
+
+ skb = skb_array_consume(&d->skb_array);
+ if (!skb)
+ return;
+
+ dev = skb->dev;
+ dev_queue_xmit(skb);
+ dev_put(dev);
+ }
+
+ tasklet_schedule(&d->mirred_tasklet);
+}
+
+static int mirred_xmit(struct sk_buff *skb)
+{
+ struct mirred_tx_data *d = this_cpu_ptr(&mirred_tx_data);
+
+ skb_dst_force(skb);
+ dev_hold(skb->dev);
+
+ if (skb_array_produce_bh(&d->skb_array, skb)) {
+ dev_put(skb->dev);
+ kfree_skb(skb);
+ return -ENOBUFS;
+ }
+
+ tasklet_schedule(&d->mirred_tasklet);
+ return 0;
+}
+
static int tcf_mirred(struct sk_buff *skb, const struct tc_action *a,
struct tcf_result *res)
{
@@ -217,7 +265,7 @@ static int tcf_mirred(struct sk_buff *skb, const struct tc_action *a,
skb2->skb_iif = skb->dev->ifindex;
skb2->dev = dev;
if (!tcf_mirred_act_wants_ingress(m_eaction))
- err = dev_queue_xmit(skb2);
+ err = mirred_xmit(skb2);
else
err = netif_receive_skb(skb2);
@@ -365,20 +413,66 @@ MODULE_AUTHOR("Jamal Hadi Salim(2002)");
MODULE_DESCRIPTION("Device Mirror/redirect actions");
MODULE_LICENSE("GPL");
+static void mirred_cleanup_pcpu(void)
+{
+ int cpu;
+
+ for_each_possible_cpu(cpu) {
+ struct mirred_tx_data *data;
+
+ data = per_cpu_ptr(&mirred_tx_data, cpu);
+
+ skb_array_cleanup(&data->skb_array);
+ tasklet_kill(&data->mirred_tasklet);
+ }
+}
+
static int __init mirred_init_module(void)
{
- int err = register_netdevice_notifier(&mirred_device_notifier);
- if (err)
+ int cpu, err;
+
+ for_each_possible_cpu(cpu) {
+ struct mirred_tx_data *data;
+
+ data = per_cpu_ptr(&mirred_tx_data, cpu);
+
+ tasklet_init(&data->mirred_tasklet, mirred_tasklet,
+ (unsigned long)data);
+ err = skb_array_init(&data->skb_array, MIRRED_TXLEN, GFP_KERNEL);
+ if (err) {
+ unregister_netdevice_notifier(&mirred_device_notifier);
+
+ while (cpu) {
+ data = per_cpu_ptr(&mirred_tx_data, --cpu);
+ skb_array_cleanup(&data->skb_array);
+ }
+
+ return err;
+ }
+ }
+
+ err = register_netdevice_notifier(&mirred_device_notifier);
+ if (err) {
+ mirred_cleanup_pcpu();
+ return err;
+ }
+
+ err = tcf_register_action(&act_mirred_ops, &mirred_net_ops);
+ if (err) {
+ unregister_netdevice_notifier(&mirred_device_notifier);
+ mirred_cleanup_pcpu();
return err;
+ }
pr_info("Mirror/redirect action on\n");
- return tcf_register_action(&act_mirred_ops, &mirred_net_ops);
+ return 0;
}
static void __exit mirred_cleanup_module(void)
{
tcf_unregister_action(&act_mirred_ops, &mirred_net_ops);
unregister_netdevice_notifier(&mirred_device_notifier);
+ mirred_cleanup_pcpu();
}
module_init(mirred_init_module);
--
2.10.2
^ permalink raw reply related [flat|nested] 3+ messages in thread
* Re: [PATCH next RFC] net: sched: mirred: move xmit to tasklet
2017-04-18 13:03 [PATCH next RFC] net: sched: mirred: move xmit to tasklet Florian Westphal
@ 2017-04-18 16:18 ` Eric Dumazet
2017-04-18 17:15 ` David Miller
1 sibling, 0 replies; 3+ messages in thread
From: Eric Dumazet @ 2017-04-18 16:18 UTC (permalink / raw)
To: Florian Westphal; +Cc: netdev
On Tue, 2017-04-18 at 15:03 +0200, Florian Westphal wrote:
> mirred is prone to deadlocks as it invokes dev_queue_xmit while
> holding one or more qdisc locks.
>
> Avoid lock recursions by moving tx context to a tasklet.
>
> Signed-off-by: Florian Westphal <fw@strlen.de>
> ---
> This a stab at removing the lock recursions discussed during netconf.
>
> Taking the cost of the tasklet appears to be the only solution;
> i tried to use a percpu 'history' instead but its not clear to
> me that this avoids all corner cases.
>
> While this patch doesn't avoid loops we don't hang the kernel
> anymore and removing the 'looping' filter makes things calm
> down again (there are also other ways to create such loops anyway,
> including use of a cable... )
>
> diff --git a/net/sched/act_mirred.c b/net/sched/act_mirred.c
> index 1b5549ababd4..3dd61eba4741 100644
> --- a/net/sched/act_mirred.c
> +++ b/net/sched/act_mirred.c
> @@ -17,6 +17,7 @@
> #include <linux/string.h>
> #include <linux/errno.h>
> #include <linux/skbuff.h>
> +#include <linux/skb_array.h>
> #include <linux/rtnetlink.h>
> #include <linux/module.h>
> #include <linux/init.h>
> @@ -25,10 +26,19 @@
> #include <net/net_namespace.h>
> #include <net/netlink.h>
> #include <net/pkt_sched.h>
> +#include <net/dst.h>
> #include <linux/tc_act/tc_mirred.h>
> #include <net/tc_act/tc_mirred.h>
>
> +#define MIRRED_TXLEN 512
Using an skb array looks overkill to me, especially if using per cpu
queue. A standard skb list should be good enough ?
> +static void mirred_cleanup_pcpu(void)
> +{
> + int cpu;
> +
> + for_each_possible_cpu(cpu) {
> + struct mirred_tx_data *data;
> +
> + data = per_cpu_ptr(&mirred_tx_data, cpu);
> +
> + skb_array_cleanup(&data->skb_array);
This wont do the dev_put() on skb->dev
> + tasklet_kill(&data->mirred_tasklet);
You might need to kill the tasklet _before_ doing the cleanup ?
> + }
> +}
> +
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [PATCH next RFC] net: sched: mirred: move xmit to tasklet
2017-04-18 13:03 [PATCH next RFC] net: sched: mirred: move xmit to tasklet Florian Westphal
2017-04-18 16:18 ` Eric Dumazet
@ 2017-04-18 17:15 ` David Miller
1 sibling, 0 replies; 3+ messages in thread
From: David Miller @ 2017-04-18 17:15 UTC (permalink / raw)
To: fw; +Cc: netdev
From: Florian Westphal <fw@strlen.de>
Date: Tue, 18 Apr 2017 15:03:31 +0200
> mirred is prone to deadlocks as it invokes dev_queue_xmit while
> holding one or more qdisc locks.
>
> Avoid lock recursions by moving tx context to a tasklet.
>
> Signed-off-by: Florian Westphal <fw@strlen.de>
> ---
> This a stab at removing the lock recursions discussed during netconf.
>
> Taking the cost of the tasklet appears to be the only solution;
> i tried to use a percpu 'history' instead but its not clear to
> me that this avoids all corner cases.
>
> While this patch doesn't avoid loops we don't hang the kernel
> anymore and removing the 'looping' filter makes things calm
> down again (there are also other ways to create such loops anyway,
> including use of a cable... )
The thing is, those qdisc's are per-device.
So we could return the locking cpu and back out of deadlocks just like
we do for the netdev tx lock.
Given that, maybe the remaining part of the equation could possibly be
handled by a ttl.
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2017-04-18 17:15 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2017-04-18 13:03 [PATCH next RFC] net: sched: mirred: move xmit to tasklet Florian Westphal
2017-04-18 16:18 ` Eric Dumazet
2017-04-18 17:15 ` David Miller
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.