* [PATCH net-next 0/2] move ndo_xdp_xmit stats to peer veth_rq @ 2020-03-25 19:22 Lorenzo Bianconi 2020-03-25 19:22 ` [PATCH net-next 1/2] veth: rely on veth_rq in veth_xdp_flush_bq signature Lorenzo Bianconi 2020-03-25 19:22 ` [PATCH net-next 2/2] veth: rely on peer veth_rq for ndo_xdp_xmit accounting Lorenzo Bianconi 0 siblings, 2 replies; 5+ messages in thread From: Lorenzo Bianconi @ 2020-03-25 19:22 UTC (permalink / raw) To: netdev; +Cc: davem, toshiaki.makita1, brouer, dsahern, lorenzo.bianconi, toke Move ndo_xdp_xmit ethtool stats accounting to peer veth_rq. Move XDP_TX accounting to veth_xdp_flush_bq routine. Lorenzo Bianconi (2): veth: rely on veth_rq in veth_xdp_flush_bq signature veth: rely on peer veth_rq for ndo_xdp_xmit accounting drivers/net/veth.c | 143 ++++++++++++++++++++++++++++----------------- 1 file changed, 89 insertions(+), 54 deletions(-) -- 2.25.1 ^ permalink raw reply [flat|nested] 5+ messages in thread
* [PATCH net-next 1/2] veth: rely on veth_rq in veth_xdp_flush_bq signature 2020-03-25 19:22 [PATCH net-next 0/2] move ndo_xdp_xmit stats to peer veth_rq Lorenzo Bianconi @ 2020-03-25 19:22 ` Lorenzo Bianconi 2020-03-26 6:36 ` Toshiaki Makita 2020-03-25 19:22 ` [PATCH net-next 2/2] veth: rely on peer veth_rq for ndo_xdp_xmit accounting Lorenzo Bianconi 1 sibling, 1 reply; 5+ messages in thread From: Lorenzo Bianconi @ 2020-03-25 19:22 UTC (permalink / raw) To: netdev; +Cc: davem, toshiaki.makita1, brouer, dsahern, lorenzo.bianconi, toke Substitute net_device point with veth_rq one in veth_xdp_flush_bq, veth_xdp_flush and veth_xdp_tx signature. This is a preliminary patch to account xdp_xmit counter on 'receiving' veth_rq Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org> --- drivers/net/veth.c | 30 +++++++++++++++--------------- 1 file changed, 15 insertions(+), 15 deletions(-) diff --git a/drivers/net/veth.c b/drivers/net/veth.c index b6505a6c7102..2041152da716 100644 --- a/drivers/net/veth.c +++ b/drivers/net/veth.c @@ -468,46 +468,46 @@ static int veth_ndo_xdp_xmit(struct net_device *dev, int n, return veth_xdp_xmit(dev, n, frames, flags, true); } -static void veth_xdp_flush_bq(struct net_device *dev, struct veth_xdp_tx_bq *bq) +static void veth_xdp_flush_bq(struct veth_rq *rq, struct veth_xdp_tx_bq *bq) { int sent, i, err = 0; - sent = veth_xdp_xmit(dev, bq->count, bq->q, 0, false); + sent = veth_xdp_xmit(rq->dev, bq->count, bq->q, 0, false); if (sent < 0) { err = sent; sent = 0; for (i = 0; i < bq->count; i++) xdp_return_frame(bq->q[i]); } - trace_xdp_bulk_tx(dev, sent, bq->count - sent, err); + trace_xdp_bulk_tx(rq->dev, sent, bq->count - sent, err); bq->count = 0; } -static void veth_xdp_flush(struct net_device *dev, struct veth_xdp_tx_bq *bq) +static void veth_xdp_flush(struct veth_rq *rq, struct veth_xdp_tx_bq *bq) { - struct veth_priv *rcv_priv, *priv = netdev_priv(dev); + struct veth_priv *rcv_priv, *priv = netdev_priv(rq->dev); struct net_device *rcv; - struct veth_rq *rq; + struct veth_rq *rcv_rq; rcu_read_lock(); - veth_xdp_flush_bq(dev, bq); + veth_xdp_flush_bq(rq, bq); rcv = rcu_dereference(priv->peer); if (unlikely(!rcv)) goto out; rcv_priv = netdev_priv(rcv); - rq = &rcv_priv->rq[veth_select_rxq(rcv)]; + rcv_rq = &rcv_priv->rq[veth_select_rxq(rcv)]; /* xdp_ring is initialized on receive side? */ - if (unlikely(!rcu_access_pointer(rq->xdp_prog))) + if (unlikely(!rcu_access_pointer(rcv_rq->xdp_prog))) goto out; - __veth_xdp_flush(rq); + __veth_xdp_flush(rcv_rq); out: rcu_read_unlock(); } -static int veth_xdp_tx(struct net_device *dev, struct xdp_buff *xdp, +static int veth_xdp_tx(struct veth_rq *rq, struct xdp_buff *xdp, struct veth_xdp_tx_bq *bq) { struct xdp_frame *frame = convert_to_xdp_frame(xdp); @@ -516,7 +516,7 @@ static int veth_xdp_tx(struct net_device *dev, struct xdp_buff *xdp, return -EOVERFLOW; if (unlikely(bq->count == VETH_XDP_TX_BULK_SIZE)) - veth_xdp_flush_bq(dev, bq); + veth_xdp_flush_bq(rq, bq); bq->q[bq->count++] = frame; @@ -559,7 +559,7 @@ static struct sk_buff *veth_xdp_rcv_one(struct veth_rq *rq, orig_frame = *frame; xdp.data_hard_start = head; xdp.rxq->mem = frame->mem; - if (unlikely(veth_xdp_tx(rq->dev, &xdp, bq) < 0)) { + if (unlikely(veth_xdp_tx(rq, &xdp, bq) < 0)) { trace_xdp_exception(rq->dev, xdp_prog, act); frame = &orig_frame; stats->rx_drops++; @@ -692,7 +692,7 @@ static struct sk_buff *veth_xdp_rcv_skb(struct veth_rq *rq, get_page(virt_to_page(xdp.data)); consume_skb(skb); xdp.rxq->mem = rq->xdp_mem; - if (unlikely(veth_xdp_tx(rq->dev, &xdp, bq) < 0)) { + if (unlikely(veth_xdp_tx(rq, &xdp, bq) < 0)) { trace_xdp_exception(rq->dev, xdp_prog, act); stats->rx_drops++; goto err_xdp; @@ -817,7 +817,7 @@ static int veth_poll(struct napi_struct *napi, int budget) } if (stats.xdp_tx > 0) - veth_xdp_flush(rq->dev, &bq); + veth_xdp_flush(rq, &bq); if (stats.xdp_redirect > 0) xdp_do_flush(); xdp_clear_return_frame_no_direct(); -- 2.25.1 ^ permalink raw reply related [flat|nested] 5+ messages in thread
* Re: [PATCH net-next 1/2] veth: rely on veth_rq in veth_xdp_flush_bq signature 2020-03-25 19:22 ` [PATCH net-next 1/2] veth: rely on veth_rq in veth_xdp_flush_bq signature Lorenzo Bianconi @ 2020-03-26 6:36 ` Toshiaki Makita 0 siblings, 0 replies; 5+ messages in thread From: Toshiaki Makita @ 2020-03-26 6:36 UTC (permalink / raw) To: Lorenzo Bianconi, netdev; +Cc: davem, brouer, dsahern, lorenzo.bianconi, toke On 2020/03/26 4:22, Lorenzo Bianconi wrote: > Substitute net_device point with veth_rq one in veth_xdp_flush_bq, > veth_xdp_flush and veth_xdp_tx signature. This is a preliminary patch > to account xdp_xmit counter on 'receiving' veth_rq > > Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org> Acked-by: Toshiaki Makita <toshiaki.makita1@gmail.com> > --- > drivers/net/veth.c | 30 +++++++++++++++--------------- > 1 file changed, 15 insertions(+), 15 deletions(-) > > diff --git a/drivers/net/veth.c b/drivers/net/veth.c > index b6505a6c7102..2041152da716 100644 > --- a/drivers/net/veth.c > +++ b/drivers/net/veth.c > @@ -468,46 +468,46 @@ static int veth_ndo_xdp_xmit(struct net_device *dev, int n, > return veth_xdp_xmit(dev, n, frames, flags, true); > } > > -static void veth_xdp_flush_bq(struct net_device *dev, struct veth_xdp_tx_bq *bq) > +static void veth_xdp_flush_bq(struct veth_rq *rq, struct veth_xdp_tx_bq *bq) > { > int sent, i, err = 0; > > - sent = veth_xdp_xmit(dev, bq->count, bq->q, 0, false); > + sent = veth_xdp_xmit(rq->dev, bq->count, bq->q, 0, false); > if (sent < 0) { > err = sent; > sent = 0; > for (i = 0; i < bq->count; i++) > xdp_return_frame(bq->q[i]); > } > - trace_xdp_bulk_tx(dev, sent, bq->count - sent, err); > + trace_xdp_bulk_tx(rq->dev, sent, bq->count - sent, err); > > bq->count = 0; > } > > -static void veth_xdp_flush(struct net_device *dev, struct veth_xdp_tx_bq *bq) > +static void veth_xdp_flush(struct veth_rq *rq, struct veth_xdp_tx_bq *bq) > { > - struct veth_priv *rcv_priv, *priv = netdev_priv(dev); > + struct veth_priv *rcv_priv, *priv = netdev_priv(rq->dev); > struct net_device *rcv; > - struct veth_rq *rq; > + struct veth_rq *rcv_rq; > > rcu_read_lock(); > - veth_xdp_flush_bq(dev, bq); > + veth_xdp_flush_bq(rq, bq); > rcv = rcu_dereference(priv->peer); > if (unlikely(!rcv)) > goto out; > > rcv_priv = netdev_priv(rcv); > - rq = &rcv_priv->rq[veth_select_rxq(rcv)]; > + rcv_rq = &rcv_priv->rq[veth_select_rxq(rcv)]; > /* xdp_ring is initialized on receive side? */ > - if (unlikely(!rcu_access_pointer(rq->xdp_prog))) > + if (unlikely(!rcu_access_pointer(rcv_rq->xdp_prog))) > goto out; > > - __veth_xdp_flush(rq); > + __veth_xdp_flush(rcv_rq); > out: > rcu_read_unlock(); > } > > -static int veth_xdp_tx(struct net_device *dev, struct xdp_buff *xdp, > +static int veth_xdp_tx(struct veth_rq *rq, struct xdp_buff *xdp, > struct veth_xdp_tx_bq *bq) > { > struct xdp_frame *frame = convert_to_xdp_frame(xdp); > @@ -516,7 +516,7 @@ static int veth_xdp_tx(struct net_device *dev, struct xdp_buff *xdp, > return -EOVERFLOW; > > if (unlikely(bq->count == VETH_XDP_TX_BULK_SIZE)) > - veth_xdp_flush_bq(dev, bq); > + veth_xdp_flush_bq(rq, bq); > > bq->q[bq->count++] = frame; > > @@ -559,7 +559,7 @@ static struct sk_buff *veth_xdp_rcv_one(struct veth_rq *rq, > orig_frame = *frame; > xdp.data_hard_start = head; > xdp.rxq->mem = frame->mem; > - if (unlikely(veth_xdp_tx(rq->dev, &xdp, bq) < 0)) { > + if (unlikely(veth_xdp_tx(rq, &xdp, bq) < 0)) { > trace_xdp_exception(rq->dev, xdp_prog, act); > frame = &orig_frame; > stats->rx_drops++; > @@ -692,7 +692,7 @@ static struct sk_buff *veth_xdp_rcv_skb(struct veth_rq *rq, > get_page(virt_to_page(xdp.data)); > consume_skb(skb); > xdp.rxq->mem = rq->xdp_mem; > - if (unlikely(veth_xdp_tx(rq->dev, &xdp, bq) < 0)) { > + if (unlikely(veth_xdp_tx(rq, &xdp, bq) < 0)) { > trace_xdp_exception(rq->dev, xdp_prog, act); > stats->rx_drops++; > goto err_xdp; > @@ -817,7 +817,7 @@ static int veth_poll(struct napi_struct *napi, int budget) > } > > if (stats.xdp_tx > 0) > - veth_xdp_flush(rq->dev, &bq); > + veth_xdp_flush(rq, &bq); > if (stats.xdp_redirect > 0) > xdp_do_flush(); > xdp_clear_return_frame_no_direct(); > ^ permalink raw reply [flat|nested] 5+ messages in thread
* [PATCH net-next 2/2] veth: rely on peer veth_rq for ndo_xdp_xmit accounting 2020-03-25 19:22 [PATCH net-next 0/2] move ndo_xdp_xmit stats to peer veth_rq Lorenzo Bianconi 2020-03-25 19:22 ` [PATCH net-next 1/2] veth: rely on veth_rq in veth_xdp_flush_bq signature Lorenzo Bianconi @ 2020-03-25 19:22 ` Lorenzo Bianconi 2020-03-26 6:57 ` Toshiaki Makita 1 sibling, 1 reply; 5+ messages in thread From: Lorenzo Bianconi @ 2020-03-25 19:22 UTC (permalink / raw) To: netdev; +Cc: davem, toshiaki.makita1, brouer, dsahern, lorenzo.bianconi, toke Rely on 'remote' veth_rq to account ndo_xdp_xmit ethtool counters. Move XDP_TX accounting to veth_xdp_flush_bq routine. Remove 'rx' prefix in rx xdp ethool counters Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org> --- drivers/net/veth.c | 113 +++++++++++++++++++++++++++++---------------- 1 file changed, 74 insertions(+), 39 deletions(-) diff --git a/drivers/net/veth.c b/drivers/net/veth.c index 2041152da716..1b95ed0c6d67 100644 --- a/drivers/net/veth.c +++ b/drivers/net/veth.c @@ -92,17 +92,22 @@ struct veth_q_stat_desc { static const struct veth_q_stat_desc veth_rq_stats_desc[] = { { "xdp_packets", VETH_RQ_STAT(xdp_packets) }, { "xdp_bytes", VETH_RQ_STAT(xdp_bytes) }, - { "rx_drops", VETH_RQ_STAT(rx_drops) }, - { "rx_xdp_redirect", VETH_RQ_STAT(xdp_redirect) }, - { "rx_xdp_drops", VETH_RQ_STAT(xdp_drops) }, - { "rx_xdp_tx", VETH_RQ_STAT(xdp_tx) }, - { "rx_xdp_tx_errors", VETH_RQ_STAT(xdp_tx_err) }, - { "tx_xdp_xmit", VETH_RQ_STAT(xdp_xmit) }, - { "tx_xdp_xmit_errors", VETH_RQ_STAT(xdp_xmit_err) }, + { "drops", VETH_RQ_STAT(rx_drops) }, + { "xdp_redirect", VETH_RQ_STAT(xdp_redirect) }, + { "xdp_drops", VETH_RQ_STAT(xdp_drops) }, + { "xdp_tx", VETH_RQ_STAT(xdp_tx) }, + { "xdp_tx_errors", VETH_RQ_STAT(xdp_tx_err) }, }; #define VETH_RQ_STATS_LEN ARRAY_SIZE(veth_rq_stats_desc) +static const struct veth_q_stat_desc veth_tq_stats_desc[] = { + { "xdp_xmit", VETH_RQ_STAT(xdp_xmit) }, + { "xdp_xmit_errors", VETH_RQ_STAT(xdp_xmit_err) }, +}; + +#define VETH_TQ_STATS_LEN ARRAY_SIZE(veth_tq_stats_desc) + static struct { const char string[ETH_GSTRING_LEN]; } ethtool_stats_keys[] = { @@ -142,6 +147,14 @@ static void veth_get_strings(struct net_device *dev, u32 stringset, u8 *buf) p += ETH_GSTRING_LEN; } } + for (i = 0; i < dev->real_num_tx_queues; i++) { + for (j = 0; j < VETH_TQ_STATS_LEN; j++) { + snprintf(p, ETH_GSTRING_LEN, + "tx_queue_%u_%.18s", + i, veth_tq_stats_desc[j].desc); + p += ETH_GSTRING_LEN; + } + } break; } } @@ -151,7 +164,8 @@ static int veth_get_sset_count(struct net_device *dev, int sset) switch (sset) { case ETH_SS_STATS: return ARRAY_SIZE(ethtool_stats_keys) + - VETH_RQ_STATS_LEN * dev->real_num_rx_queues; + VETH_RQ_STATS_LEN * dev->real_num_rx_queues + + VETH_TQ_STATS_LEN * dev->real_num_tx_queues; default: return -EOPNOTSUPP; } @@ -160,7 +174,7 @@ static int veth_get_sset_count(struct net_device *dev, int sset) static void veth_get_ethtool_stats(struct net_device *dev, struct ethtool_stats *stats, u64 *data) { - struct veth_priv *priv = netdev_priv(dev); + struct veth_priv *rcv_priv, *priv = netdev_priv(dev); struct net_device *peer = rtnl_dereference(priv->peer); int i, j, idx; @@ -181,6 +195,26 @@ static void veth_get_ethtool_stats(struct net_device *dev, } while (u64_stats_fetch_retry_irq(&rq_stats->syncp, start)); idx += VETH_RQ_STATS_LEN; } + + if (!peer) + return; + + rcv_priv = netdev_priv(peer); + for (i = 0; i < peer->real_num_rx_queues; i++) { + const struct veth_rq_stats *rq_stats = &rcv_priv->rq[i].stats; + const void *base = (void *)&rq_stats->vs; + unsigned int start, tx_idx = idx; + size_t offset; + + tx_idx += (i % dev->real_num_tx_queues) * VETH_TQ_STATS_LEN; + do { + start = u64_stats_fetch_begin_irq(&rq_stats->syncp); + for (j = 0; j < VETH_TQ_STATS_LEN; j++) { + offset = veth_tq_stats_desc[j].offset; + data[tx_idx + j] += *(u64 *)(base + offset); + } + } while (u64_stats_fetch_retry_irq(&rq_stats->syncp, start)); + } } static const struct ethtool_ops veth_ethtool_ops = { @@ -340,8 +374,8 @@ static void veth_get_stats64(struct net_device *dev, tot->tx_packets = packets; veth_stats_rx(&rx, dev); - tot->tx_dropped += rx.xdp_xmit_err + rx.xdp_tx_err; - tot->rx_dropped = rx.rx_drops; + tot->tx_dropped += rx.xdp_tx_err; + tot->rx_dropped = rx.rx_drops + rx.xdp_xmit_err; tot->rx_bytes = rx.xdp_bytes; tot->rx_packets = rx.xdp_packets; @@ -353,7 +387,8 @@ static void veth_get_stats64(struct net_device *dev, tot->rx_packets += packets; veth_stats_rx(&rx, peer); - tot->rx_dropped += rx.xdp_xmit_err + rx.xdp_tx_err; + tot->tx_dropped += rx.xdp_xmit_err; + tot->rx_dropped += rx.xdp_tx_err; tot->tx_bytes += rx.xdp_bytes; tot->tx_packets += rx.xdp_packets; } @@ -394,38 +429,28 @@ static int veth_xdp_xmit(struct net_device *dev, int n, u32 flags, bool ndo_xmit) { struct veth_priv *rcv_priv, *priv = netdev_priv(dev); - unsigned int qidx, max_len; + int i, ret = -ENXIO, drops = 0; struct net_device *rcv; - int i, ret, drops = n; + unsigned int max_len; struct veth_rq *rq; - rcu_read_lock(); - if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK)) { - rcu_read_unlock(); - atomic64_add(drops, &priv->dropped); + if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK)) return -EINVAL; - } + rcu_read_lock(); rcv = rcu_dereference(priv->peer); - if (unlikely(!rcv)) { - rcu_read_unlock(); - atomic64_add(drops, &priv->dropped); - return -ENXIO; - } + if (unlikely(!rcv)) + goto out; rcv_priv = netdev_priv(rcv); - qidx = veth_select_rxq(rcv); - rq = &rcv_priv->rq[qidx]; + rq = &rcv_priv->rq[veth_select_rxq(rcv)]; /* Non-NULL xdp_prog ensures that xdp_ring is initialized on receive * side. This means an XDP program is loaded on the peer and the peer * device is up. */ - if (!rcu_access_pointer(rq->xdp_prog)) { - ret = -ENXIO; - goto drop; - } + if (!rcu_access_pointer(rq->xdp_prog)) + goto out; - drops = 0; max_len = rcv->mtu + rcv->hard_header_len + VLAN_HLEN; spin_lock(&rq->xdp_ring.producer_lock); @@ -445,18 +470,14 @@ static int veth_xdp_xmit(struct net_device *dev, int n, __veth_xdp_flush(rq); ret = n - drops; -drop: - rq = &priv->rq[qidx]; - u64_stats_update_begin(&rq->stats.syncp); if (ndo_xmit) { + u64_stats_update_begin(&rq->stats.syncp); rq->stats.vs.xdp_xmit += n - drops; rq->stats.vs.xdp_xmit_err += drops; - } else { - rq->stats.vs.xdp_tx += n - drops; - rq->stats.vs.xdp_tx_err += drops; + u64_stats_update_end(&rq->stats.syncp); } - u64_stats_update_end(&rq->stats.syncp); +out: rcu_read_unlock(); return ret; @@ -465,7 +486,16 @@ static int veth_xdp_xmit(struct net_device *dev, int n, static int veth_ndo_xdp_xmit(struct net_device *dev, int n, struct xdp_frame **frames, u32 flags) { - return veth_xdp_xmit(dev, n, frames, flags, true); + int err; + + err = veth_xdp_xmit(dev, n, frames, flags, true); + if (err < 0) { + struct veth_priv *priv = netdev_priv(dev); + + atomic64_add(n, &priv->dropped); + } + + return err; } static void veth_xdp_flush_bq(struct veth_rq *rq, struct veth_xdp_tx_bq *bq) @@ -481,6 +511,11 @@ static void veth_xdp_flush_bq(struct veth_rq *rq, struct veth_xdp_tx_bq *bq) } trace_xdp_bulk_tx(rq->dev, sent, bq->count - sent, err); + u64_stats_update_begin(&rq->stats.syncp); + rq->stats.vs.xdp_tx += sent; + rq->stats.vs.xdp_tx_err += bq->count - sent; + u64_stats_update_end(&rq->stats.syncp); + bq->count = 0; } -- 2.25.1 ^ permalink raw reply related [flat|nested] 5+ messages in thread
* Re: [PATCH net-next 2/2] veth: rely on peer veth_rq for ndo_xdp_xmit accounting 2020-03-25 19:22 ` [PATCH net-next 2/2] veth: rely on peer veth_rq for ndo_xdp_xmit accounting Lorenzo Bianconi @ 2020-03-26 6:57 ` Toshiaki Makita 0 siblings, 0 replies; 5+ messages in thread From: Toshiaki Makita @ 2020-03-26 6:57 UTC (permalink / raw) To: Lorenzo Bianconi, netdev; +Cc: davem, brouer, dsahern, lorenzo.bianconi, toke On 2020/03/26 4:22, Lorenzo Bianconi wrote: > Rely on 'remote' veth_rq to account ndo_xdp_xmit ethtool counters. > Move XDP_TX accounting to veth_xdp_flush_bq routine. > Remove 'rx' prefix in rx xdp ethool counters > > Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org> > --- > drivers/net/veth.c | 113 +++++++++++++++++++++++++++++---------------- > 1 file changed, 74 insertions(+), 39 deletions(-) > > diff --git a/drivers/net/veth.c b/drivers/net/veth.c > index 2041152da716..1b95ed0c6d67 100644 > --- a/drivers/net/veth.c > +++ b/drivers/net/veth.c > @@ -92,17 +92,22 @@ struct veth_q_stat_desc { > static const struct veth_q_stat_desc veth_rq_stats_desc[] = { > { "xdp_packets", VETH_RQ_STAT(xdp_packets) }, > { "xdp_bytes", VETH_RQ_STAT(xdp_bytes) }, > - { "rx_drops", VETH_RQ_STAT(rx_drops) }, > - { "rx_xdp_redirect", VETH_RQ_STAT(xdp_redirect) }, > - { "rx_xdp_drops", VETH_RQ_STAT(xdp_drops) }, > - { "rx_xdp_tx", VETH_RQ_STAT(xdp_tx) }, > - { "rx_xdp_tx_errors", VETH_RQ_STAT(xdp_tx_err) }, > - { "tx_xdp_xmit", VETH_RQ_STAT(xdp_xmit) }, > - { "tx_xdp_xmit_errors", VETH_RQ_STAT(xdp_xmit_err) }, > + { "drops", VETH_RQ_STAT(rx_drops) }, > + { "xdp_redirect", VETH_RQ_STAT(xdp_redirect) }, > + { "xdp_drops", VETH_RQ_STAT(xdp_drops) }, > + { "xdp_tx", VETH_RQ_STAT(xdp_tx) }, > + { "xdp_tx_errors", VETH_RQ_STAT(xdp_tx_err) }, > }; > > #define VETH_RQ_STATS_LEN ARRAY_SIZE(veth_rq_stats_desc) > > +static const struct veth_q_stat_desc veth_tq_stats_desc[] = { > + { "xdp_xmit", VETH_RQ_STAT(xdp_xmit) }, > + { "xdp_xmit_errors", VETH_RQ_STAT(xdp_xmit_err) }, > +}; > + > +#define VETH_TQ_STATS_LEN ARRAY_SIZE(veth_tq_stats_desc) > + > static struct { > const char string[ETH_GSTRING_LEN]; > } ethtool_stats_keys[] = { > @@ -142,6 +147,14 @@ static void veth_get_strings(struct net_device *dev, u32 stringset, u8 *buf) > p += ETH_GSTRING_LEN; > } > } > + for (i = 0; i < dev->real_num_tx_queues; i++) { > + for (j = 0; j < VETH_TQ_STATS_LEN; j++) { > + snprintf(p, ETH_GSTRING_LEN, > + "tx_queue_%u_%.18s", > + i, veth_tq_stats_desc[j].desc); > + p += ETH_GSTRING_LEN; > + } > + } > break; > } > } > @@ -151,7 +164,8 @@ static int veth_get_sset_count(struct net_device *dev, int sset) > switch (sset) { > case ETH_SS_STATS: > return ARRAY_SIZE(ethtool_stats_keys) + > - VETH_RQ_STATS_LEN * dev->real_num_rx_queues; > + VETH_RQ_STATS_LEN * dev->real_num_rx_queues + > + VETH_TQ_STATS_LEN * dev->real_num_tx_queues; > default: > return -EOPNOTSUPP; > } > @@ -160,7 +174,7 @@ static int veth_get_sset_count(struct net_device *dev, int sset) > static void veth_get_ethtool_stats(struct net_device *dev, > struct ethtool_stats *stats, u64 *data) > { > - struct veth_priv *priv = netdev_priv(dev); > + struct veth_priv *rcv_priv, *priv = netdev_priv(dev); > struct net_device *peer = rtnl_dereference(priv->peer); > int i, j, idx; > > @@ -181,6 +195,26 @@ static void veth_get_ethtool_stats(struct net_device *dev, > } while (u64_stats_fetch_retry_irq(&rq_stats->syncp, start)); > idx += VETH_RQ_STATS_LEN; > } > + > + if (!peer) > + return; > + > + rcv_priv = netdev_priv(peer); > + for (i = 0; i < peer->real_num_rx_queues; i++) { > + const struct veth_rq_stats *rq_stats = &rcv_priv->rq[i].stats; > + const void *base = (void *)&rq_stats->vs; > + unsigned int start, tx_idx = idx; > + size_t offset; > + > + tx_idx += (i % dev->real_num_tx_queues) * VETH_TQ_STATS_LEN; > + do { > + start = u64_stats_fetch_begin_irq(&rq_stats->syncp); > + for (j = 0; j < VETH_TQ_STATS_LEN; j++) { > + offset = veth_tq_stats_desc[j].offset; > + data[tx_idx + j] += *(u64 *)(base + offset); > + } > + } while (u64_stats_fetch_retry_irq(&rq_stats->syncp, start)); > + } > } > > static const struct ethtool_ops veth_ethtool_ops = { > @@ -340,8 +374,8 @@ static void veth_get_stats64(struct net_device *dev, > tot->tx_packets = packets; > > veth_stats_rx(&rx, dev); > - tot->tx_dropped += rx.xdp_xmit_err + rx.xdp_tx_err; > - tot->rx_dropped = rx.rx_drops; > + tot->tx_dropped += rx.xdp_tx_err; > + tot->rx_dropped = rx.rx_drops + rx.xdp_xmit_err; This looks confusing. How about changing the variable name a bit? struct veth_stats { u64 rx_drops; /* xdp */ u64 xdp_packets; u64 xdp_bytes; u64 xdp_redirect; u64 xdp_drops; u64 xdp_tx; u64 xdp_tx_err; - u64 xdp_xmit; - u64 xdp_xmit_err; + u64 peer_tq_xdp_xmit; + u64 peer_tq_xdp_xmit_err; }; +static const struct veth_q_stat_desc veth_tq_stats_desc[] = { + { "xdp_xmit", VETH_RQ_STAT(peer_tq_xdp_xmit) }, + { "xdp_xmit_errors", VETH_RQ_STAT(peer_tq_xdp_xmit_err) }, +}; And then, tot->rx_dropped = rx.rx_drops + rx.peer_tq_xdp_xmit_err; Otherwise looks good. Thanks! Toshiaki Makita > tot->rx_bytes = rx.xdp_bytes; > tot->rx_packets = rx.xdp_packets; > > @@ -353,7 +387,8 @@ static void veth_get_stats64(struct net_device *dev, > tot->rx_packets += packets; > > veth_stats_rx(&rx, peer); > - tot->rx_dropped += rx.xdp_xmit_err + rx.xdp_tx_err; > + tot->tx_dropped += rx.xdp_xmit_err; > + tot->rx_dropped += rx.xdp_tx_err; > tot->tx_bytes += rx.xdp_bytes; > tot->tx_packets += rx.xdp_packets; > } > @@ -394,38 +429,28 @@ static int veth_xdp_xmit(struct net_device *dev, int n, > u32 flags, bool ndo_xmit) > { > struct veth_priv *rcv_priv, *priv = netdev_priv(dev); > - unsigned int qidx, max_len; > + int i, ret = -ENXIO, drops = 0; > struct net_device *rcv; > - int i, ret, drops = n; > + unsigned int max_len; > struct veth_rq *rq; > > - rcu_read_lock(); > - if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK)) { > - rcu_read_unlock(); > - atomic64_add(drops, &priv->dropped); > + if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK)) > return -EINVAL; > - } > > + rcu_read_lock(); > rcv = rcu_dereference(priv->peer); > - if (unlikely(!rcv)) { > - rcu_read_unlock(); > - atomic64_add(drops, &priv->dropped); > - return -ENXIO; > - } > + if (unlikely(!rcv)) > + goto out; > > rcv_priv = netdev_priv(rcv); > - qidx = veth_select_rxq(rcv); > - rq = &rcv_priv->rq[qidx]; > + rq = &rcv_priv->rq[veth_select_rxq(rcv)]; > /* Non-NULL xdp_prog ensures that xdp_ring is initialized on receive > * side. This means an XDP program is loaded on the peer and the peer > * device is up. > */ > - if (!rcu_access_pointer(rq->xdp_prog)) { > - ret = -ENXIO; > - goto drop; > - } > + if (!rcu_access_pointer(rq->xdp_prog)) > + goto out; > > - drops = 0; > max_len = rcv->mtu + rcv->hard_header_len + VLAN_HLEN; > > spin_lock(&rq->xdp_ring.producer_lock); > @@ -445,18 +470,14 @@ static int veth_xdp_xmit(struct net_device *dev, int n, > __veth_xdp_flush(rq); > > ret = n - drops; > -drop: > - rq = &priv->rq[qidx]; > - u64_stats_update_begin(&rq->stats.syncp); > if (ndo_xmit) { > + u64_stats_update_begin(&rq->stats.syncp); > rq->stats.vs.xdp_xmit += n - drops; > rq->stats.vs.xdp_xmit_err += drops; > - } else { > - rq->stats.vs.xdp_tx += n - drops; > - rq->stats.vs.xdp_tx_err += drops; > + u64_stats_update_end(&rq->stats.syncp); > } > - u64_stats_update_end(&rq->stats.syncp); > > +out: > rcu_read_unlock(); > > return ret; > @@ -465,7 +486,16 @@ static int veth_xdp_xmit(struct net_device *dev, int n, > static int veth_ndo_xdp_xmit(struct net_device *dev, int n, > struct xdp_frame **frames, u32 flags) > { > - return veth_xdp_xmit(dev, n, frames, flags, true); > + int err; > + > + err = veth_xdp_xmit(dev, n, frames, flags, true); > + if (err < 0) { > + struct veth_priv *priv = netdev_priv(dev); > + > + atomic64_add(n, &priv->dropped); > + } > + > + return err; > } > > static void veth_xdp_flush_bq(struct veth_rq *rq, struct veth_xdp_tx_bq *bq) > @@ -481,6 +511,11 @@ static void veth_xdp_flush_bq(struct veth_rq *rq, struct veth_xdp_tx_bq *bq) > } > trace_xdp_bulk_tx(rq->dev, sent, bq->count - sent, err); > > + u64_stats_update_begin(&rq->stats.syncp); > + rq->stats.vs.xdp_tx += sent; > + rq->stats.vs.xdp_tx_err += bq->count - sent; > + u64_stats_update_end(&rq->stats.syncp); > + > bq->count = 0; > } > > ^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2020-03-26 6:57 UTC | newest] Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2020-03-25 19:22 [PATCH net-next 0/2] move ndo_xdp_xmit stats to peer veth_rq Lorenzo Bianconi 2020-03-25 19:22 ` [PATCH net-next 1/2] veth: rely on veth_rq in veth_xdp_flush_bq signature Lorenzo Bianconi 2020-03-26 6:36 ` Toshiaki Makita 2020-03-25 19:22 ` [PATCH net-next 2/2] veth: rely on peer veth_rq for ndo_xdp_xmit accounting Lorenzo Bianconi 2020-03-26 6:57 ` Toshiaki Makita
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).