bpf.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Jesper Dangaard Brouer <hawk@kernel.org>
To: Sebastian Andrzej Siewior <bigeasy@linutronix.de>
Cc: "Toke Høiland-Jørgensen" <toke@redhat.com>,
	bpf@vger.kernel.org, netdev@vger.kernel.org
Subject: Re: [PATCH RFC net-next 1/2] net: Reference bpf_redirect_info via task_struct on PREEMPT_RT.
Date: Tue, 20 Feb 2024 13:57:24 +0100	[thread overview]
Message-ID: <07620deb-2b96-4bcc-a045-480568a27c58@kernel.org> (raw)
In-Reply-To: <20240220120821.1Tbz6IeI@linutronix.de>



On 20/02/2024 13.08, Sebastian Andrzej Siewior wrote:
> On 2024-02-20 11:42:57 [+0100], Jesper Dangaard Brouer wrote:
>> This seems low...
>> Have you remembered to disable Ethernet flow-control?
> 
> No but one side says:
> | i40e 0000:3d:00.1 eno2np1: NIC Link is Up, 10 Gbps Full Duplex, Flow Control: None
> 
> but I did this
> 
>>   # ethtool -A ixgbe1 rx off tx off
>>   # ethtool -A i40e2 rx off tx off
> 
> and it didn't change much.
> 
>>
>>> | Summary                 8,436,294 rx/s                  0 err/s
>>
>> You want to see the "extended" info via cmdline (or Ctrl+\)
>>
>>   # xdp-bench drop -e eth1
>>
>>
>>>
>>> with "-t 8 -b 64". I started with 2 and then increased until rx/s was
>>> falling again. I have ixgbe on the sending side and i40e on the
>>
>> With ixgbe on the sending side, my testlab shows I need -t 2.
>>
>> With -t 2 :
>> Summary                14,678,170 rx/s                  0 err/s
>>    receive total        14,678,170 pkt/s        14,678,170 drop/s         0
>> error/s
>>      cpu:1              14,678,170 pkt/s        14,678,170 drop/s         0
>> error/s
>>    xdp_exception                 0 hit/s
>>
>> with -t 4:
>>
>> Summary                10,255,385 rx/s                  0 err/s
>>    receive total        10,255,385 pkt/s        10,255,385 drop/s         0
>> error/s
>>      cpu:1              10,255,385 pkt/s        10,255,385 drop/s         0
>> error/s
>>    xdp_exception                 0 hit/s
>>
>>> receiving side. I tried to receive on ixgbe but this ended with -ENOMEM
>>> | # xdp-bench drop eth1
>>> | Failed to attach XDP program: Cannot allocate memory
>>>
>>> This is v6.8-rc5 on both sides. Let me see where this is coming from…
>>>
>>
>> Another pitfall with ixgbe is that it does a full link reset when
>> adding/removing XDP prog on device.  This can be annoying if connected
>> back-to-back, because "remote" pktgen will stop on link reset.
> 
> so I replaced nr_cpu_ids with 64 and bootet maxcpus=64 so that I can run
> xdp-bench on the ixgbe.
> 

Yes, ixgbe HW have limited TX queues, and XDP tries to allocate a
hardware TX queue for every CPU in the system.  So, I guess you have too
many CPUs in your system - lol.

Other drivers have a fallback to a locked XDP TX path, so this is also
something to lookout for in the machine with i40e.

> so. i40 send, ixgbe receive.
> 
> -t 2
> 
> | Summary                 2,348,800 rx/s                  0 err/s
> |   receive total         2,348,800 pkt/s         2,348,800 drop/s                0 error/s
> |     cpu:0               2,348,800 pkt/s         2,348,800 drop/s                0 error/s
> |   xdp_exception                 0 hit/s
>

This is way too low, with i40e sending.

On my system with only -t 1 my i40e driver can send with approx 15Mpps:

  Ethtool(i40e2) stat:  15028585 (  15,028,585) <= tx-0.packets /sec
  Ethtool(i40e2) stat:  15028589 (  15,028,589) <= tx_packets /sec


> -t 4
> | Summary                 4,158,199 rx/s                  0 err/s
> |   receive total         4,158,199 pkt/s         4,158,199 drop/s                0 error/s
> |     cpu:0               4,158,199 pkt/s         4,158,199 drop/s                0 error/s
> |   xdp_exception                 0 hit/s
> 

Do notice that this is all hitting CPU:0
(this is by design from pktgen_sample03_burst_single_flow.sh)


> -t 8
> | Summary                 5,612,861 rx/s                  0 err/s
> |   receive total         5,612,861 pkt/s         5,612,861 drop/s                0 error/s
> |     cpu:0               5,612,861 pkt/s         5,612,861 drop/s                0 error/s
> |   xdp_exception                 0 hit/s
> 
> going higher makes the rate drop. With 8 it floats between 5,5… 5,7…
> 

At -t 8 we seem to have hit limit on RX side, which also seems too low.

I recommend checking what speeds packet generator is sending with.
I use this tool: ethtool_stats.pl
 
https://github.com/netoptimizer/network-testing/blob/master/bin/ethtool_stats.pl

What we are basically asking: Make sure packet generator is not the
limiting factor in your tests.
As we need to make sure DUT (Device Under Test) is being overloaded
100%.  I often also check (via per record) that DUT don't have idle CPU
cycles (yes, this can easily happen... and happens when we hit a limit
in hardware either NIC or PCIe slot)


> Doing "ethtool -G eno2np1 tx 4096 rx 4096" on the i40 makes it worse,
> using the default 512/512 gets the numbers from above, going below 256
> makes it worse.
> 
> receiving on i40, sending on ixgbe:
> 
> -t 2
> |Summary                 3,042,957 rx/s                  0 err/s
> |  receive total         3,042,957 pkt/s         3,042,957 drop/s                0 error/s
> |    cpu:60              3,042,957 pkt/s         3,042,957 drop/s                0 error/s
> |  xdp_exception                 0 hit/s
> 
> -t 4
> |Summary                 5,442,166 rx/s                  0 err/s
> |  receive total         5,442,166 pkt/s         5,442,166 drop/s                0 error/s
> |    cpu:60              5,442,166 pkt/s         5,442,166 drop/s                0 error/s
> |  xdp_exception                 0 hit/s
> 
> 
> -t 6
> | Summary                 7,023,406 rx/s                  0 err/s
> |   receive total         7,023,406 pkt/s         7,023,406 drop/s                0 error/s
> |     cpu:60              7,023,406 pkt/s         7,023,406 drop/s                0 error/s
> |   xdp_exception                 0 hit/s
> 
> 
> -t 8
> | Summary                 7,540,915 rx/s                  0 err/s
> |   receive total         7,540,915 pkt/s         7,540,915 drop/s                0 error/s
> |     cpu:60              7,540,915 pkt/s         7,540,915 drop/s                0 error/s
> |   xdp_exception                 0 hit/s
> 
> -t 10
> |Summary                 7,699,143 rx/s                  0 err/s
> |  receive total         7,699,143 pkt/s         7,699,143 drop/s                0 error/s
> |    cpu:60              7,699,143 pkt/s         7,699,143 drop/s                0 error/s
> |  xdp_exception                 0 hit/s
> 

At this level, if you can verify that CPU:60 is 100% loaded, and packet
generator is sending more than rx number, then it could work as a valid
experiment.

> -t 18
> | Summary                 7,784,946 rx/s                  0 err/s
> |   receive total         7,784,946 pkt/s         7,784,946 drop/s                0 error/s
> |     cpu:60              7,784,946 pkt/s         7,784,946 drop/s                0 error/s
> |   xdp_exception                 0 hit/s
> 
> after t18 it drop down to 2,…
> Now I got worse than before since -t8 says 7,5… and it did 8,4 in the
> morning. Do you have maybe a .config for me in case I did not enable the
> performance switch?
> 

I would look for root-cause with perf record +
  perf report --sort cpu,comm,dso,symbol --no-children


--Jesper

  reply	other threads:[~2024-02-20 12:57 UTC|newest]

Thread overview: 25+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-02-13 14:58 [PATCH RFC net-next 0/2] Use per-task storage for XDP-redirects on PREEMPT_RT Sebastian Andrzej Siewior
2024-02-13 14:58 ` [PATCH RFC net-next 1/2] net: Reference bpf_redirect_info via task_struct " Sebastian Andrzej Siewior
2024-02-13 20:50   ` Jesper Dangaard Brouer
2024-02-14 12:19     ` Sebastian Andrzej Siewior
2024-02-14 13:23       ` Toke Høiland-Jørgensen
2024-02-14 14:28         ` Sebastian Andrzej Siewior
2024-02-14 16:08           ` Toke Høiland-Jørgensen
2024-02-14 16:36             ` Sebastian Andrzej Siewior
2024-02-15 20:23               ` Toke Høiland-Jørgensen
2024-02-16 16:57                 ` Sebastian Andrzej Siewior
2024-02-19 19:01                   ` Toke Høiland-Jørgensen
2024-02-20  9:17                     ` Jesper Dangaard Brouer
2024-02-20 10:17                       ` Sebastian Andrzej Siewior
2024-02-20 10:42                         ` Jesper Dangaard Brouer
2024-02-20 12:08                           ` Sebastian Andrzej Siewior
2024-02-20 12:57                             ` Jesper Dangaard Brouer [this message]
2024-02-20 15:32                               ` Sebastian Andrzej Siewior
2024-02-22  9:22                                 ` Sebastian Andrzej Siewior
2024-02-22 10:10                                   ` Jesper Dangaard Brouer
2024-02-22 10:58                                     ` Sebastian Andrzej Siewior
2024-02-20 12:10                           ` Dave Taht
2024-02-14 16:13   ` Toke Høiland-Jørgensen
2024-02-15  9:04     ` Sebastian Andrzej Siewior
2024-02-15 12:11       ` Toke Høiland-Jørgensen
2024-02-13 14:58 ` [PATCH RFC net-next 2/2] net: Move per-CPU flush-lists to bpf_xdp_storage " Sebastian Andrzej Siewior

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=07620deb-2b96-4bcc-a045-480568a27c58@kernel.org \
    --to=hawk@kernel.org \
    --cc=bigeasy@linutronix.de \
    --cc=bpf@vger.kernel.org \
    --cc=netdev@vger.kernel.org \
    --cc=toke@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).