linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Yunsheng Lin <linyunsheng@huawei.com>
To: Ilias Apalodimas <ilias.apalodimas@linaro.org>
Cc: Jesper Dangaard Brouer <jbrouer@redhat.com>, <brouer@redhat.com>,
	Alexander Duyck <alexander.duyck@gmail.com>,
	<davem@davemloft.net>, <kuba@kernel.org>,
	<netdev@vger.kernel.org>, <linux-kernel@vger.kernel.org>,
	<linuxarm@openeuler.org>, <hawk@kernel.org>,
	<jonathan.lemon@gmail.com>, <alobakin@pm.me>,
	<willemb@google.com>, <cong.wang@bytedance.com>,
	<pabeni@redhat.com>, <haokexin@gmail.com>, <nogikh@google.com>,
	<elver@google.com>, <memxor@gmail.com>, <edumazet@google.com>,
	<dsahern@gmail.com>
Subject: Re: [Linuxarm] Re: [PATCH net-next v2 3/3] skbuff: keep track of pp page when __skb_frag_ref() is called
Date: Thu, 16 Sep 2021 17:33:39 +0800	[thread overview]
Message-ID: <ac16cc82-8d98-6a2c-b0a6-7c186808c72c@huawei.com> (raw)
In-Reply-To: <YUMD2v7ffs1xAjaW@apalos.home>

On 2021/9/16 16:44, Ilias Apalodimas wrote:
>>>> appear if we try to pull in your patches on using page pool and recycling
> 
> [...]
> 
>>>> for Tx where TSO and skb_split are used?
>>
>> As my understanding, the problem might exists without tx recycling, because a
>> skb from wire would be passed down to the tcp stack and retransmited back to
>> the wire theoretically. As I am not able to setup a configuration to verify
>> and test it and the handling seems tricky, so I am targetting net-next branch
>> instead of net branch.
>>
>>>>
>>>> I'll be honest, when I came up with the recycling idea for page pool, I
>>>> never intended to support Tx.  I agree with Alexander here,  If people want
>>>> to use it on Tx and think there's value,  we might need to go back to the
>>>> drawing board and see what I've missed.  It's still early and there's a
>>>> handful of drivers using it,  so it will less painful now.
>>
>> Yes, we also need to prototype it to see if there is something missing in the
>> drawing board and how much improvement we get from that:)
>>
>>>
>>> I agree, page_pool is NOT designed or intended for TX support.
>>> E.g. it doesn't make sense to allocate a page_pool instance per socket, as the backing memory structures for page_pool are too much.
>>> As the number RX-queues are more limited it was deemed okay that we use page_pool per RX-queue, which sacrifice some memory to gain speed.
>>
>> As memtioned before, Tx recycling is based on page_pool instance per socket.
>> it shares the page_pool instance with rx.
>>
>> Anyway, based on feedback from edumazet and dsahern, I am still trying to
>> see if the page pool is meaningful for tx.
>>
>>>
>>>
>>>> The pp_recycle_bit was introduced to make the checking faster, instead of
>>>> getting stuff into cache and check the page signature.  If that ends up
>>>> being counterproductive, we could just replace the entire logic with the
>>>> frag count and the page signature, couldn't we?  In that case we should be
>>>> very cautious and measure potential regression on the standard path.
>>>
>>> +1
>>
>> I am not sure "pp_recycle_bit was introduced to make the checking faster" is a
>> valid. The size of "struct page" is only about 9 words(36/72 bytes), which is
>> mostly to be in the same cache line, and both standard path and recycle path have
>> been touching the "struct page", so it seems the overhead for checking signature
>> seems minimal.
>>
>> I agree that we need to be cautious and measure potential regression on the
>> standard path.
> 
> well pp_recycle is on the same cache line boundary with the head_frag we
> need to decide on recycling. After that we start checking page signatures
> etc,  which means the default release path remains mostly unaffected.  
> 
> I guess what you are saying here, is that 'struct page' is going to be
> accessed eventually by the default network path,  so there won't be any 
> noticeable performance hit?  What about the other usecases we have

Yes.

> for pp_recycle right now?  __skb_frag_unref() in skb_shift() or
> skb_try_coalesce() (the latter can probably be removed tbh).

If we decide to go with accurate indicator of a pp page, we just need
to make sure network stack use __skb_frag_unref() and __skb_frag_ref()
to put and get a page frag, the indicator checking need only done in
__skb_frag_unref() and __skb_frag_ref(), so the skb_shift() and
skb_try_coalesce() should be fine too.

> 
>>
>> Another way is to use the bit 0 of frag->bv_page ptr to indicate if a frag
>> page is from page pool.
> 
> Instead of the 'struct page' signature?  And the pp_recycle bit will
> continue to exist?  

pp_recycle bit might only exist or is only used for the head page for the skb.
The bit 0 of frag->bv_page ptr can be used to indicate a frag page uniquely.
Doing a memcpying of shinfo or "*fragto = *fragfrom" automatically pass the
indicator to the new shinfo before doing a __skb_frag_ref(), and __skb_frag_ref()
will increment the _refcount or pp_frag_count according to the bit 0 of
frag->bv_page.

By the way, I also prototype the above idea, and it seems to work well too.

> .
> Right now the 'naive' explanation on the recycling decision is something like:
> 
> if (pp_recycle) <--- recycling bit is set
>     (check page signature) <--- signature matches page pool
> 		(check fragment refcnt) <--- If frags are enabled and is the last consumer
> 			recycle
> 
> If we can proove the performance is unaffected when we eliminate the first if,
> then obviously we should remove it.  I'll try running that test here and see,
> but keep in mind I am only testing on an 1GB interface.  Any chance we can get 
> measurements on a beefier hardware using hns3 ?

Sure, I will try it.
As the kind of performance overhead is small, any performance testcase in mind?

> 
>>
>>>
>>>> But in general,  I'd be happier if we only had a simple logic in our
>>>> testing for the pages we have to recycle.  Debugging and understanding this
>>>> otherwise will end up being a mess.
>>>
>>>
> 
> [...]
> 
> Regards
> /Ilias
> .
> 

  reply	other threads:[~2021-09-16  9:33 UTC|newest]

Thread overview: 24+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-09-14 12:11 [PATCH net-next v2 0/3] some optimization for page pool Yunsheng Lin
2021-09-14 12:11 ` [PATCH net-next v2 1/3] page_pool: support non-split page with PP_FLAG_PAGE_FRAG Yunsheng Lin
2021-09-14 12:11 ` [PATCH net-next v2 2/3] pool_pool: avoid calling compound_head() for skb frag page Yunsheng Lin
2021-09-14 12:11 ` [PATCH net-next v2 3/3] skbuff: keep track of pp page when __skb_frag_ref() is called Yunsheng Lin
2021-09-15  0:59   ` Alexander Duyck
2021-09-15  9:07     ` Yunsheng Lin
2021-09-15 12:56       ` Ilias Apalodimas
2021-09-15 15:42         ` Jesper Dangaard Brouer
2021-09-16  2:05           ` [Linuxarm] " Yunsheng Lin
2021-09-16  8:44             ` Ilias Apalodimas
2021-09-16  9:33               ` Yunsheng Lin [this message]
2021-09-16 10:38                 ` Ilias Apalodimas
2021-09-16 11:04                   ` Yunsheng Lin
2021-09-16 11:21                     ` Yunsheng Lin
2021-09-16 11:57                     ` [Linuxarm] " Ilias Apalodimas
2021-09-17  3:57                       ` Yunsheng Lin
2021-09-17  6:38                         ` Ilias Apalodimas
2021-09-17  9:17                           ` Yunsheng Lin
2021-09-17 15:01                             ` Ilias Apalodimas
2021-09-18  1:43                               ` Yunsheng Lin
2021-09-18  9:23                                 ` Ilias Apalodimas
2021-09-22  3:38                                   ` Yunsheng Lin
2021-09-17 17:15             ` [Linuxarm] " Eric Dumazet
2021-09-18  2:42               ` Yunsheng Lin

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=ac16cc82-8d98-6a2c-b0a6-7c186808c72c@huawei.com \
    --to=linyunsheng@huawei.com \
    --cc=alexander.duyck@gmail.com \
    --cc=alobakin@pm.me \
    --cc=brouer@redhat.com \
    --cc=cong.wang@bytedance.com \
    --cc=davem@davemloft.net \
    --cc=dsahern@gmail.com \
    --cc=edumazet@google.com \
    --cc=elver@google.com \
    --cc=haokexin@gmail.com \
    --cc=hawk@kernel.org \
    --cc=ilias.apalodimas@linaro.org \
    --cc=jbrouer@redhat.com \
    --cc=jonathan.lemon@gmail.com \
    --cc=kuba@kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linuxarm@openeuler.org \
    --cc=memxor@gmail.com \
    --cc=netdev@vger.kernel.org \
    --cc=nogikh@google.com \
    --cc=pabeni@redhat.com \
    --cc=willemb@google.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).