All of lore.kernel.org
 help / color / mirror / Atom feed
From: "Huang\, Ying" <ying.huang@intel.com>
To: Minchan Kim <minchan@kernel.org>
Cc: "Huang\, Ying" <ying.huang@intel.com>,
	Andrew Morton <akpm@linux-foundation.org>, <linux-mm@kvack.org>,
	<linux-kernel@vger.kernel.org>, Hugh Dickins <hughd@google.com>,
	Shaohua Li <shli@kernel.org>, Rik van Riel <riel@redhat.com>
Subject: Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
Date: Thu, 20 Apr 2017 15:15:25 +0800	[thread overview]
Message-ID: <874lxjim7m.fsf@yhuang-dev.intel.com> (raw)
In-Reply-To: <20170420063834.GB3720@bbox> (Minchan Kim's message of "Thu, 20 Apr 2017 15:38:34 +0900")

Minchan Kim <minchan@kernel.org> writes:

> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> Minchan Kim <minchan@kernel.org> writes:
>> 
>> > Hi Huang,
>> >
>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> >> From: Huang Ying <ying.huang@intel.com>
>> >> 
>> >> To reduce the lock contention of swap_info_struct->lock when freeing
>> >> swap entry.  The freed swap entries will be collected in a per-CPU
>> >> buffer firstly, and be really freed later in batch.  During the batch
>> >> freeing, if the consecutive swap entries in the per-CPU buffer belongs
>> >> to same swap device, the swap_info_struct->lock needs to be
>> >> acquired/released only once, so that the lock contention could be
>> >> reduced greatly.  But if there are multiple swap devices, it is
>> >> possible that the lock may be unnecessarily released/acquired because
>> >> the swap entries belong to the same swap device are non-consecutive in
>> >> the per-CPU buffer.
>> >> 
>> >> To solve the issue, the per-CPU buffer is sorted according to the swap
>> >> device before freeing the swap entries.  Test shows that the time
>> >> spent by swapcache_free_entries() could be reduced after the patch.
>> >> 
>> >> Test the patch via measuring the run time of swap_cache_free_entries()
>> >> during the exit phase of the applications use much swap space.  The
>> >> results shows that the average run time of swap_cache_free_entries()
>> >> reduced about 20% after applying the patch.
>> >> 
>> >> Signed-off-by: Huang Ying <ying.huang@intel.com>
>> >> Acked-by: Tim Chen <tim.c.chen@intel.com>
>> >> Cc: Hugh Dickins <hughd@google.com>
>> >> Cc: Shaohua Li <shli@kernel.org>
>> >> Cc: Minchan Kim <minchan@kernel.org>
>> >> Cc: Rik van Riel <riel@redhat.com>
>> >> 
>> >> v3:
>> >> 
>> >> - Add some comments in code per Rik's suggestion.
>> >> 
>> >> v2:
>> >> 
>> >> - Avoid sort swap entries if there is only one swap device.
>> >> ---
>> >>  mm/swapfile.c | 12 ++++++++++++
>> >>  1 file changed, 12 insertions(+)
>> >> 
>> >> diff --git a/mm/swapfile.c b/mm/swapfile.c
>> >> index 90054f3c2cdc..f23c56e9be39 100644
>> >> --- a/mm/swapfile.c
>> >> +++ b/mm/swapfile.c
>> >> @@ -37,6 +37,7 @@
>> >>  #include <linux/swapfile.h>
>> >>  #include <linux/export.h>
>> >>  #include <linux/swap_slots.h>
>> >> +#include <linux/sort.h>
>> >>  
>> >>  #include <asm/pgtable.h>
>> >>  #include <asm/tlbflush.h>
>> >> @@ -1065,6 +1066,13 @@ void swapcache_free(swp_entry_t entry)
>> >>  	}
>> >>  }
>> >>  
>> >> +static int swp_entry_cmp(const void *ent1, const void *ent2)
>> >> +{
>> >> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
>> >> +
>> >> +	return (long)(swp_type(*e1) - swp_type(*e2));
>> >> +}
>> >> +
>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>  {
>> >>  	struct swap_info_struct *p, *prev;
>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>  
>> >>  	prev = NULL;
>> >>  	p = NULL;
>> >> +
>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> +	if (nr_swapfiles > 1)
>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >
>> > Let's think on other cases.
>> >
>> > There are two swaps and they are configured by priority so a swap's usage
>> > would be zero unless other swap used up. In case of that, this sorting
>> > is pointless.
>> >
>> > As well, nr_swapfiles is never decreased so if we enable multiple
>> > swaps and then disable until a swap is remained, this sorting is
>> > pointelss, too.
>> >
>> > How about lazy sorting approach? IOW, if we found prev != p and,
>> > then we can sort it.
>> 
>> Yes.  That should be better.  I just don't know whether the added
>> complexity is necessary, given the array is short and sort is fast.
>
> Huh?
>
> 1. swapon /dev/XXX1
> 2. swapon /dev/XXX2
> 3. swapoff /dev/XXX2
> 4. use only one swap
> 5. then, always pointless sort.

Yes.  In this situation we will do unnecessary sorting.  What I don't
know is whether the unnecessary sorting will hurt performance in real
life.  I can do some measurement.

Best Regards,
Huang, Ying

> Do not add such bogus code.
>
> Nacked.

WARNING: multiple messages have this Message-ID (diff)
From: "Huang\, Ying" <ying.huang@intel.com>
To: Minchan Kim <minchan@kernel.org>
Cc: "Huang, Ying" <ying.huang@intel.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	linux-mm@kvack.org, linux-kernel@vger.kernel.org,
	Hugh Dickins <hughd@google.com>, Shaohua Li <shli@kernel.org>,
	Rik van Riel <riel@redhat.com>
Subject: Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
Date: Thu, 20 Apr 2017 15:15:25 +0800	[thread overview]
Message-ID: <874lxjim7m.fsf@yhuang-dev.intel.com> (raw)
In-Reply-To: <20170420063834.GB3720@bbox> (Minchan Kim's message of "Thu, 20 Apr 2017 15:38:34 +0900")

Minchan Kim <minchan@kernel.org> writes:

> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> Minchan Kim <minchan@kernel.org> writes:
>> 
>> > Hi Huang,
>> >
>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> >> From: Huang Ying <ying.huang@intel.com>
>> >> 
>> >> To reduce the lock contention of swap_info_struct->lock when freeing
>> >> swap entry.  The freed swap entries will be collected in a per-CPU
>> >> buffer firstly, and be really freed later in batch.  During the batch
>> >> freeing, if the consecutive swap entries in the per-CPU buffer belongs
>> >> to same swap device, the swap_info_struct->lock needs to be
>> >> acquired/released only once, so that the lock contention could be
>> >> reduced greatly.  But if there are multiple swap devices, it is
>> >> possible that the lock may be unnecessarily released/acquired because
>> >> the swap entries belong to the same swap device are non-consecutive in
>> >> the per-CPU buffer.
>> >> 
>> >> To solve the issue, the per-CPU buffer is sorted according to the swap
>> >> device before freeing the swap entries.  Test shows that the time
>> >> spent by swapcache_free_entries() could be reduced after the patch.
>> >> 
>> >> Test the patch via measuring the run time of swap_cache_free_entries()
>> >> during the exit phase of the applications use much swap space.  The
>> >> results shows that the average run time of swap_cache_free_entries()
>> >> reduced about 20% after applying the patch.
>> >> 
>> >> Signed-off-by: Huang Ying <ying.huang@intel.com>
>> >> Acked-by: Tim Chen <tim.c.chen@intel.com>
>> >> Cc: Hugh Dickins <hughd@google.com>
>> >> Cc: Shaohua Li <shli@kernel.org>
>> >> Cc: Minchan Kim <minchan@kernel.org>
>> >> Cc: Rik van Riel <riel@redhat.com>
>> >> 
>> >> v3:
>> >> 
>> >> - Add some comments in code per Rik's suggestion.
>> >> 
>> >> v2:
>> >> 
>> >> - Avoid sort swap entries if there is only one swap device.
>> >> ---
>> >>  mm/swapfile.c | 12 ++++++++++++
>> >>  1 file changed, 12 insertions(+)
>> >> 
>> >> diff --git a/mm/swapfile.c b/mm/swapfile.c
>> >> index 90054f3c2cdc..f23c56e9be39 100644
>> >> --- a/mm/swapfile.c
>> >> +++ b/mm/swapfile.c
>> >> @@ -37,6 +37,7 @@
>> >>  #include <linux/swapfile.h>
>> >>  #include <linux/export.h>
>> >>  #include <linux/swap_slots.h>
>> >> +#include <linux/sort.h>
>> >>  
>> >>  #include <asm/pgtable.h>
>> >>  #include <asm/tlbflush.h>
>> >> @@ -1065,6 +1066,13 @@ void swapcache_free(swp_entry_t entry)
>> >>  	}
>> >>  }
>> >>  
>> >> +static int swp_entry_cmp(const void *ent1, const void *ent2)
>> >> +{
>> >> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
>> >> +
>> >> +	return (long)(swp_type(*e1) - swp_type(*e2));
>> >> +}
>> >> +
>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>  {
>> >>  	struct swap_info_struct *p, *prev;
>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>  
>> >>  	prev = NULL;
>> >>  	p = NULL;
>> >> +
>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> +	if (nr_swapfiles > 1)
>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >
>> > Let's think on other cases.
>> >
>> > There are two swaps and they are configured by priority so a swap's usage
>> > would be zero unless other swap used up. In case of that, this sorting
>> > is pointless.
>> >
>> > As well, nr_swapfiles is never decreased so if we enable multiple
>> > swaps and then disable until a swap is remained, this sorting is
>> > pointelss, too.
>> >
>> > How about lazy sorting approach? IOW, if we found prev != p and,
>> > then we can sort it.
>> 
>> Yes.  That should be better.  I just don't know whether the added
>> complexity is necessary, given the array is short and sort is fast.
>
> Huh?
>
> 1. swapon /dev/XXX1
> 2. swapon /dev/XXX2
> 3. swapoff /dev/XXX2
> 4. use only one swap
> 5. then, always pointless sort.

Yes.  In this situation we will do unnecessary sorting.  What I don't
know is whether the unnecessary sorting will hurt performance in real
life.  I can do some measurement.

Best Regards,
Huang, Ying

> Do not add such bogus code.
>
> Nacked.

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  reply	other threads:[~2017-04-20  7:15 UTC|newest]

Thread overview: 61+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2017-04-07  6:49 [PATCH -mm -v3] mm, swap: Sort swap entries before free Huang, Ying
2017-04-07  6:49 ` Huang, Ying
2017-04-07 13:05 ` Rik van Riel
2017-04-07 13:05   ` Rik van Riel
2017-04-07 21:43 ` Andrew Morton
2017-04-07 21:43   ` Andrew Morton
2017-04-11  7:03   ` Huang, Ying
2017-04-11  7:03     ` Huang, Ying
2017-04-14  1:36   ` Huang, Ying
2017-04-14  1:36     ` Huang, Ying
2017-04-14  1:41     ` Huang, Ying
2017-04-18  4:59 ` Minchan Kim
2017-04-18  4:59   ` Minchan Kim
2017-04-19  8:14   ` Huang, Ying
2017-04-19  8:14     ` Huang, Ying
2017-04-20  6:38     ` Minchan Kim
2017-04-20  6:38       ` Minchan Kim
2017-04-20  7:15       ` Huang, Ying [this message]
2017-04-20  7:15         ` Huang, Ying
2017-04-21 12:29         ` Huang, Ying
2017-04-21 12:29           ` Huang, Ying
2017-04-21 23:29           ` Tim Chen
2017-04-21 23:29             ` Tim Chen
2017-04-23 13:16             ` Huang, Ying
2017-04-23 13:16               ` Huang, Ying
2017-04-24 16:03               ` Tim Chen
2017-04-24 16:03                 ` Tim Chen
2017-04-24  4:52           ` Minchan Kim
2017-04-24  4:52             ` Minchan Kim
2017-04-24  6:47             ` Huang, Ying
2017-04-24  6:47               ` Huang, Ying
2017-04-26 12:42             ` Huang, Ying
2017-04-26 12:42               ` Huang, Ying
2017-04-26 20:13               ` Tim Chen
2017-04-26 20:13                 ` Tim Chen
2017-04-27  1:21                 ` Huang, Ying
2017-04-27  1:21                   ` Huang, Ying
2017-04-27 16:48                   ` Tim Chen
2017-04-27 16:48                     ` Tim Chen
2017-04-27  4:35               ` Minchan Kim
2017-04-27  4:35                 ` Minchan Kim
2017-04-28  1:09                 ` Huang, Ying
2017-04-28  1:09                   ` Huang, Ying
2017-04-28  7:42                   ` Minchan Kim
2017-04-28  7:42                     ` Minchan Kim
2017-04-28  8:05                     ` Huang, Ying
2017-04-28  8:05                       ` Huang, Ying
2017-04-28  9:00                       ` Minchan Kim
2017-04-28  9:00                         ` Minchan Kim
2017-04-28 11:48                         ` Huang, Ying
2017-04-28 11:48                           ` Huang, Ying
2017-04-28 13:35                           ` Huang, Ying
2017-04-28 13:35                             ` Huang, Ying
2017-05-02  5:02                             ` Minchan Kim
2017-05-02  5:02                               ` Minchan Kim
2017-05-02  5:35                               ` Huang, Ying
2017-05-02  5:35                                 ` Huang, Ying
2017-05-02  5:48                                 ` Minchan Kim
2017-05-02  5:48                                   ` Minchan Kim
2017-05-02  6:08                                   ` Huang, Ying
2017-05-02  6:08                                     ` Huang, Ying

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=874lxjim7m.fsf@yhuang-dev.intel.com \
    --to=ying.huang@intel.com \
    --cc=akpm@linux-foundation.org \
    --cc=hughd@google.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=minchan@kernel.org \
    --cc=riel@redhat.com \
    --cc=shli@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.