linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH -mm -v3] mm, swap: Sort swap entries before free
@ 2017-04-07  6:49 Huang, Ying
  2017-04-07 13:05 ` Rik van Riel
                   ` (2 more replies)
  0 siblings, 3 replies; 30+ messages in thread
From: Huang, Ying @ 2017-04-07  6:49 UTC (permalink / raw)
  To: Andrew Morton
  Cc: linux-mm, linux-kernel, Huang Ying, Hugh Dickins, Shaohua Li,
	Minchan Kim, Rik van Riel

From: Huang Ying <ying.huang@intel.com>

To reduce the lock contention of swap_info_struct->lock when freeing
swap entry.  The freed swap entries will be collected in a per-CPU
buffer firstly, and be really freed later in batch.  During the batch
freeing, if the consecutive swap entries in the per-CPU buffer belongs
to same swap device, the swap_info_struct->lock needs to be
acquired/released only once, so that the lock contention could be
reduced greatly.  But if there are multiple swap devices, it is
possible that the lock may be unnecessarily released/acquired because
the swap entries belong to the same swap device are non-consecutive in
the per-CPU buffer.

To solve the issue, the per-CPU buffer is sorted according to the swap
device before freeing the swap entries.  Test shows that the time
spent by swapcache_free_entries() could be reduced after the patch.

Test the patch via measuring the run time of swap_cache_free_entries()
during the exit phase of the applications use much swap space.  The
results shows that the average run time of swap_cache_free_entries()
reduced about 20% after applying the patch.

Signed-off-by: Huang Ying <ying.huang@intel.com>
Acked-by: Tim Chen <tim.c.chen@intel.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Shaohua Li <shli@kernel.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Rik van Riel <riel@redhat.com>

v3:

- Add some comments in code per Rik's suggestion.

v2:

- Avoid sort swap entries if there is only one swap device.
---
 mm/swapfile.c | 12 ++++++++++++
 1 file changed, 12 insertions(+)

diff --git a/mm/swapfile.c b/mm/swapfile.c
index 90054f3c2cdc..f23c56e9be39 100644
--- a/mm/swapfile.c
+++ b/mm/swapfile.c
@@ -37,6 +37,7 @@
 #include <linux/swapfile.h>
 #include <linux/export.h>
 #include <linux/swap_slots.h>
+#include <linux/sort.h>
 
 #include <asm/pgtable.h>
 #include <asm/tlbflush.h>
@@ -1065,6 +1066,13 @@ void swapcache_free(swp_entry_t entry)
 	}
 }
 
+static int swp_entry_cmp(const void *ent1, const void *ent2)
+{
+	const swp_entry_t *e1 = ent1, *e2 = ent2;
+
+	return (long)(swp_type(*e1) - swp_type(*e2));
+}
+
 void swapcache_free_entries(swp_entry_t *entries, int n)
 {
 	struct swap_info_struct *p, *prev;
@@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
 
 	prev = NULL;
 	p = NULL;
+
+	/* Sort swap entries by swap device, so each lock is only taken once. */
+	if (nr_swapfiles > 1)
+		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
 	for (i = 0; i < n; ++i) {
 		p = swap_info_get_cont(entries[i], prev);
 		if (p)
-- 
2.11.0

^ permalink raw reply related	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-07  6:49 [PATCH -mm -v3] mm, swap: Sort swap entries before free Huang, Ying
@ 2017-04-07 13:05 ` Rik van Riel
  2017-04-07 21:43 ` Andrew Morton
  2017-04-18  4:59 ` Minchan Kim
  2 siblings, 0 replies; 30+ messages in thread
From: Rik van Riel @ 2017-04-07 13:05 UTC (permalink / raw)
  To: Huang, Ying, Andrew Morton
  Cc: linux-mm, linux-kernel, Hugh Dickins, Shaohua Li, Minchan Kim

On Fri, 2017-04-07 at 14:49 +0800, Huang, Ying wrote:

> To solve the issue, the per-CPU buffer is sorted according to the
> swap
> device before freeing the swap entries.  Test shows that the time
> spent by swapcache_free_entries() could be reduced after the patch.
> 
> Test the patch via measuring the run time of
> swap_cache_free_entries()
> during the exit phase of the applications use much swap space.  The
> results shows that the average run time of swap_cache_free_entries()
> reduced about 20% after applying the patch.
> 
> Signed-off-by: Huang Ying <ying.huang@intel.com>
> Acked-by: Tim Chen <tim.c.chen@intel.com>
> Cc: Hugh Dickins <hughd@google.com>
> Cc: Shaohua Li <shli@kernel.org>
> Cc: Minchan Kim <minchan@kernel.org>
> Cc: Rik van Riel <riel@redhat.com>

Acked-by: Rik van Riel <riel@redhat.com>

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-07  6:49 [PATCH -mm -v3] mm, swap: Sort swap entries before free Huang, Ying
  2017-04-07 13:05 ` Rik van Riel
@ 2017-04-07 21:43 ` Andrew Morton
  2017-04-11  7:03   ` Huang, Ying
  2017-04-14  1:36   ` Huang, Ying
  2017-04-18  4:59 ` Minchan Kim
  2 siblings, 2 replies; 30+ messages in thread
From: Andrew Morton @ 2017-04-07 21:43 UTC (permalink / raw)
  To: Huang, Ying
  Cc: linux-mm, linux-kernel, Hugh Dickins, Shaohua Li, Minchan Kim,
	Rik van Riel

On Fri,  7 Apr 2017 14:49:01 +0800 "Huang, Ying" <ying.huang@intel.com> wrote:

> To reduce the lock contention of swap_info_struct->lock when freeing
> swap entry.  The freed swap entries will be collected in a per-CPU
> buffer firstly, and be really freed later in batch.  During the batch
> freeing, if the consecutive swap entries in the per-CPU buffer belongs
> to same swap device, the swap_info_struct->lock needs to be
> acquired/released only once, so that the lock contention could be
> reduced greatly.  But if there are multiple swap devices, it is
> possible that the lock may be unnecessarily released/acquired because
> the swap entries belong to the same swap device are non-consecutive in
> the per-CPU buffer.
> 
> To solve the issue, the per-CPU buffer is sorted according to the swap
> device before freeing the swap entries.  Test shows that the time
> spent by swapcache_free_entries() could be reduced after the patch.
> 
> Test the patch via measuring the run time of swap_cache_free_entries()
> during the exit phase of the applications use much swap space.  The
> results shows that the average run time of swap_cache_free_entries()
> reduced about 20% after applying the patch.

"20%" is useful info, but it is much better to present the absolute
numbers, please.  If it's "20% of one nanosecond" then the patch isn't
very interesting.  If it's "20% of 35 seconds" then we know we have
more work to do.

If there is indeed still a significant problem here then perhaps it
would be better to move the percpu swp_entry_t buffer into the
per-device structure swap_info_struct, so it becomes "per cpu, per
device".  That way we should be able to reduce contention further.

Or maybe we do something else - it all depends upon the significance of
this problem, which is why a full description of your measurements is
useful.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-07 21:43 ` Andrew Morton
@ 2017-04-11  7:03   ` Huang, Ying
  2017-04-14  1:36   ` Huang, Ying
  1 sibling, 0 replies; 30+ messages in thread
From: Huang, Ying @ 2017-04-11  7:03 UTC (permalink / raw)
  To: Andrew Morton
  Cc: Huang, Ying, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Minchan Kim, Rik van Riel

Andrew Morton <akpm@linux-foundation.org> writes:

> On Fri,  7 Apr 2017 14:49:01 +0800 "Huang, Ying" <ying.huang@intel.com> wrote:
>
>> To reduce the lock contention of swap_info_struct->lock when freeing
>> swap entry.  The freed swap entries will be collected in a per-CPU
>> buffer firstly, and be really freed later in batch.  During the batch
>> freeing, if the consecutive swap entries in the per-CPU buffer belongs
>> to same swap device, the swap_info_struct->lock needs to be
>> acquired/released only once, so that the lock contention could be
>> reduced greatly.  But if there are multiple swap devices, it is
>> possible that the lock may be unnecessarily released/acquired because
>> the swap entries belong to the same swap device are non-consecutive in
>> the per-CPU buffer.
>> 
>> To solve the issue, the per-CPU buffer is sorted according to the swap
>> device before freeing the swap entries.  Test shows that the time
>> spent by swapcache_free_entries() could be reduced after the patch.
>> 
>> Test the patch via measuring the run time of swap_cache_free_entries()
>> during the exit phase of the applications use much swap space.  The
>> results shows that the average run time of swap_cache_free_entries()
>> reduced about 20% after applying the patch.
>
> "20%" is useful info, but it is much better to present the absolute
> numbers, please.  If it's "20% of one nanosecond" then the patch isn't
> very interesting.  If it's "20% of 35 seconds" then we know we have
> more work to do.

The average run time of swap_cache_free_entries() is reduced from
about ~137us to ~111us. The total samples of swap_cache_free_entries()
is about 200000, run on 16 CPUs, so the wall time is about 1.7s.  I will
revise the tests to get the total run time reduction.

> If there is indeed still a significant problem here then perhaps it
> would be better to move the percpu swp_entry_t buffer into the
> per-device structure swap_info_struct, so it becomes "per cpu, per
> device".  That way we should be able to reduce contention further.
>
> Or maybe we do something else - it all depends upon the significance of
> this problem, which is why a full description of your measurements is
> useful.

Yes.  I will provide more and better measurement firstly.

Best Regards,
Huang, Ying

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-07 21:43 ` Andrew Morton
  2017-04-11  7:03   ` Huang, Ying
@ 2017-04-14  1:36   ` Huang, Ying
  1 sibling, 0 replies; 30+ messages in thread
From: Huang, Ying @ 2017-04-14  1:36 UTC (permalink / raw)
  To: Andrew Morton
  Cc: Huang, Ying, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Minchan Kim, Rik van Riel

Andrew Morton <akpm@linux-foundation.org> writes:

> On Fri,  7 Apr 2017 14:49:01 +0800 "Huang, Ying" <ying.huang@intel.com> wrote:
>
>> To reduce the lock contention of swap_info_struct->lock when freeing
>> swap entry.  The freed swap entries will be collected in a per-CPU
>> buffer firstly, and be really freed later in batch.  During the batch
>> freeing, if the consecutive swap entries in the per-CPU buffer belongs
>> to same swap device, the swap_info_struct->lock needs to be
>> acquired/released only once, so that the lock contention could be
>> reduced greatly.  But if there are multiple swap devices, it is
>> possible that the lock may be unnecessarily released/acquired because
>> the swap entries belong to the same swap device are non-consecutive in
>> the per-CPU buffer.
>> 
>> To solve the issue, the per-CPU buffer is sorted according to the swap
>> device before freeing the swap entries.  Test shows that the time
>> spent by swapcache_free_entries() could be reduced after the patch.
>> 
>> Test the patch via measuring the run time of swap_cache_free_entries()
>> during the exit phase of the applications use much swap space.  The
>> results shows that the average run time of swap_cache_free_entries()
>> reduced about 20% after applying the patch.
>
> "20%" is useful info, but it is much better to present the absolute
> numbers, please.  If it's "20% of one nanosecond" then the patch isn't
> very interesting.  If it's "20% of 35 seconds" then we know we have
> more work to do.

I added memory freeing timing capability to vm-scalability test suite.
The result shows the memory freeing time reduced from 2.64s to 2.31s
(about -12.5%).

Best Regards,
Huang, Ying

> If there is indeed still a significant problem here then perhaps it
> would be better to move the percpu swp_entry_t buffer into the
> per-device structure swap_info_struct, so it becomes "per cpu, per
> device".  That way we should be able to reduce contention further.
>
> Or maybe we do something else - it all depends upon the significance of
> this problem, which is why a full description of your measurements is
> useful.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-07  6:49 [PATCH -mm -v3] mm, swap: Sort swap entries before free Huang, Ying
  2017-04-07 13:05 ` Rik van Riel
  2017-04-07 21:43 ` Andrew Morton
@ 2017-04-18  4:59 ` Minchan Kim
  2017-04-19  8:14   ` Huang, Ying
  2 siblings, 1 reply; 30+ messages in thread
From: Minchan Kim @ 2017-04-18  4:59 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel

Hi Huang,

On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
> From: Huang Ying <ying.huang@intel.com>
> 
> To reduce the lock contention of swap_info_struct->lock when freeing
> swap entry.  The freed swap entries will be collected in a per-CPU
> buffer firstly, and be really freed later in batch.  During the batch
> freeing, if the consecutive swap entries in the per-CPU buffer belongs
> to same swap device, the swap_info_struct->lock needs to be
> acquired/released only once, so that the lock contention could be
> reduced greatly.  But if there are multiple swap devices, it is
> possible that the lock may be unnecessarily released/acquired because
> the swap entries belong to the same swap device are non-consecutive in
> the per-CPU buffer.
> 
> To solve the issue, the per-CPU buffer is sorted according to the swap
> device before freeing the swap entries.  Test shows that the time
> spent by swapcache_free_entries() could be reduced after the patch.
> 
> Test the patch via measuring the run time of swap_cache_free_entries()
> during the exit phase of the applications use much swap space.  The
> results shows that the average run time of swap_cache_free_entries()
> reduced about 20% after applying the patch.
> 
> Signed-off-by: Huang Ying <ying.huang@intel.com>
> Acked-by: Tim Chen <tim.c.chen@intel.com>
> Cc: Hugh Dickins <hughd@google.com>
> Cc: Shaohua Li <shli@kernel.org>
> Cc: Minchan Kim <minchan@kernel.org>
> Cc: Rik van Riel <riel@redhat.com>
> 
> v3:
> 
> - Add some comments in code per Rik's suggestion.
> 
> v2:
> 
> - Avoid sort swap entries if there is only one swap device.
> ---
>  mm/swapfile.c | 12 ++++++++++++
>  1 file changed, 12 insertions(+)
> 
> diff --git a/mm/swapfile.c b/mm/swapfile.c
> index 90054f3c2cdc..f23c56e9be39 100644
> --- a/mm/swapfile.c
> +++ b/mm/swapfile.c
> @@ -37,6 +37,7 @@
>  #include <linux/swapfile.h>
>  #include <linux/export.h>
>  #include <linux/swap_slots.h>
> +#include <linux/sort.h>
>  
>  #include <asm/pgtable.h>
>  #include <asm/tlbflush.h>
> @@ -1065,6 +1066,13 @@ void swapcache_free(swp_entry_t entry)
>  	}
>  }
>  
> +static int swp_entry_cmp(const void *ent1, const void *ent2)
> +{
> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
> +
> +	return (long)(swp_type(*e1) - swp_type(*e2));
> +}
> +
>  void swapcache_free_entries(swp_entry_t *entries, int n)
>  {
>  	struct swap_info_struct *p, *prev;
> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>  
>  	prev = NULL;
>  	p = NULL;
> +
> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> +	if (nr_swapfiles > 1)
> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);

Let's think on other cases.

There are two swaps and they are configured by priority so a swap's usage
would be zero unless other swap used up. In case of that, this sorting
is pointless.

As well, nr_swapfiles is never decreased so if we enable multiple
swaps and then disable until a swap is remained, this sorting is
pointelss, too.

How about lazy sorting approach? IOW, if we found prev != p and,
then we can sort it.

Thanks.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-18  4:59 ` Minchan Kim
@ 2017-04-19  8:14   ` Huang, Ying
  2017-04-20  6:38     ` Minchan Kim
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-04-19  8:14 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Minchan Kim <minchan@kernel.org> writes:

> Hi Huang,
>
> On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> From: Huang Ying <ying.huang@intel.com>
>> 
>> To reduce the lock contention of swap_info_struct->lock when freeing
>> swap entry.  The freed swap entries will be collected in a per-CPU
>> buffer firstly, and be really freed later in batch.  During the batch
>> freeing, if the consecutive swap entries in the per-CPU buffer belongs
>> to same swap device, the swap_info_struct->lock needs to be
>> acquired/released only once, so that the lock contention could be
>> reduced greatly.  But if there are multiple swap devices, it is
>> possible that the lock may be unnecessarily released/acquired because
>> the swap entries belong to the same swap device are non-consecutive in
>> the per-CPU buffer.
>> 
>> To solve the issue, the per-CPU buffer is sorted according to the swap
>> device before freeing the swap entries.  Test shows that the time
>> spent by swapcache_free_entries() could be reduced after the patch.
>> 
>> Test the patch via measuring the run time of swap_cache_free_entries()
>> during the exit phase of the applications use much swap space.  The
>> results shows that the average run time of swap_cache_free_entries()
>> reduced about 20% after applying the patch.
>> 
>> Signed-off-by: Huang Ying <ying.huang@intel.com>
>> Acked-by: Tim Chen <tim.c.chen@intel.com>
>> Cc: Hugh Dickins <hughd@google.com>
>> Cc: Shaohua Li <shli@kernel.org>
>> Cc: Minchan Kim <minchan@kernel.org>
>> Cc: Rik van Riel <riel@redhat.com>
>> 
>> v3:
>> 
>> - Add some comments in code per Rik's suggestion.
>> 
>> v2:
>> 
>> - Avoid sort swap entries if there is only one swap device.
>> ---
>>  mm/swapfile.c | 12 ++++++++++++
>>  1 file changed, 12 insertions(+)
>> 
>> diff --git a/mm/swapfile.c b/mm/swapfile.c
>> index 90054f3c2cdc..f23c56e9be39 100644
>> --- a/mm/swapfile.c
>> +++ b/mm/swapfile.c
>> @@ -37,6 +37,7 @@
>>  #include <linux/swapfile.h>
>>  #include <linux/export.h>
>>  #include <linux/swap_slots.h>
>> +#include <linux/sort.h>
>>  
>>  #include <asm/pgtable.h>
>>  #include <asm/tlbflush.h>
>> @@ -1065,6 +1066,13 @@ void swapcache_free(swp_entry_t entry)
>>  	}
>>  }
>>  
>> +static int swp_entry_cmp(const void *ent1, const void *ent2)
>> +{
>> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
>> +
>> +	return (long)(swp_type(*e1) - swp_type(*e2));
>> +}
>> +
>>  void swapcache_free_entries(swp_entry_t *entries, int n)
>>  {
>>  	struct swap_info_struct *p, *prev;
>> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>>  
>>  	prev = NULL;
>>  	p = NULL;
>> +
>> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> +	if (nr_swapfiles > 1)
>> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>
> Let's think on other cases.
>
> There are two swaps and they are configured by priority so a swap's usage
> would be zero unless other swap used up. In case of that, this sorting
> is pointless.
>
> As well, nr_swapfiles is never decreased so if we enable multiple
> swaps and then disable until a swap is remained, this sorting is
> pointelss, too.
>
> How about lazy sorting approach? IOW, if we found prev != p and,
> then we can sort it.

Yes.  That should be better.  I just don't know whether the added
complexity is necessary, given the array is short and sort is fast.

Best Regards,
Huang, Ying

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-19  8:14   ` Huang, Ying
@ 2017-04-20  6:38     ` Minchan Kim
  2017-04-20  7:15       ` Huang, Ying
  0 siblings, 1 reply; 30+ messages in thread
From: Minchan Kim @ 2017-04-20  6:38 UTC (permalink / raw)
  To: Huang, Ying, Andrew Morton
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel

On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
> Minchan Kim <minchan@kernel.org> writes:
> 
> > Hi Huang,
> >
> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
> >> From: Huang Ying <ying.huang@intel.com>
> >> 
> >> To reduce the lock contention of swap_info_struct->lock when freeing
> >> swap entry.  The freed swap entries will be collected in a per-CPU
> >> buffer firstly, and be really freed later in batch.  During the batch
> >> freeing, if the consecutive swap entries in the per-CPU buffer belongs
> >> to same swap device, the swap_info_struct->lock needs to be
> >> acquired/released only once, so that the lock contention could be
> >> reduced greatly.  But if there are multiple swap devices, it is
> >> possible that the lock may be unnecessarily released/acquired because
> >> the swap entries belong to the same swap device are non-consecutive in
> >> the per-CPU buffer.
> >> 
> >> To solve the issue, the per-CPU buffer is sorted according to the swap
> >> device before freeing the swap entries.  Test shows that the time
> >> spent by swapcache_free_entries() could be reduced after the patch.
> >> 
> >> Test the patch via measuring the run time of swap_cache_free_entries()
> >> during the exit phase of the applications use much swap space.  The
> >> results shows that the average run time of swap_cache_free_entries()
> >> reduced about 20% after applying the patch.
> >> 
> >> Signed-off-by: Huang Ying <ying.huang@intel.com>
> >> Acked-by: Tim Chen <tim.c.chen@intel.com>
> >> Cc: Hugh Dickins <hughd@google.com>
> >> Cc: Shaohua Li <shli@kernel.org>
> >> Cc: Minchan Kim <minchan@kernel.org>
> >> Cc: Rik van Riel <riel@redhat.com>
> >> 
> >> v3:
> >> 
> >> - Add some comments in code per Rik's suggestion.
> >> 
> >> v2:
> >> 
> >> - Avoid sort swap entries if there is only one swap device.
> >> ---
> >>  mm/swapfile.c | 12 ++++++++++++
> >>  1 file changed, 12 insertions(+)
> >> 
> >> diff --git a/mm/swapfile.c b/mm/swapfile.c
> >> index 90054f3c2cdc..f23c56e9be39 100644
> >> --- a/mm/swapfile.c
> >> +++ b/mm/swapfile.c
> >> @@ -37,6 +37,7 @@
> >>  #include <linux/swapfile.h>
> >>  #include <linux/export.h>
> >>  #include <linux/swap_slots.h>
> >> +#include <linux/sort.h>
> >>  
> >>  #include <asm/pgtable.h>
> >>  #include <asm/tlbflush.h>
> >> @@ -1065,6 +1066,13 @@ void swapcache_free(swp_entry_t entry)
> >>  	}
> >>  }
> >>  
> >> +static int swp_entry_cmp(const void *ent1, const void *ent2)
> >> +{
> >> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
> >> +
> >> +	return (long)(swp_type(*e1) - swp_type(*e2));
> >> +}
> >> +
> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
> >>  {
> >>  	struct swap_info_struct *p, *prev;
> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
> >>  
> >>  	prev = NULL;
> >>  	p = NULL;
> >> +
> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> >> +	if (nr_swapfiles > 1)
> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> >
> > Let's think on other cases.
> >
> > There are two swaps and they are configured by priority so a swap's usage
> > would be zero unless other swap used up. In case of that, this sorting
> > is pointless.
> >
> > As well, nr_swapfiles is never decreased so if we enable multiple
> > swaps and then disable until a swap is remained, this sorting is
> > pointelss, too.
> >
> > How about lazy sorting approach? IOW, if we found prev != p and,
> > then we can sort it.
> 
> Yes.  That should be better.  I just don't know whether the added
> complexity is necessary, given the array is short and sort is fast.

Huh?

1. swapon /dev/XXX1
2. swapon /dev/XXX2
3. swapoff /dev/XXX2
4. use only one swap
5. then, always pointless sort.

Do not add such bogus code.

Nacked.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-20  6:38     ` Minchan Kim
@ 2017-04-20  7:15       ` Huang, Ying
  2017-04-21 12:29         ` Huang, Ying
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-04-20  7:15 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Minchan Kim <minchan@kernel.org> writes:

> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> Minchan Kim <minchan@kernel.org> writes:
>> 
>> > Hi Huang,
>> >
>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> >> From: Huang Ying <ying.huang@intel.com>
>> >> 
>> >> To reduce the lock contention of swap_info_struct->lock when freeing
>> >> swap entry.  The freed swap entries will be collected in a per-CPU
>> >> buffer firstly, and be really freed later in batch.  During the batch
>> >> freeing, if the consecutive swap entries in the per-CPU buffer belongs
>> >> to same swap device, the swap_info_struct->lock needs to be
>> >> acquired/released only once, so that the lock contention could be
>> >> reduced greatly.  But if there are multiple swap devices, it is
>> >> possible that the lock may be unnecessarily released/acquired because
>> >> the swap entries belong to the same swap device are non-consecutive in
>> >> the per-CPU buffer.
>> >> 
>> >> To solve the issue, the per-CPU buffer is sorted according to the swap
>> >> device before freeing the swap entries.  Test shows that the time
>> >> spent by swapcache_free_entries() could be reduced after the patch.
>> >> 
>> >> Test the patch via measuring the run time of swap_cache_free_entries()
>> >> during the exit phase of the applications use much swap space.  The
>> >> results shows that the average run time of swap_cache_free_entries()
>> >> reduced about 20% after applying the patch.
>> >> 
>> >> Signed-off-by: Huang Ying <ying.huang@intel.com>
>> >> Acked-by: Tim Chen <tim.c.chen@intel.com>
>> >> Cc: Hugh Dickins <hughd@google.com>
>> >> Cc: Shaohua Li <shli@kernel.org>
>> >> Cc: Minchan Kim <minchan@kernel.org>
>> >> Cc: Rik van Riel <riel@redhat.com>
>> >> 
>> >> v3:
>> >> 
>> >> - Add some comments in code per Rik's suggestion.
>> >> 
>> >> v2:
>> >> 
>> >> - Avoid sort swap entries if there is only one swap device.
>> >> ---
>> >>  mm/swapfile.c | 12 ++++++++++++
>> >>  1 file changed, 12 insertions(+)
>> >> 
>> >> diff --git a/mm/swapfile.c b/mm/swapfile.c
>> >> index 90054f3c2cdc..f23c56e9be39 100644
>> >> --- a/mm/swapfile.c
>> >> +++ b/mm/swapfile.c
>> >> @@ -37,6 +37,7 @@
>> >>  #include <linux/swapfile.h>
>> >>  #include <linux/export.h>
>> >>  #include <linux/swap_slots.h>
>> >> +#include <linux/sort.h>
>> >>  
>> >>  #include <asm/pgtable.h>
>> >>  #include <asm/tlbflush.h>
>> >> @@ -1065,6 +1066,13 @@ void swapcache_free(swp_entry_t entry)
>> >>  	}
>> >>  }
>> >>  
>> >> +static int swp_entry_cmp(const void *ent1, const void *ent2)
>> >> +{
>> >> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
>> >> +
>> >> +	return (long)(swp_type(*e1) - swp_type(*e2));
>> >> +}
>> >> +
>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>  {
>> >>  	struct swap_info_struct *p, *prev;
>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>  
>> >>  	prev = NULL;
>> >>  	p = NULL;
>> >> +
>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> +	if (nr_swapfiles > 1)
>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >
>> > Let's think on other cases.
>> >
>> > There are two swaps and they are configured by priority so a swap's usage
>> > would be zero unless other swap used up. In case of that, this sorting
>> > is pointless.
>> >
>> > As well, nr_swapfiles is never decreased so if we enable multiple
>> > swaps and then disable until a swap is remained, this sorting is
>> > pointelss, too.
>> >
>> > How about lazy sorting approach? IOW, if we found prev != p and,
>> > then we can sort it.
>> 
>> Yes.  That should be better.  I just don't know whether the added
>> complexity is necessary, given the array is short and sort is fast.
>
> Huh?
>
> 1. swapon /dev/XXX1
> 2. swapon /dev/XXX2
> 3. swapoff /dev/XXX2
> 4. use only one swap
> 5. then, always pointless sort.

Yes.  In this situation we will do unnecessary sorting.  What I don't
know is whether the unnecessary sorting will hurt performance in real
life.  I can do some measurement.

Best Regards,
Huang, Ying

> Do not add such bogus code.
>
> Nacked.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-20  7:15       ` Huang, Ying
@ 2017-04-21 12:29         ` Huang, Ying
  2017-04-21 23:29           ` Tim Chen
  2017-04-24  4:52           ` Minchan Kim
  0 siblings, 2 replies; 30+ messages in thread
From: Huang, Ying @ 2017-04-21 12:29 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Minchan Kim, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

"Huang, Ying" <ying.huang@intel.com> writes:

> Minchan Kim <minchan@kernel.org> writes:
>
>> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>>> Minchan Kim <minchan@kernel.org> writes:
>>> 
>>> > Hi Huang,
>>> >
>>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>>> >> From: Huang Ying <ying.huang@intel.com>
>>> >> 
>>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>>> >>  {
>>> >>  	struct swap_info_struct *p, *prev;
>>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>>> >>  
>>> >>  	prev = NULL;
>>> >>  	p = NULL;
>>> >> +
>>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>>> >> +	if (nr_swapfiles > 1)
>>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>>> >
>>> > Let's think on other cases.
>>> >
>>> > There are two swaps and they are configured by priority so a swap's usage
>>> > would be zero unless other swap used up. In case of that, this sorting
>>> > is pointless.
>>> >
>>> > As well, nr_swapfiles is never decreased so if we enable multiple
>>> > swaps and then disable until a swap is remained, this sorting is
>>> > pointelss, too.
>>> >
>>> > How about lazy sorting approach? IOW, if we found prev != p and,
>>> > then we can sort it.
>>> 
>>> Yes.  That should be better.  I just don't know whether the added
>>> complexity is necessary, given the array is short and sort is fast.
>>
>> Huh?
>>
>> 1. swapon /dev/XXX1
>> 2. swapon /dev/XXX2
>> 3. swapoff /dev/XXX2
>> 4. use only one swap
>> 5. then, always pointless sort.
>
> Yes.  In this situation we will do unnecessary sorting.  What I don't
> know is whether the unnecessary sorting will hurt performance in real
> life.  I can do some measurement.

I tested the patch with 1 swap device and 1 process to eat memory
(remove the "if (nr_swapfiles > 1)" for test).  I think this is the
worse case because there is no lock contention.  The memory freeing time
increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
overhead for some cases.  I change the algorithm to something like
below,

 void swapcache_free_entries(swp_entry_t *entries, int n)
 {
 	struct swap_info_struct *p, *prev;
 	int i;
+	swp_entry_t entry;
+	unsigned int prev_swp_type;
 
 	if (n <= 0)
 		return;
 
+	prev_swp_type = swp_type(entries[0]);
+	for (i = n - 1; i > 0; i--) {
+		if (swp_type(entries[i]) != prev_swp_type)
+			break;
+	}
+
+	/* Sort swap entries by swap device, so each lock is only taken once. */
+	if (i)
+		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
 	prev = NULL;
 	p = NULL;
 	for (i = 0; i < n; ++i) {
-		p = swap_info_get_cont(entries[i], prev);
+		entry = entries[i];
+		p = swap_info_get_cont(entry, prev);
 		if (p)
-			swap_entry_free(p, entries[i]);
+			swap_entry_free(p, entry);
 		prev = p;
 	}
 	if (p)

With this patch, the memory freeing time increased from 1.94s to 1.97s.
I think this is good enough.  Do you think so?

I will send out the formal patch soon.

Best Regards,
Huang, Ying

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-21 12:29         ` Huang, Ying
@ 2017-04-21 23:29           ` Tim Chen
  2017-04-23 13:16             ` Huang, Ying
  2017-04-24  4:52           ` Minchan Kim
  1 sibling, 1 reply; 30+ messages in thread
From: Tim Chen @ 2017-04-21 23:29 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Minchan Kim, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

On Fri, 2017-04-21 at 20:29 +0800, Huang, Ying wrote:
> "Huang, Ying" <ying.huang@intel.com> writes:
> 
> > 
> > Minchan Kim <minchan@kernel.org> writes:
> > 
> > > 
> > > On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
> > > > 
> > > > Minchan Kim <minchan@kernel.org> writes:
> > > > 
> > > > > 
> > > > > Hi Huang,
> > > > > 
> > > > > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
> > > > > > 
> > > > > > From: Huang Ying <ying.huang@intel.com>
> > > > > > 
> > > > > >  void swapcache_free_entries(swp_entry_t *entries, int n)
> > > > > >  {
> > > > > >  	struct swap_info_struct *p, *prev;
> > > > > > @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
> > > > > >  
> > > > > >  	prev = NULL;
> > > > > >  	p = NULL;
> > > > > > +
> > > > > > +	/* Sort swap entries by swap device, so each lock is only taken once. */
> > > > > > +	if (nr_swapfiles > 1)
> > > > > > +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> > > > > Let's think on other cases.
> > > > > 
> > > > > There are two swaps and they are configured by priority so a swap's usage
> > > > > would be zero unless other swap used up. In case of that, this sorting
> > > > > is pointless.
> > > > > 
> > > > > As well, nr_swapfiles is never decreased so if we enable multiple
> > > > > swaps and then disable until a swap is remained, this sorting is
> > > > > pointelss, too.
> > > > > 
> > > > > How about lazy sorting approach? IOW, if we found prev != p and,
> > > > > then we can sort it.
> > > > Yes.  That should be better.  I just don't know whether the added
> > > > complexity is necessary, given the array is short and sort is fast.
> > > Huh?
> > > 
> > > 1. swapon /dev/XXX1
> > > 2. swapon /dev/XXX2
> > > 3. swapoff /dev/XXX2
> > > 4. use only one swap
> > > 5. then, always pointless sort.
> > Yes.  In this situation we will do unnecessary sorting.  What I don't
> > know is whether the unnecessary sorting will hurt performance in real
> > life.  I can do some measurement.
> I tested the patch with 1 swap device and 1 process to eat memory
> (remove the "if (nr_swapfiles > 1)" for test).  

It is possible that nr_swapfiles > 1 when we have only 1 swapfile due
to swapoff.  The nr_swapfiles never decrement on swapoff.
We will need to use another counter in alloc_swap_info and
swapoff to track the true number of swapfiles in use to have a fast path
that avoid the search and sort for the 1 swap case.

Tim

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-21 23:29           ` Tim Chen
@ 2017-04-23 13:16             ` Huang, Ying
  2017-04-24 16:03               ` Tim Chen
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-04-23 13:16 UTC (permalink / raw)
  To: Tim Chen
  Cc: Huang, Ying, Minchan Kim, Andrew Morton, linux-mm, linux-kernel,
	Hugh Dickins, Shaohua Li, Rik van Riel

Tim Chen <tim.c.chen@linux.intel.com> writes:

> On Fri, 2017-04-21 at 20:29 +0800, Huang, Ying wrote:
>> "Huang, Ying" <ying.huang@intel.com> writes:
>> 
>> > 
>> > Minchan Kim <minchan@kernel.org> writes:
>> > 
>> > > 
>> > > On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> > > > 
>> > > > Minchan Kim <minchan@kernel.org> writes:
>> > > > 
>> > > > > 
>> > > > > Hi Huang,
>> > > > > 
>> > > > > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> > > > > > 
>> > > > > > From: Huang Ying <ying.huang@intel.com>
>> > > > > > 
>> > > > > >  void swapcache_free_entries(swp_entry_t *entries, int n)
>> > > > > >  {
>> > > > > >  	struct swap_info_struct *p, *prev;
>> > > > > > @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> > > > > >  
>> > > > > >  	prev = NULL;
>> > > > > >  	p = NULL;
>> > > > > > +
>> > > > > > +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> > > > > > +	if (nr_swapfiles > 1)
>> > > > > > +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> > > > > Let's think on other cases.
>> > > > > 
>> > > > > There are two swaps and they are configured by priority so a swap's usage
>> > > > > would be zero unless other swap used up. In case of that, this sorting
>> > > > > is pointless.
>> > > > > 
>> > > > > As well, nr_swapfiles is never decreased so if we enable multiple
>> > > > > swaps and then disable until a swap is remained, this sorting is
>> > > > > pointelss, too.
>> > > > > 
>> > > > > How about lazy sorting approach? IOW, if we found prev != p and,
>> > > > > then we can sort it.
>> > > > Yes.  That should be better.  I just don't know whether the added
>> > > > complexity is necessary, given the array is short and sort is fast.
>> > > Huh?
>> > > 
>> > > 1. swapon /dev/XXX1
>> > > 2. swapon /dev/XXX2
>> > > 3. swapoff /dev/XXX2
>> > > 4. use only one swap
>> > > 5. then, always pointless sort.
>> > Yes.  In this situation we will do unnecessary sorting.  What I don't
>> > know is whether the unnecessary sorting will hurt performance in real
>> > life.  I can do some measurement.
>> I tested the patch with 1 swap device and 1 process to eat memory
>> (remove the "if (nr_swapfiles > 1)" for test).  
>
> It is possible that nr_swapfiles > 1 when we have only 1 swapfile due
> to swapoff.  The nr_swapfiles never decrement on swapoff.
> We will need to use another counter in alloc_swap_info and
> swapoff to track the true number of swapfiles in use to have a fast path
> that avoid the search and sort for the 1 swap case.

Yes.  That is a possible optimization.  But it doesn't cover another use
cases raised by Minchan (two swap device with different priority).  So
in general, we still need to check whether there are entries from
multiple swap devices in the array.  Given the cost of the checking code
is really low, I think maybe we can just always use the checking code.
Do you think so?

Best Regards,
Huang, Ying

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-21 12:29         ` Huang, Ying
  2017-04-21 23:29           ` Tim Chen
@ 2017-04-24  4:52           ` Minchan Kim
  2017-04-24  6:47             ` Huang, Ying
  2017-04-26 12:42             ` Huang, Ying
  1 sibling, 2 replies; 30+ messages in thread
From: Minchan Kim @ 2017-04-24  4:52 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel

On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
> "Huang, Ying" <ying.huang@intel.com> writes:
> 
> > Minchan Kim <minchan@kernel.org> writes:
> >
> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
> >>> Minchan Kim <minchan@kernel.org> writes:
> >>> 
> >>> > Hi Huang,
> >>> >
> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
> >>> >> From: Huang Ying <ying.huang@intel.com>
> >>> >> 
> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
> >>> >>  {
> >>> >>  	struct swap_info_struct *p, *prev;
> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
> >>> >>  
> >>> >>  	prev = NULL;
> >>> >>  	p = NULL;
> >>> >> +
> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> >>> >> +	if (nr_swapfiles > 1)
> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> >>> >
> >>> > Let's think on other cases.
> >>> >
> >>> > There are two swaps and they are configured by priority so a swap's usage
> >>> > would be zero unless other swap used up. In case of that, this sorting
> >>> > is pointless.
> >>> >
> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
> >>> > swaps and then disable until a swap is remained, this sorting is
> >>> > pointelss, too.
> >>> >
> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
> >>> > then we can sort it.
> >>> 
> >>> Yes.  That should be better.  I just don't know whether the added
> >>> complexity is necessary, given the array is short and sort is fast.
> >>
> >> Huh?
> >>
> >> 1. swapon /dev/XXX1
> >> 2. swapon /dev/XXX2
> >> 3. swapoff /dev/XXX2
> >> 4. use only one swap
> >> 5. then, always pointless sort.
> >
> > Yes.  In this situation we will do unnecessary sorting.  What I don't
> > know is whether the unnecessary sorting will hurt performance in real
> > life.  I can do some measurement.
> 
> I tested the patch with 1 swap device and 1 process to eat memory
> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
> worse case because there is no lock contention.  The memory freeing time
> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
> overhead for some cases.  I change the algorithm to something like
> below,
> 
>  void swapcache_free_entries(swp_entry_t *entries, int n)
>  {
>  	struct swap_info_struct *p, *prev;
>  	int i;
> +	swp_entry_t entry;
> +	unsigned int prev_swp_type;
>  
>  	if (n <= 0)
>  		return;
>  
> +	prev_swp_type = swp_type(entries[0]);
> +	for (i = n - 1; i > 0; i--) {
> +		if (swp_type(entries[i]) != prev_swp_type)
> +			break;
> +	}

That's really what I want to avoid. For many swap usecases,
it adds unnecessary overhead.

> +
> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> +	if (i)
> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>  	prev = NULL;
>  	p = NULL;
>  	for (i = 0; i < n; ++i) {
> -		p = swap_info_get_cont(entries[i], prev);
> +		entry = entries[i];
> +		p = swap_info_get_cont(entry, prev);
>  		if (p)
> -			swap_entry_free(p, entries[i]);
> +			swap_entry_free(p, entry);
>  		prev = p;
>  	}
>  	if (p)
> 
> With this patch, the memory freeing time increased from 1.94s to 1.97s.
> I think this is good enough.  Do you think so?

What I mean is as follows(I didn't test it at all):

With this, sort entries if we found multiple entries in current
entries. It adds some condition checks for non-multiple swap
usecase but it would be more cheaper than the sorting.
And it adds a [un]lock overhead for multiple swap usecase but
it should be a compromise for single-swap usecase which is more
popular.

diff --git a/mm/swapfile.c b/mm/swapfile.c
index f23c56e9be39..0d76a492786f 100644
--- a/mm/swapfile.c
+++ b/mm/swapfile.c
@@ -1073,30 +1073,40 @@ static int swp_entry_cmp(const void *ent1, const void *ent2)
 	return (long)(swp_type(*e1) - swp_type(*e2));
 }
 
-void swapcache_free_entries(swp_entry_t *entries, int n)
+void swapcache_free_entries(swp_entry_t *entries, int nr)
 {
-	struct swap_info_struct *p, *prev;
 	int i;
+	struct swap_info_struct *cur, *prev = NULL;
+	bool sorted = false;
 
-	if (n <= 0)
+	if (nr <= 0)
 		return;
 
-	prev = NULL;
-	p = NULL;
-
-	/* Sort swap entries by swap device, so each lock is only taken once. */
-	if (nr_swapfiles > 1)
-		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
-	for (i = 0; i < n; ++i) {
-		p = swap_info_get_cont(entries[i], prev);
-		if (p)
-			swap_entry_free(p, entries[i]);
-		else
+	for (i = 0; i < nr; i++) {
+		cur = swap_info_get_cont(entries[i], prev);
+		if (!cur)
 			break;
-		prev = p;
+		if (cur != prev && !sorted && prev) {
+			spin_unlock(&cur->lock);
+			/*
+			 * Sort swap entries by swap device,
+			 * so each lock is only taken once.
+			 */
+			sort(entries + i, nr - i,
+					sizeof(swp_entry_t),
+					swp_entry_cmp, NULL);
+			sorted = true;
+			prev = NULL;
+			i--;
+			continue;
+		}
+
+		swap_entry_free(cur, entries[i]);
+		prev = cur;
 	}
-	if (p)
-		spin_unlock(&p->lock);
+
+	if (cur)
+		spin_unlock(&cur->lock);
 }
 
 /*

^ permalink raw reply related	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-24  4:52           ` Minchan Kim
@ 2017-04-24  6:47             ` Huang, Ying
  2017-04-26 12:42             ` Huang, Ying
  1 sibling, 0 replies; 30+ messages in thread
From: Huang, Ying @ 2017-04-24  6:47 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Minchan Kim <minchan@kernel.org> writes:

> On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
>> "Huang, Ying" <ying.huang@intel.com> writes:
>> 
>> > Minchan Kim <minchan@kernel.org> writes:
>> >
>> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> >>> Minchan Kim <minchan@kernel.org> writes:
>> >>> 
>> >>> > Hi Huang,
>> >>> >
>> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> >>> >> From: Huang Ying <ying.huang@intel.com>
>> >>> >> 
>> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>> >>  {
>> >>> >>  	struct swap_info_struct *p, *prev;
>> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>> >>  
>> >>> >>  	prev = NULL;
>> >>> >>  	p = NULL;
>> >>> >> +
>> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >>> >> +	if (nr_swapfiles > 1)
>> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >>> >
>> >>> > Let's think on other cases.
>> >>> >
>> >>> > There are two swaps and they are configured by priority so a swap's usage
>> >>> > would be zero unless other swap used up. In case of that, this sorting
>> >>> > is pointless.
>> >>> >
>> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
>> >>> > swaps and then disable until a swap is remained, this sorting is
>> >>> > pointelss, too.
>> >>> >
>> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
>> >>> > then we can sort it.
>> >>> 
>> >>> Yes.  That should be better.  I just don't know whether the added
>> >>> complexity is necessary, given the array is short and sort is fast.
>> >>
>> >> Huh?
>> >>
>> >> 1. swapon /dev/XXX1
>> >> 2. swapon /dev/XXX2
>> >> 3. swapoff /dev/XXX2
>> >> 4. use only one swap
>> >> 5. then, always pointless sort.
>> >
>> > Yes.  In this situation we will do unnecessary sorting.  What I don't
>> > know is whether the unnecessary sorting will hurt performance in real
>> > life.  I can do some measurement.
>> 
>> I tested the patch with 1 swap device and 1 process to eat memory
>> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
>> worse case because there is no lock contention.  The memory freeing time
>> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
>> overhead for some cases.  I change the algorithm to something like
>> below,
>> 
>>  void swapcache_free_entries(swp_entry_t *entries, int n)
>>  {
>>  	struct swap_info_struct *p, *prev;
>>  	int i;
>> +	swp_entry_t entry;
>> +	unsigned int prev_swp_type;
>>  
>>  	if (n <= 0)
>>  		return;
>>  
>> +	prev_swp_type = swp_type(entries[0]);
>> +	for (i = n - 1; i > 0; i--) {
>> +		if (swp_type(entries[i]) != prev_swp_type)
>> +			break;
>> +	}
>
> That's really what I want to avoid. For many swap usecases,
> it adds unnecessary overhead.
>
>> +
>> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> +	if (i)
>> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>>  	prev = NULL;
>>  	p = NULL;
>>  	for (i = 0; i < n; ++i) {
>> -		p = swap_info_get_cont(entries[i], prev);
>> +		entry = entries[i];
>> +		p = swap_info_get_cont(entry, prev);
>>  		if (p)
>> -			swap_entry_free(p, entries[i]);
>> +			swap_entry_free(p, entry);
>>  		prev = p;
>>  	}
>>  	if (p)
>> 
>> With this patch, the memory freeing time increased from 1.94s to 1.97s.
>> I think this is good enough.  Do you think so?
>
> What I mean is as follows(I didn't test it at all):
>
> With this, sort entries if we found multiple entries in current
> entries. It adds some condition checks for non-multiple swap
> usecase but it would be more cheaper than the sorting.
> And it adds a [un]lock overhead for multiple swap usecase but
> it should be a compromise for single-swap usecase which is more
> popular.

Yes.  What I concerned is that one swap device may be locked twice
instead of once during the freeing.  I will give it some test.

Best Regards,
Huang, Ying

> diff --git a/mm/swapfile.c b/mm/swapfile.c
> index f23c56e9be39..0d76a492786f 100644
> --- a/mm/swapfile.c
> +++ b/mm/swapfile.c
> @@ -1073,30 +1073,40 @@ static int swp_entry_cmp(const void *ent1, const void *ent2)
>  	return (long)(swp_type(*e1) - swp_type(*e2));
>  }
>  
> -void swapcache_free_entries(swp_entry_t *entries, int n)
> +void swapcache_free_entries(swp_entry_t *entries, int nr)
>  {
> -	struct swap_info_struct *p, *prev;
>  	int i;
> +	struct swap_info_struct *cur, *prev = NULL;
> +	bool sorted = false;
>  
> -	if (n <= 0)
> +	if (nr <= 0)
>  		return;
>  
> -	prev = NULL;
> -	p = NULL;
> -
> -	/* Sort swap entries by swap device, so each lock is only taken once. */
> -	if (nr_swapfiles > 1)
> -		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> -	for (i = 0; i < n; ++i) {
> -		p = swap_info_get_cont(entries[i], prev);
> -		if (p)
> -			swap_entry_free(p, entries[i]);
> -		else
> +	for (i = 0; i < nr; i++) {
> +		cur = swap_info_get_cont(entries[i], prev);
> +		if (!cur)
>  			break;
> -		prev = p;
> +		if (cur != prev && !sorted && prev) {
> +			spin_unlock(&cur->lock);
> +			/*
> +			 * Sort swap entries by swap device,
> +			 * so each lock is only taken once.
> +			 */
> +			sort(entries + i, nr - i,
> +					sizeof(swp_entry_t),
> +					swp_entry_cmp, NULL);
> +			sorted = true;
> +			prev = NULL;
> +			i--;
> +			continue;
> +		}
> +
> +		swap_entry_free(cur, entries[i]);
> +		prev = cur;
>  	}
> -	if (p)
> -		spin_unlock(&p->lock);
> +
> +	if (cur)
> +		spin_unlock(&cur->lock);
>  }
>  
>  /*

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-23 13:16             ` Huang, Ying
@ 2017-04-24 16:03               ` Tim Chen
  0 siblings, 0 replies; 30+ messages in thread
From: Tim Chen @ 2017-04-24 16:03 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Minchan Kim, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

On Sun, 2017-04-23 at 21:16 +0800, Huang, Ying wrote:
> Tim Chen <tim.c.chen@linux.intel.com> writes:
> 
> > 
> > On Fri, 2017-04-21 at 20:29 +0800, Huang, Ying wrote:
> > > 
> > > "Huang, Ying" <ying.huang@intel.com> writes:
> > > 
> > > > 
> > > > 
> > > > Minchan Kim <minchan@kernel.org> writes:
> > > > 
> > > > > 
> > > > > 
> > > > > On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
> > > > > > 
> > > > > > 
> > > > > > Minchan Kim <minchan@kernel.org> writes:
> > > > > > 
> > > > > > > 
> > > > > > > 
> > > > > > > Hi Huang,
> > > > > > > 
> > > > > > > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
> > > > > > > > 
> > > > > > > > 
> > > > > > > > From: Huang Ying <ying.huang@intel.com>
> > > > > > > > 
> > > > > > > >  void swapcache_free_entries(swp_entry_t *entries, int n)
> > > > > > > >  {
> > > > > > > >  	struct swap_info_struct *p, *prev;
> > > > > > > > @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
> > > > > > > >  
> > > > > > > >  	prev = NULL;
> > > > > > > >  	p = NULL;
> > > > > > > > +
> > > > > > > > +	/* Sort swap entries by swap device, so each lock is only taken once. */
> > > > > > > > +	if (nr_swapfiles > 1)
> > > > > > > > +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> > > > > > > Let's think on other cases.
> > > > > > > 
> > > > > > > There are two swaps and they are configured by priority so a swap's usage
> > > > > > > would be zero unless other swap used up. In case of that, this sorting
> > > > > > > is pointless.
> > > > > > > 
> > > > > > > As well, nr_swapfiles is never decreased so if we enable multiple
> > > > > > > swaps and then disable until a swap is remained, this sorting is
> > > > > > > pointelss, too.
> > > > > > > 
> > > > > > > How about lazy sorting approach? IOW, if we found prev != p and,
> > > > > > > then we can sort it.
> > > > > > Yes.  That should be better.  I just don't know whether the added
> > > > > > complexity is necessary, given the array is short and sort is fast.
> > > > > Huh?
> > > > > 
> > > > > 1. swapon /dev/XXX1
> > > > > 2. swapon /dev/XXX2
> > > > > 3. swapoff /dev/XXX2
> > > > > 4. use only one swap
> > > > > 5. then, always pointless sort.
> > > > Yes.  In this situation we will do unnecessary sorting.  What I don't
> > > > know is whether the unnecessary sorting will hurt performance in real
> > > > life.  I can do some measurement.
> > > I tested the patch with 1 swap device and 1 process to eat memory
> > > (remove the "if (nr_swapfiles > 1)" for test).  
> > It is possible that nr_swapfiles > 1 when we have only 1 swapfile due
> > to swapoff.  The nr_swapfiles never decrement on swapoff.
> > We will need to use another counter in alloc_swap_info and
> > swapoff to track the true number of swapfiles in use to have a fast path
> > that avoid the search and sort for the 1 swap case.
> Yes.  That is a possible optimization.  But it doesn't cover another use
> cases raised by Minchan (two swap device with different priority).  So
> in general, we still need to check whether there are entries from
> multiple swap devices in the array.  Given the cost of the checking code
> is really low, I think maybe we can just always use the checking code.
> Do you think so?

The single swap case is very common. It will be better if we can bypass the
extra logic and cost for multiple swap.  Yes, we still need the proper
check to see if sort is necessary as you proposed for the multiple swap case.

Tim

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-24  4:52           ` Minchan Kim
  2017-04-24  6:47             ` Huang, Ying
@ 2017-04-26 12:42             ` Huang, Ying
  2017-04-26 20:13               ` Tim Chen
  2017-04-27  4:35               ` Minchan Kim
  1 sibling, 2 replies; 30+ messages in thread
From: Huang, Ying @ 2017-04-26 12:42 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Minchan Kim <minchan@kernel.org> writes:

> On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
>> "Huang, Ying" <ying.huang@intel.com> writes:
>> 
>> > Minchan Kim <minchan@kernel.org> writes:
>> >
>> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> >>> Minchan Kim <minchan@kernel.org> writes:
>> >>> 
>> >>> > Hi Huang,
>> >>> >
>> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> >>> >> From: Huang Ying <ying.huang@intel.com>
>> >>> >> 
>> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>> >>  {
>> >>> >>  	struct swap_info_struct *p, *prev;
>> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>> >>  
>> >>> >>  	prev = NULL;
>> >>> >>  	p = NULL;
>> >>> >> +
>> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >>> >> +	if (nr_swapfiles > 1)
>> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >>> >
>> >>> > Let's think on other cases.
>> >>> >
>> >>> > There are two swaps and they are configured by priority so a swap's usage
>> >>> > would be zero unless other swap used up. In case of that, this sorting
>> >>> > is pointless.
>> >>> >
>> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
>> >>> > swaps and then disable until a swap is remained, this sorting is
>> >>> > pointelss, too.
>> >>> >
>> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
>> >>> > then we can sort it.
>> >>> 
>> >>> Yes.  That should be better.  I just don't know whether the added
>> >>> complexity is necessary, given the array is short and sort is fast.
>> >>
>> >> Huh?
>> >>
>> >> 1. swapon /dev/XXX1
>> >> 2. swapon /dev/XXX2
>> >> 3. swapoff /dev/XXX2
>> >> 4. use only one swap
>> >> 5. then, always pointless sort.
>> >
>> > Yes.  In this situation we will do unnecessary sorting.  What I don't
>> > know is whether the unnecessary sorting will hurt performance in real
>> > life.  I can do some measurement.
>> 
>> I tested the patch with 1 swap device and 1 process to eat memory
>> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
>> worse case because there is no lock contention.  The memory freeing time
>> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
>> overhead for some cases.  I change the algorithm to something like
>> below,
>> 
>>  void swapcache_free_entries(swp_entry_t *entries, int n)
>>  {
>>  	struct swap_info_struct *p, *prev;
>>  	int i;
>> +	swp_entry_t entry;
>> +	unsigned int prev_swp_type;
>>  
>>  	if (n <= 0)
>>  		return;
>>  
>> +	prev_swp_type = swp_type(entries[0]);
>> +	for (i = n - 1; i > 0; i--) {
>> +		if (swp_type(entries[i]) != prev_swp_type)
>> +			break;
>> +	}
>
> That's really what I want to avoid. For many swap usecases,
> it adds unnecessary overhead.
>
>> +
>> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> +	if (i)
>> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>>  	prev = NULL;
>>  	p = NULL;
>>  	for (i = 0; i < n; ++i) {
>> -		p = swap_info_get_cont(entries[i], prev);
>> +		entry = entries[i];
>> +		p = swap_info_get_cont(entry, prev);
>>  		if (p)
>> -			swap_entry_free(p, entries[i]);
>> +			swap_entry_free(p, entry);
>>  		prev = p;
>>  	}
>>  	if (p)
>> 
>> With this patch, the memory freeing time increased from 1.94s to 1.97s.
>> I think this is good enough.  Do you think so?
>
> What I mean is as follows(I didn't test it at all):
>
> With this, sort entries if we found multiple entries in current
> entries. It adds some condition checks for non-multiple swap
> usecase but it would be more cheaper than the sorting.
> And it adds a [un]lock overhead for multiple swap usecase but
> it should be a compromise for single-swap usecase which is more
> popular.
>

How about the following solution?  It can avoid [un]lock overhead and
double lock issue for multiple swap user case and has good performance
for one swap user case too.

Best Regards,
Huang, Ying

>From 7bd903c42749c448ef6acbbdee8dcbc1c5b498b9 Mon Sep 17 00:00:00 2001
From: Huang Ying <ying.huang@intel.com>
Date: Thu, 23 Feb 2017 13:05:20 +0800
Subject: [PATCH -v5] mm, swap: Sort swap entries before free

To reduce the lock contention of swap_info_struct->lock when freeing
swap entry.  The freed swap entries will be collected in a per-CPU
buffer firstly, and be really freed later in batch.  During the batch
freeing, if the consecutive swap entries in the per-CPU buffer belongs
to same swap device, the swap_info_struct->lock needs to be
acquired/released only once, so that the lock contention could be
reduced greatly.  But if there are multiple swap devices, it is
possible that the lock may be unnecessarily released/acquired because
the swap entries belong to the same swap device are non-consecutive in
the per-CPU buffer.

To solve the issue, the per-CPU buffer is sorted according to the swap
device before freeing the swap entries.  Test shows that the time
spent by swapcache_free_entries() could be reduced after the patch.

With the patch, the memory (some swapped out) free time reduced
13.6% (from 2.59s to 2.28s) in the vm-scalability swap-w-rand test
case with 16 processes.  The test is done on a Xeon E5 v3 system.  The
swap device used is a RAM simulated PMEM (persistent memory) device.
To test swapping, the test case creates 16 processes, which allocate
and write to the anonymous pages until the RAM and part of the swap
device is used up, finally the memory (some swapped out) is freed
before exit.

Signed-off-by: Huang Ying <ying.huang@intel.com>
Acked-by: Tim Chen <tim.c.chen@intel.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Shaohua Li <shli@kernel.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Rik van Riel <riel@redhat.com>

v5:

- Use a smarter way to determine whether sort is necessary.

v4:

- Avoid unnecessary sort if all entries are from one swap device.

v3:

- Add some comments in code per Rik's suggestion.

v2:

- Avoid sort swap entries if there is only one swap device.
---
 mm/swapfile.c | 43 ++++++++++++++++++++++++++++++++++++++-----
 1 file changed, 38 insertions(+), 5 deletions(-)

diff --git a/mm/swapfile.c b/mm/swapfile.c
index 71890061f653..10e75f9e8ac1 100644
--- a/mm/swapfile.c
+++ b/mm/swapfile.c
@@ -37,6 +37,7 @@
 #include <linux/swapfile.h>
 #include <linux/export.h>
 #include <linux/swap_slots.h>
+#include <linux/sort.h>
 
 #include <asm/pgtable.h>
 #include <asm/tlbflush.h>
@@ -1065,20 +1066,52 @@ void swapcache_free(swp_entry_t entry)
 	}
 }
 
+static int swp_entry_cmp(const void *ent1, const void *ent2)
+{
+	const swp_entry_t *e1 = ent1, *e2 = ent2;
+
+	return (int)(swp_type(*e1) - swp_type(*e2));
+}
+
 void swapcache_free_entries(swp_entry_t *entries, int n)
 {
 	struct swap_info_struct *p, *prev;
-	int i;
+	int i, m;
+	swp_entry_t entry;
+	unsigned int prev_swp_type;
 
 	if (n <= 0)
 		return;
 
 	prev = NULL;
 	p = NULL;
-	for (i = 0; i < n; ++i) {
-		p = swap_info_get_cont(entries[i], prev);
-		if (p)
-			swap_entry_free(p, entries[i]);
+	m = 0;
+	prev_swp_type = swp_type(entries[0]);
+	for (i = 0; i < n; i++) {
+		entry = entries[i];
+		if (likely(swp_type(entry) == prev_swp_type)) {
+			p = swap_info_get_cont(entry, prev);
+			if (likely(p))
+				swap_entry_free(p, entry);
+			prev = p;
+		} else if (!m)
+			m = i;
+	}
+	if (p)
+		spin_unlock(&p->lock);
+	if (likely(!m))
+		return;
+
+	/* Sort swap entries by swap device, so each lock is only taken once. */
+	sort(entries + m, n - m, sizeof(entries[0]), swp_entry_cmp, NULL);
+	prev = NULL;
+	for (i = m; i < n; i++) {
+		entry = entries[i];
+		if (swp_type(entry) == prev_swp_type)
+			continue;
+		p = swap_info_get_cont(entry, prev);
+		if (likely(p))
+			swap_entry_free(p, entry);
 		prev = p;
 	}
 	if (p)
-- 
2.11.0

^ permalink raw reply related	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-26 12:42             ` Huang, Ying
@ 2017-04-26 20:13               ` Tim Chen
  2017-04-27  1:21                 ` Huang, Ying
  2017-04-27  4:35               ` Minchan Kim
  1 sibling, 1 reply; 30+ messages in thread
From: Tim Chen @ 2017-04-26 20:13 UTC (permalink / raw)
  To: Huang, Ying, Minchan Kim
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel


> 
> From 7bd903c42749c448ef6acbbdee8dcbc1c5b498b9 Mon Sep 17 00:00:00 2001
> From: Huang Ying <ying.huang@intel.com>
> Date: Thu, 23 Feb 2017 13:05:20 +0800
> Subject: [PATCH -v5] mm, swap: Sort swap entries before free
> 
> 
> ---
>  mm/swapfile.c | 43 ++++++++++++++++++++++++++++++++++++++-----
>  1 file changed, 38 insertions(+), 5 deletions(-)
> diff --git a/mm/swapfile.c b/mm/swapfile.c
> index 71890061f653..10e75f9e8ac1 100644
> --- a/mm/swapfile.c
> +++ b/mm/swapfile.c
> @@ -37,6 +37,7 @@
>  #include <linux/swapfile.h>
>  #include <linux/export.h>
>  #include <linux/swap_slots.h>
> +#include <linux/sort.h>
>  
>  #include <asm/pgtable.h>
>  #include <asm/tlbflush.h>
> @@ -1065,20 +1066,52 @@ void swapcache_free(swp_entry_t entry)
>  	}
>  }
>  
> +static int swp_entry_cmp(const void *ent1, const void *ent2)
> +{
> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
> +
> +	return (int)(swp_type(*e1) - swp_type(*e2));
> +}
> +
>  void swapcache_free_entries(swp_entry_t *entries, int n)
>  {
>  	struct swap_info_struct *p, *prev;
> -	int i;
> +	int i, m;
> +	swp_entry_t entry;
> +	unsigned int prev_swp_type;

I think it will be clearer to name prev_swp_type as first_swp_type
as this is the swp type of the first entry.

>  
>  	if (n <= 0)
>  		return;
>  
>  	prev = NULL;
>  	p = NULL;
> -	for (i = 0; i < n; ++i) {
> -		p = swap_info_get_cont(entries[i], prev);
> -		if (p)
> -			swap_entry_free(p, entries[i]);
> +	m = 0;
> +	prev_swp_type = swp_type(entries[0]);
> +	for (i = 0; i < n; i++) {
> +		entry = entries[i];
> +		if (likely(swp_type(entry) == prev_swp_type)) {
> +			p = swap_info_get_cont(entry, prev);
> +			if (likely(p))
> +				swap_entry_free(p, entry);
> +			prev = p;
> +		} else if (!m)
> +			m = i;
> +	}
> +	if (p)
> +		spin_unlock(&p->lock);
> +	if (likely(!m))
> +		return;
> +

We could still have prev_swp_type at the first entry after sorting.
and we can avoid an unlock/relock for this case if we do this:

	if (likely(!m)) {
		if (p)
			spin_unlock(&p->lock);
		return;
	}
		
> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> +	sort(entries + m, n - m, sizeof(entries[0]), swp_entry_cmp, NULL);
> +	prev = NULL;

Can eliminate prev=NULL if we adopt the above change.

> +	for (i = m; i < n; i++) {
> +		entry = entries[i];
> +		if (swp_type(entry) == prev_swp_type)
> +			continue;

The if/continue statement seems incorrect. When swp_type(entry) == prev_swp_type
we also need to free entry.  The if/continue statement should be deleted.

Say we have 3 entries with swp_type
1,2,1

We will get prev_swp_type as 1 and free the first entry
and sort the remaining two.  The last entry with
swp_type 1 will not be freed.

> +		p = swap_info_get_cont(entry, prev);
> +		if (likely(p))
> +			swap_entry_free(p, entry);
>  		prev = p;
>  	}
>  	if (p)

Thanks.

Tim

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-26 20:13               ` Tim Chen
@ 2017-04-27  1:21                 ` Huang, Ying
  2017-04-27 16:48                   ` Tim Chen
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-04-27  1:21 UTC (permalink / raw)
  To: Tim Chen
  Cc: Huang, Ying, Minchan Kim, Andrew Morton, linux-mm, linux-kernel,
	Hugh Dickins, Shaohua Li, Rik van Riel

Tim Chen <tim.c.chen@linux.intel.com> writes:

>> 
>> From 7bd903c42749c448ef6acbbdee8dcbc1c5b498b9 Mon Sep 17 00:00:00 2001
>> From: Huang Ying <ying.huang@intel.com>
>> Date: Thu, 23 Feb 2017 13:05:20 +0800
>> Subject: [PATCH -v5] mm, swap: Sort swap entries before free
>> 
>> 
>> ---
>>  mm/swapfile.c | 43 ++++++++++++++++++++++++++++++++++++++-----
>>  1 file changed, 38 insertions(+), 5 deletions(-)
>> diff --git a/mm/swapfile.c b/mm/swapfile.c
>> index 71890061f653..10e75f9e8ac1 100644
>> --- a/mm/swapfile.c
>> +++ b/mm/swapfile.c
>> @@ -37,6 +37,7 @@
>>  #include <linux/swapfile.h>
>>  #include <linux/export.h>
>>  #include <linux/swap_slots.h>
>> +#include <linux/sort.h>
>>  
>>  #include <asm/pgtable.h>
>>  #include <asm/tlbflush.h>
>> @@ -1065,20 +1066,52 @@ void swapcache_free(swp_entry_t entry)
>>  	}
>>  }
>>  
>> +static int swp_entry_cmp(const void *ent1, const void *ent2)
>> +{
>> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
>> +
>> +	return (int)(swp_type(*e1) - swp_type(*e2));
>> +}
>> +
>>  void swapcache_free_entries(swp_entry_t *entries, int n)
>>  {
>>  	struct swap_info_struct *p, *prev;
>> -	int i;
>> +	int i, m;
>> +	swp_entry_t entry;
>> +	unsigned int prev_swp_type;
>
> I think it will be clearer to name prev_swp_type as first_swp_type
> as this is the swp type of the first entry.

Yes.  That is better!  Will do that.

>>  
>>  	if (n <= 0)
>>  		return;
>>  
>>  	prev = NULL;
>>  	p = NULL;
>> -	for (i = 0; i < n; ++i) {
>> -		p = swap_info_get_cont(entries[i], prev);
>> -		if (p)
>> -			swap_entry_free(p, entries[i]);
>> +	m = 0;
>> +	prev_swp_type = swp_type(entries[0]);
>> +	for (i = 0; i < n; i++) {
>> +		entry = entries[i];
>> +		if (likely(swp_type(entry) == prev_swp_type)) {
>> +			p = swap_info_get_cont(entry, prev);
>> +			if (likely(p))
>> +				swap_entry_free(p, entry);
>> +			prev = p;
>> +		} else if (!m)
>> +			m = i;
>> +	}
>> +	if (p)
>> +		spin_unlock(&p->lock);
>> +	if (likely(!m))
>> +		return;
>> +
>
> We could still have prev_swp_type at the first entry after sorting.
> and we can avoid an unlock/relock for this case if we do this:
>
> 	if (likely(!m)) {
> 		if (p)
> 			spin_unlock(&p->lock);
> 		return;
> 	}
> 		
>> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> +	sort(entries + m, n - m, sizeof(entries[0]), swp_entry_cmp, NULL);
>> +	prev = NULL;
>
> Can eliminate prev=NULL if we adopt the above change.
>
>> +	for (i = m; i < n; i++) {
>> +		entry = entries[i];
>> +		if (swp_type(entry) == prev_swp_type)
>> +			continue;
>
> The if/continue statement seems incorrect. When swp_type(entry) == prev_swp_type
> we also need to free entry.  The if/continue statement should be deleted.
>
> Say we have 3 entries with swp_type
> 1,2,1
>
> We will get prev_swp_type as 1 and free the first entry
> and sort the remaining two.  The last entry with
> swp_type 1 will not be freed.

The first loop in the function will scan all elements of the array, so
the first and third entry will be freed in the first loop.  Then the the
second and the third entry will be sorted.  But all entries with the
same swap type (device) of the first entry needn't to be freed again.
The key point is that we will scan all elements of the array in the
first loop, record the first entry that has different swap type
(device).

Best Regards,
Huang, Ying

>> +		p = swap_info_get_cont(entry, prev);
>> +		if (likely(p))
>> +			swap_entry_free(p, entry);
>>  		prev = p;
>>  	}
>>  	if (p)
>
> Thanks.
>
> Tim

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-26 12:42             ` Huang, Ying
  2017-04-26 20:13               ` Tim Chen
@ 2017-04-27  4:35               ` Minchan Kim
  2017-04-28  1:09                 ` Huang, Ying
  1 sibling, 1 reply; 30+ messages in thread
From: Minchan Kim @ 2017-04-27  4:35 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel

On Wed, Apr 26, 2017 at 08:42:10PM +0800, Huang, Ying wrote:
> Minchan Kim <minchan@kernel.org> writes:
> 
> > On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
> >> "Huang, Ying" <ying.huang@intel.com> writes:
> >> 
> >> > Minchan Kim <minchan@kernel.org> writes:
> >> >
> >> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
> >> >>> Minchan Kim <minchan@kernel.org> writes:
> >> >>> 
> >> >>> > Hi Huang,
> >> >>> >
> >> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
> >> >>> >> From: Huang Ying <ying.huang@intel.com>
> >> >>> >> 
> >> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
> >> >>> >>  {
> >> >>> >>  	struct swap_info_struct *p, *prev;
> >> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
> >> >>> >>  
> >> >>> >>  	prev = NULL;
> >> >>> >>  	p = NULL;
> >> >>> >> +
> >> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> >> >>> >> +	if (nr_swapfiles > 1)
> >> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> >> >>> >
> >> >>> > Let's think on other cases.
> >> >>> >
> >> >>> > There are two swaps and they are configured by priority so a swap's usage
> >> >>> > would be zero unless other swap used up. In case of that, this sorting
> >> >>> > is pointless.
> >> >>> >
> >> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
> >> >>> > swaps and then disable until a swap is remained, this sorting is
> >> >>> > pointelss, too.
> >> >>> >
> >> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
> >> >>> > then we can sort it.
> >> >>> 
> >> >>> Yes.  That should be better.  I just don't know whether the added
> >> >>> complexity is necessary, given the array is short and sort is fast.
> >> >>
> >> >> Huh?
> >> >>
> >> >> 1. swapon /dev/XXX1
> >> >> 2. swapon /dev/XXX2
> >> >> 3. swapoff /dev/XXX2
> >> >> 4. use only one swap
> >> >> 5. then, always pointless sort.
> >> >
> >> > Yes.  In this situation we will do unnecessary sorting.  What I don't
> >> > know is whether the unnecessary sorting will hurt performance in real
> >> > life.  I can do some measurement.
> >> 
> >> I tested the patch with 1 swap device and 1 process to eat memory
> >> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
> >> worse case because there is no lock contention.  The memory freeing time
> >> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
> >> overhead for some cases.  I change the algorithm to something like
> >> below,
> >> 
> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
> >>  {
> >>  	struct swap_info_struct *p, *prev;
> >>  	int i;
> >> +	swp_entry_t entry;
> >> +	unsigned int prev_swp_type;
> >>  
> >>  	if (n <= 0)
> >>  		return;
> >>  
> >> +	prev_swp_type = swp_type(entries[0]);
> >> +	for (i = n - 1; i > 0; i--) {
> >> +		if (swp_type(entries[i]) != prev_swp_type)
> >> +			break;
> >> +	}
> >
> > That's really what I want to avoid. For many swap usecases,
> > it adds unnecessary overhead.
> >
> >> +
> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> >> +	if (i)
> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> >>  	prev = NULL;
> >>  	p = NULL;
> >>  	for (i = 0; i < n; ++i) {
> >> -		p = swap_info_get_cont(entries[i], prev);
> >> +		entry = entries[i];
> >> +		p = swap_info_get_cont(entry, prev);
> >>  		if (p)
> >> -			swap_entry_free(p, entries[i]);
> >> +			swap_entry_free(p, entry);
> >>  		prev = p;
> >>  	}
> >>  	if (p)
> >> 
> >> With this patch, the memory freeing time increased from 1.94s to 1.97s.
> >> I think this is good enough.  Do you think so?
> >
> > What I mean is as follows(I didn't test it at all):
> >
> > With this, sort entries if we found multiple entries in current
> > entries. It adds some condition checks for non-multiple swap
> > usecase but it would be more cheaper than the sorting.
> > And it adds a [un]lock overhead for multiple swap usecase but
> > it should be a compromise for single-swap usecase which is more
> > popular.
> >
> 
> How about the following solution?  It can avoid [un]lock overhead and
> double lock issue for multiple swap user case and has good performance
> for one swap user case too.

How worse with approach I suggested compared to as-is?
Unless it's too bad, let's not add more complicated thing to just
enhance the minor usecase in such even *slow* path.
It adds code size/maintainance overead.
With your suggestion, it might enhance a bit with speicific benchmark
but not sure it's really worth for real practice.

> 
> Best Regards,
> Huang, Ying
> 
> From 7bd903c42749c448ef6acbbdee8dcbc1c5b498b9 Mon Sep 17 00:00:00 2001
> From: Huang Ying <ying.huang@intel.com>
> Date: Thu, 23 Feb 2017 13:05:20 +0800
> Subject: [PATCH -v5] mm, swap: Sort swap entries before free
> 
> To reduce the lock contention of swap_info_struct->lock when freeing
> swap entry.  The freed swap entries will be collected in a per-CPU
> buffer firstly, and be really freed later in batch.  During the batch
> freeing, if the consecutive swap entries in the per-CPU buffer belongs
> to same swap device, the swap_info_struct->lock needs to be
> acquired/released only once, so that the lock contention could be
> reduced greatly.  But if there are multiple swap devices, it is
> possible that the lock may be unnecessarily released/acquired because
> the swap entries belong to the same swap device are non-consecutive in
> the per-CPU buffer.
> 
> To solve the issue, the per-CPU buffer is sorted according to the swap
> device before freeing the swap entries.  Test shows that the time
> spent by swapcache_free_entries() could be reduced after the patch.
> 
> With the patch, the memory (some swapped out) free time reduced
> 13.6% (from 2.59s to 2.28s) in the vm-scalability swap-w-rand test
> case with 16 processes.  The test is done on a Xeon E5 v3 system.  The
> swap device used is a RAM simulated PMEM (persistent memory) device.
> To test swapping, the test case creates 16 processes, which allocate
> and write to the anonymous pages until the RAM and part of the swap
> device is used up, finally the memory (some swapped out) is freed
> before exit.
> 
> Signed-off-by: Huang Ying <ying.huang@intel.com>
> Acked-by: Tim Chen <tim.c.chen@intel.com>
> Cc: Hugh Dickins <hughd@google.com>
> Cc: Shaohua Li <shli@kernel.org>
> Cc: Minchan Kim <minchan@kernel.org>
> Cc: Rik van Riel <riel@redhat.com>
> 
> v5:
> 
> - Use a smarter way to determine whether sort is necessary.
> 
> v4:
> 
> - Avoid unnecessary sort if all entries are from one swap device.
> 
> v3:
> 
> - Add some comments in code per Rik's suggestion.
> 
> v2:
> 
> - Avoid sort swap entries if there is only one swap device.
> ---
>  mm/swapfile.c | 43 ++++++++++++++++++++++++++++++++++++++-----
>  1 file changed, 38 insertions(+), 5 deletions(-)
> 
> diff --git a/mm/swapfile.c b/mm/swapfile.c
> index 71890061f653..10e75f9e8ac1 100644
> --- a/mm/swapfile.c
> +++ b/mm/swapfile.c
> @@ -37,6 +37,7 @@
>  #include <linux/swapfile.h>
>  #include <linux/export.h>
>  #include <linux/swap_slots.h>
> +#include <linux/sort.h>
>  
>  #include <asm/pgtable.h>
>  #include <asm/tlbflush.h>
> @@ -1065,20 +1066,52 @@ void swapcache_free(swp_entry_t entry)
>  	}
>  }
>  
> +static int swp_entry_cmp(const void *ent1, const void *ent2)
> +{
> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
> +
> +	return (int)(swp_type(*e1) - swp_type(*e2));
> +}
> +
>  void swapcache_free_entries(swp_entry_t *entries, int n)
>  {
>  	struct swap_info_struct *p, *prev;
> -	int i;
> +	int i, m;
> +	swp_entry_t entry;
> +	unsigned int prev_swp_type;
>  
>  	if (n <= 0)
>  		return;
>  
>  	prev = NULL;
>  	p = NULL;
> -	for (i = 0; i < n; ++i) {
> -		p = swap_info_get_cont(entries[i], prev);
> -		if (p)
> -			swap_entry_free(p, entries[i]);
> +	m = 0;
> +	prev_swp_type = swp_type(entries[0]);
> +	for (i = 0; i < n; i++) {
> +		entry = entries[i];
> +		if (likely(swp_type(entry) == prev_swp_type)) {
> +			p = swap_info_get_cont(entry, prev);
> +			if (likely(p))
> +				swap_entry_free(p, entry);
> +			prev = p;
> +		} else if (!m)
> +			m = i;
> +	}
> +	if (p)
> +		spin_unlock(&p->lock);
> +	if (likely(!m))
> +		return;
> +
> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> +	sort(entries + m, n - m, sizeof(entries[0]), swp_entry_cmp, NULL);
> +	prev = NULL;
> +	for (i = m; i < n; i++) {
> +		entry = entries[i];
> +		if (swp_type(entry) == prev_swp_type)
> +			continue;
> +		p = swap_info_get_cont(entry, prev);
> +		if (likely(p))
> +			swap_entry_free(p, entry);
>  		prev = p;
>  	}
>  	if (p)
> -- 
> 2.11.0
> 

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-27  1:21                 ` Huang, Ying
@ 2017-04-27 16:48                   ` Tim Chen
  0 siblings, 0 replies; 30+ messages in thread
From: Tim Chen @ 2017-04-27 16:48 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Minchan Kim, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

On Thu, 2017-04-27 at 09:21 +0800, Huang, Ying wrote:
> Tim Chen <tim.c.chen@linux.intel.com> writes:
> 
> > 
> > > 
> > > 
> > > From 7bd903c42749c448ef6acbbdee8dcbc1c5b498b9 Mon Sep 17 00:00:00 2001
> > > From: Huang Ying <ying.huang@intel.com>
> > > Date: Thu, 23 Feb 2017 13:05:20 +0800
> > > Subject: [PATCH -v5] mm, swap: Sort swap entries before free
> > > 
> > >  
> > > ---
> > >  mm/swapfile.c | 43 ++++++++++++++++++++++++++++++++++++++-----
> > >  1 file changed, 38 insertions(+), 5 deletions(-)
> > > diff --git a/mm/swapfile.c b/mm/swapfile.c
> > > index 71890061f653..10e75f9e8ac1 100644
> > > --- a/mm/swapfile.c
> > > +++ b/mm/swapfile.c
> > > @@ -37,6 +37,7 @@
> > >  #include <linux/swapfile.h>
> > >  #include <linux/export.h>
> > >  #include <linux/swap_slots.h>
> > > +#include <linux/sort.h>
> > >  
> > >  #include <asm/pgtable.h>
> > >  #include <asm/tlbflush.h>
> > > @@ -1065,20 +1066,52 @@ void swapcache_free(swp_entry_t entry)
> > >  	}
> > >  }
> > >  
> > > +static int swp_entry_cmp(const void *ent1, const void *ent2)
> > > +{
> > > +	const swp_entry_t *e1 = ent1, *e2 = ent2;
> > > +
> > > +	return (int)(swp_type(*e1) - swp_type(*e2));
> > > +}
> > > +
> > >  void swapcache_free_entries(swp_entry_t *entries, int n)
> > >  {
> > >  	struct swap_info_struct *p, *prev;
> > > -	int i;
> > > +	int i, m;
> > > +	swp_entry_t entry;
> > > +	unsigned int prev_swp_type;
> > I think it will be clearer to name prev_swp_type as first_swp_type
> > as this is the swp type of the first entry.
> Yes.  That is better!  Will do that.
> 
> > 
> > > 
> > >  
> > >  	if (n <= 0)
> > >  		return;
> > >  
> > >  	prev = NULL;
> > >  	p = NULL;
> > > -	for (i = 0; i < n; ++i) {
> > > -		p = swap_info_get_cont(entries[i], prev);
> > > -		if (p)
> > > -			swap_entry_free(p, entries[i]);
> > > +	m = 0;
> > > +	prev_swp_type = swp_type(entries[0]);
> > > +	for (i = 0; i < n; i++) {
> > > +		entry = entries[i];
> > > +		if (likely(swp_type(entry) == prev_swp_type)) {
> > > +			p = swap_info_get_cont(entry, prev);
> > > +			if (likely(p))
> > > +				swap_entry_free(p, entry);
> > > +			prev = p;
> > > +		} else if (!m)
> > > +			m = i;
> > > +	}
> > > +	if (p)
> > > +		spin_unlock(&p->lock);
> > > +	if (likely(!m))
> > > +		return;
> > > +
> > We could still have prev_swp_type at the first entry after sorting.
> > and we can avoid an unlock/relock for this case if we do this:
> > 
> > 	if (likely(!m)) {
> > 		if (p)
> > 			spin_unlock(&p->lock);
> > 		return;
> > 	}
> > 		
> > > 
> > > +	/* Sort swap entries by swap device, so each lock is only taken once. */
> > > +	sort(entries + m, n - m, sizeof(entries[0]), swp_entry_cmp, NULL);
> > > +	prev = NULL;
> > Can eliminate prev=NULL if we adopt the above change.
> > 
> > > 
> > > +	for (i = m; i < n; i++) {
> > > +		entry = entries[i];
> > > +		if (swp_type(entry) == prev_swp_type)
> > > +			continue;
> > The if/continue statement seems incorrect. When swp_type(entry) == prev_swp_type
> > we also need to free entry.  The if/continue statement should be deleted.
> > 
> > Say we have 3 entries with swp_type
> > 1,2,1
> > 
> > We will get prev_swp_type as 1 and free the first entry
> > and sort the remaining two.  The last entry with
> > swp_type 1 will not be freed.
> The first loop in the function will scan all elements of the array, so
> the first and third entry will be freed in the first loop.  Then the the
> second and the third entry will be sorted.  But all entries with the
> same swap type (device) of the first entry needn't to be freed again.
> The key point is that we will scan all elements of the array in the
> first loop, record the first entry that has different swap type
> (device).

I was under the wrong impression that the code break from the first
loop when it finds a different swp type.  Yes, we should skip the
free in the second loop if the first loop scan the whole list.

Thanks.

Tim

> 
> Best Regards,
> Huang, Ying
> 
> > 
> > > 
> > > +		p = swap_info_get_cont(entry, prev);
> > > +		if (likely(p))
> > > +			swap_entry_free(p, entry);
> > >  		prev = p;
> > >  	}
> > >  	if (p)
> > Thanks.
> > 
> > Tim

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-27  4:35               ` Minchan Kim
@ 2017-04-28  1:09                 ` Huang, Ying
  2017-04-28  7:42                   ` Minchan Kim
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-04-28  1:09 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Minchan Kim <minchan@kernel.org> writes:

> On Wed, Apr 26, 2017 at 08:42:10PM +0800, Huang, Ying wrote:
>> Minchan Kim <minchan@kernel.org> writes:
>> 
>> > On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
>> >> "Huang, Ying" <ying.huang@intel.com> writes:
>> >> 
>> >> > Minchan Kim <minchan@kernel.org> writes:
>> >> >
>> >> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> >> >>> Minchan Kim <minchan@kernel.org> writes:
>> >> >>> 
>> >> >>> > Hi Huang,
>> >> >>> >
>> >> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> >> >>> >> From: Huang Ying <ying.huang@intel.com>
>> >> >>> >> 
>> >> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >> >>> >>  {
>> >> >>> >>  	struct swap_info_struct *p, *prev;
>> >> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> >> >>> >>  
>> >> >>> >>  	prev = NULL;
>> >> >>> >>  	p = NULL;
>> >> >>> >> +
>> >> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> >>> >> +	if (nr_swapfiles > 1)
>> >> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >> >>> >
>> >> >>> > Let's think on other cases.
>> >> >>> >
>> >> >>> > There are two swaps and they are configured by priority so a swap's usage
>> >> >>> > would be zero unless other swap used up. In case of that, this sorting
>> >> >>> > is pointless.
>> >> >>> >
>> >> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
>> >> >>> > swaps and then disable until a swap is remained, this sorting is
>> >> >>> > pointelss, too.
>> >> >>> >
>> >> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
>> >> >>> > then we can sort it.
>> >> >>> 
>> >> >>> Yes.  That should be better.  I just don't know whether the added
>> >> >>> complexity is necessary, given the array is short and sort is fast.
>> >> >>
>> >> >> Huh?
>> >> >>
>> >> >> 1. swapon /dev/XXX1
>> >> >> 2. swapon /dev/XXX2
>> >> >> 3. swapoff /dev/XXX2
>> >> >> 4. use only one swap
>> >> >> 5. then, always pointless sort.
>> >> >
>> >> > Yes.  In this situation we will do unnecessary sorting.  What I don't
>> >> > know is whether the unnecessary sorting will hurt performance in real
>> >> > life.  I can do some measurement.
>> >> 
>> >> I tested the patch with 1 swap device and 1 process to eat memory
>> >> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
>> >> worse case because there is no lock contention.  The memory freeing time
>> >> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
>> >> overhead for some cases.  I change the algorithm to something like
>> >> below,
>> >> 
>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >>  {
>> >>  	struct swap_info_struct *p, *prev;
>> >>  	int i;
>> >> +	swp_entry_t entry;
>> >> +	unsigned int prev_swp_type;
>> >>  
>> >>  	if (n <= 0)
>> >>  		return;
>> >>  
>> >> +	prev_swp_type = swp_type(entries[0]);
>> >> +	for (i = n - 1; i > 0; i--) {
>> >> +		if (swp_type(entries[i]) != prev_swp_type)
>> >> +			break;
>> >> +	}
>> >
>> > That's really what I want to avoid. For many swap usecases,
>> > it adds unnecessary overhead.
>> >
>> >> +
>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> +	if (i)
>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >>  	prev = NULL;
>> >>  	p = NULL;
>> >>  	for (i = 0; i < n; ++i) {
>> >> -		p = swap_info_get_cont(entries[i], prev);
>> >> +		entry = entries[i];
>> >> +		p = swap_info_get_cont(entry, prev);
>> >>  		if (p)
>> >> -			swap_entry_free(p, entries[i]);
>> >> +			swap_entry_free(p, entry);
>> >>  		prev = p;
>> >>  	}
>> >>  	if (p)
>> >> 
>> >> With this patch, the memory freeing time increased from 1.94s to 1.97s.
>> >> I think this is good enough.  Do you think so?
>> >
>> > What I mean is as follows(I didn't test it at all):
>> >
>> > With this, sort entries if we found multiple entries in current
>> > entries. It adds some condition checks for non-multiple swap
>> > usecase but it would be more cheaper than the sorting.
>> > And it adds a [un]lock overhead for multiple swap usecase but
>> > it should be a compromise for single-swap usecase which is more
>> > popular.
>> >
>> 
>> How about the following solution?  It can avoid [un]lock overhead and
>> double lock issue for multiple swap user case and has good performance
>> for one swap user case too.
>
> How worse with approach I suggested compared to as-is?

The performance difference between your version and my version is small
for my testing.

> Unless it's too bad, let's not add more complicated thing to just
> enhance the minor usecase in such even *slow* path.
> It adds code size/maintainance overead.
> With your suggestion, it might enhance a bit with speicific benchmark
> but not sure it's really worth for real practice.

I don't think the code complexity has much difference between our latest
versions.  As for complexity, I think my original version which just
uses nr_swapfiles to avoid sort() for single swap device is simple and
good enough for this task.  Maybe we can just improve the correctness of
swap device counting as Tim suggested.

Best Regards,
Huang, Ying

>> 
>> Best Regards,
>> Huang, Ying
>> 
>> From 7bd903c42749c448ef6acbbdee8dcbc1c5b498b9 Mon Sep 17 00:00:00 2001
>> From: Huang Ying <ying.huang@intel.com>
>> Date: Thu, 23 Feb 2017 13:05:20 +0800
>> Subject: [PATCH -v5] mm, swap: Sort swap entries before free
>> 
>> To reduce the lock contention of swap_info_struct->lock when freeing
>> swap entry.  The freed swap entries will be collected in a per-CPU
>> buffer firstly, and be really freed later in batch.  During the batch
>> freeing, if the consecutive swap entries in the per-CPU buffer belongs
>> to same swap device, the swap_info_struct->lock needs to be
>> acquired/released only once, so that the lock contention could be
>> reduced greatly.  But if there are multiple swap devices, it is
>> possible that the lock may be unnecessarily released/acquired because
>> the swap entries belong to the same swap device are non-consecutive in
>> the per-CPU buffer.
>> 
>> To solve the issue, the per-CPU buffer is sorted according to the swap
>> device before freeing the swap entries.  Test shows that the time
>> spent by swapcache_free_entries() could be reduced after the patch.
>> 
>> With the patch, the memory (some swapped out) free time reduced
>> 13.6% (from 2.59s to 2.28s) in the vm-scalability swap-w-rand test
>> case with 16 processes.  The test is done on a Xeon E5 v3 system.  The
>> swap device used is a RAM simulated PMEM (persistent memory) device.
>> To test swapping, the test case creates 16 processes, which allocate
>> and write to the anonymous pages until the RAM and part of the swap
>> device is used up, finally the memory (some swapped out) is freed
>> before exit.
>> 
>> Signed-off-by: Huang Ying <ying.huang@intel.com>
>> Acked-by: Tim Chen <tim.c.chen@intel.com>
>> Cc: Hugh Dickins <hughd@google.com>
>> Cc: Shaohua Li <shli@kernel.org>
>> Cc: Minchan Kim <minchan@kernel.org>
>> Cc: Rik van Riel <riel@redhat.com>
>> 
>> v5:
>> 
>> - Use a smarter way to determine whether sort is necessary.
>> 
>> v4:
>> 
>> - Avoid unnecessary sort if all entries are from one swap device.
>> 
>> v3:
>> 
>> - Add some comments in code per Rik's suggestion.
>> 
>> v2:
>> 
>> - Avoid sort swap entries if there is only one swap device.
>> ---
>>  mm/swapfile.c | 43 ++++++++++++++++++++++++++++++++++++++-----
>>  1 file changed, 38 insertions(+), 5 deletions(-)
>> 
>> diff --git a/mm/swapfile.c b/mm/swapfile.c
>> index 71890061f653..10e75f9e8ac1 100644
>> --- a/mm/swapfile.c
>> +++ b/mm/swapfile.c
>> @@ -37,6 +37,7 @@
>>  #include <linux/swapfile.h>
>>  #include <linux/export.h>
>>  #include <linux/swap_slots.h>
>> +#include <linux/sort.h>
>>  
>>  #include <asm/pgtable.h>
>>  #include <asm/tlbflush.h>
>> @@ -1065,20 +1066,52 @@ void swapcache_free(swp_entry_t entry)
>>  	}
>>  }
>>  
>> +static int swp_entry_cmp(const void *ent1, const void *ent2)
>> +{
>> +	const swp_entry_t *e1 = ent1, *e2 = ent2;
>> +
>> +	return (int)(swp_type(*e1) - swp_type(*e2));
>> +}
>> +
>>  void swapcache_free_entries(swp_entry_t *entries, int n)
>>  {
>>  	struct swap_info_struct *p, *prev;
>> -	int i;
>> +	int i, m;
>> +	swp_entry_t entry;
>> +	unsigned int prev_swp_type;
>>  
>>  	if (n <= 0)
>>  		return;
>>  
>>  	prev = NULL;
>>  	p = NULL;
>> -	for (i = 0; i < n; ++i) {
>> -		p = swap_info_get_cont(entries[i], prev);
>> -		if (p)
>> -			swap_entry_free(p, entries[i]);
>> +	m = 0;
>> +	prev_swp_type = swp_type(entries[0]);
>> +	for (i = 0; i < n; i++) {
>> +		entry = entries[i];
>> +		if (likely(swp_type(entry) == prev_swp_type)) {
>> +			p = swap_info_get_cont(entry, prev);
>> +			if (likely(p))
>> +				swap_entry_free(p, entry);
>> +			prev = p;
>> +		} else if (!m)
>> +			m = i;
>> +	}
>> +	if (p)
>> +		spin_unlock(&p->lock);
>> +	if (likely(!m))
>> +		return;
>> +
>> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> +	sort(entries + m, n - m, sizeof(entries[0]), swp_entry_cmp, NULL);
>> +	prev = NULL;
>> +	for (i = m; i < n; i++) {
>> +		entry = entries[i];
>> +		if (swp_type(entry) == prev_swp_type)
>> +			continue;
>> +		p = swap_info_get_cont(entry, prev);
>> +		if (likely(p))
>> +			swap_entry_free(p, entry);
>>  		prev = p;
>>  	}
>>  	if (p)
>> -- 
>> 2.11.0
>> 

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-28  1:09                 ` Huang, Ying
@ 2017-04-28  7:42                   ` Minchan Kim
  2017-04-28  8:05                     ` Huang, Ying
  0 siblings, 1 reply; 30+ messages in thread
From: Minchan Kim @ 2017-04-28  7:42 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel

On Fri, Apr 28, 2017 at 09:09:53AM +0800, Huang, Ying wrote:
> Minchan Kim <minchan@kernel.org> writes:
> 
> > On Wed, Apr 26, 2017 at 08:42:10PM +0800, Huang, Ying wrote:
> >> Minchan Kim <minchan@kernel.org> writes:
> >> 
> >> > On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
> >> >> "Huang, Ying" <ying.huang@intel.com> writes:
> >> >> 
> >> >> > Minchan Kim <minchan@kernel.org> writes:
> >> >> >
> >> >> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
> >> >> >>> Minchan Kim <minchan@kernel.org> writes:
> >> >> >>> 
> >> >> >>> > Hi Huang,
> >> >> >>> >
> >> >> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
> >> >> >>> >> From: Huang Ying <ying.huang@intel.com>
> >> >> >>> >> 
> >> >> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
> >> >> >>> >>  {
> >> >> >>> >>  	struct swap_info_struct *p, *prev;
> >> >> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
> >> >> >>> >>  
> >> >> >>> >>  	prev = NULL;
> >> >> >>> >>  	p = NULL;
> >> >> >>> >> +
> >> >> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> >> >> >>> >> +	if (nr_swapfiles > 1)
> >> >> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> >> >> >>> >
> >> >> >>> > Let's think on other cases.
> >> >> >>> >
> >> >> >>> > There are two swaps and they are configured by priority so a swap's usage
> >> >> >>> > would be zero unless other swap used up. In case of that, this sorting
> >> >> >>> > is pointless.
> >> >> >>> >
> >> >> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
> >> >> >>> > swaps and then disable until a swap is remained, this sorting is
> >> >> >>> > pointelss, too.
> >> >> >>> >
> >> >> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
> >> >> >>> > then we can sort it.
> >> >> >>> 
> >> >> >>> Yes.  That should be better.  I just don't know whether the added
> >> >> >>> complexity is necessary, given the array is short and sort is fast.
> >> >> >>
> >> >> >> Huh?
> >> >> >>
> >> >> >> 1. swapon /dev/XXX1
> >> >> >> 2. swapon /dev/XXX2
> >> >> >> 3. swapoff /dev/XXX2
> >> >> >> 4. use only one swap
> >> >> >> 5. then, always pointless sort.
> >> >> >
> >> >> > Yes.  In this situation we will do unnecessary sorting.  What I don't
> >> >> > know is whether the unnecessary sorting will hurt performance in real
> >> >> > life.  I can do some measurement.
> >> >> 
> >> >> I tested the patch with 1 swap device and 1 process to eat memory
> >> >> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
> >> >> worse case because there is no lock contention.  The memory freeing time
> >> >> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
> >> >> overhead for some cases.  I change the algorithm to something like
> >> >> below,
> >> >> 
> >> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
> >> >>  {
> >> >>  	struct swap_info_struct *p, *prev;
> >> >>  	int i;
> >> >> +	swp_entry_t entry;
> >> >> +	unsigned int prev_swp_type;
> >> >>  
> >> >>  	if (n <= 0)
> >> >>  		return;
> >> >>  
> >> >> +	prev_swp_type = swp_type(entries[0]);
> >> >> +	for (i = n - 1; i > 0; i--) {
> >> >> +		if (swp_type(entries[i]) != prev_swp_type)
> >> >> +			break;
> >> >> +	}
> >> >
> >> > That's really what I want to avoid. For many swap usecases,
> >> > it adds unnecessary overhead.
> >> >
> >> >> +
> >> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> >> >> +	if (i)
> >> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> >> >>  	prev = NULL;
> >> >>  	p = NULL;
> >> >>  	for (i = 0; i < n; ++i) {
> >> >> -		p = swap_info_get_cont(entries[i], prev);
> >> >> +		entry = entries[i];
> >> >> +		p = swap_info_get_cont(entry, prev);
> >> >>  		if (p)
> >> >> -			swap_entry_free(p, entries[i]);
> >> >> +			swap_entry_free(p, entry);
> >> >>  		prev = p;
> >> >>  	}
> >> >>  	if (p)
> >> >> 
> >> >> With this patch, the memory freeing time increased from 1.94s to 1.97s.
> >> >> I think this is good enough.  Do you think so?
> >> >
> >> > What I mean is as follows(I didn't test it at all):
> >> >
> >> > With this, sort entries if we found multiple entries in current
> >> > entries. It adds some condition checks for non-multiple swap
> >> > usecase but it would be more cheaper than the sorting.
> >> > And it adds a [un]lock overhead for multiple swap usecase but
> >> > it should be a compromise for single-swap usecase which is more
> >> > popular.
> >> >
> >> 
> >> How about the following solution?  It can avoid [un]lock overhead and
> >> double lock issue for multiple swap user case and has good performance
> >> for one swap user case too.
> >
> > How worse with approach I suggested compared to as-is?
> 
> The performance difference between your version and my version is small
> for my testing.

If so, why should we add code to optimize further?

> 
> > Unless it's too bad, let's not add more complicated thing to just
> > enhance the minor usecase in such even *slow* path.
> > It adds code size/maintainance overead.
> > With your suggestion, it might enhance a bit with speicific benchmark
> > but not sure it's really worth for real practice.
> 
> I don't think the code complexity has much difference between our latest
> versions.  As for complexity, I think my original version which just

What I suggested is to avoid pointless overhead for *major* usecase
and the code you are adding now is to optimize further for *minor*
usecase. And now I dobut the code you are adding is really worth
unless it makes a meaningful output.
If it doesn't, it adds just overhead(code size, maintainance, power and
performance). You might argue it's really *small* so it would be okay
but think about that you would be not only one in the community so
kernel bloats day by day with code to handle corner cases.

> uses nr_swapfiles to avoid sort() for single swap device is simple and
> good enough for this task.  Maybe we can just improve the correctness of

But it hurts *major* usecase.

> swap device counting as Tim suggested.

I don't know what Tim suggested. Anyway, my point is that minor
usecase doesn't hurt major usecase and justify the benefit
if you want to put more. So I'm okay with either solution to
meet it.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-28  7:42                   ` Minchan Kim
@ 2017-04-28  8:05                     ` Huang, Ying
  2017-04-28  9:00                       ` Minchan Kim
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-04-28  8:05 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Minchan Kim <minchan@kernel.org> writes:

> On Fri, Apr 28, 2017 at 09:09:53AM +0800, Huang, Ying wrote:
>> Minchan Kim <minchan@kernel.org> writes:
>> 
>> > On Wed, Apr 26, 2017 at 08:42:10PM +0800, Huang, Ying wrote:
>> >> Minchan Kim <minchan@kernel.org> writes:
>> >> 
>> >> > On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
>> >> >> "Huang, Ying" <ying.huang@intel.com> writes:
>> >> >> 
>> >> >> > Minchan Kim <minchan@kernel.org> writes:
>> >> >> >
>> >> >> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> >> >> >>> Minchan Kim <minchan@kernel.org> writes:
>> >> >> >>> 
>> >> >> >>> > Hi Huang,
>> >> >> >>> >
>> >> >> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> >> >> >>> >> From: Huang Ying <ying.huang@intel.com>
>> >> >> >>> >> 
>> >> >> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >> >> >>> >>  {
>> >> >> >>> >>  	struct swap_info_struct *p, *prev;
>> >> >> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> >> >> >>> >>  
>> >> >> >>> >>  	prev = NULL;
>> >> >> >>> >>  	p = NULL;
>> >> >> >>> >> +
>> >> >> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> >> >>> >> +	if (nr_swapfiles > 1)
>> >> >> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >> >> >>> >
>> >> >> >>> > Let's think on other cases.
>> >> >> >>> >
>> >> >> >>> > There are two swaps and they are configured by priority so a swap's usage
>> >> >> >>> > would be zero unless other swap used up. In case of that, this sorting
>> >> >> >>> > is pointless.
>> >> >> >>> >
>> >> >> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
>> >> >> >>> > swaps and then disable until a swap is remained, this sorting is
>> >> >> >>> > pointelss, too.
>> >> >> >>> >
>> >> >> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
>> >> >> >>> > then we can sort it.
>> >> >> >>> 
>> >> >> >>> Yes.  That should be better.  I just don't know whether the added
>> >> >> >>> complexity is necessary, given the array is short and sort is fast.
>> >> >> >>
>> >> >> >> Huh?
>> >> >> >>
>> >> >> >> 1. swapon /dev/XXX1
>> >> >> >> 2. swapon /dev/XXX2
>> >> >> >> 3. swapoff /dev/XXX2
>> >> >> >> 4. use only one swap
>> >> >> >> 5. then, always pointless sort.
>> >> >> >
>> >> >> > Yes.  In this situation we will do unnecessary sorting.  What I don't
>> >> >> > know is whether the unnecessary sorting will hurt performance in real
>> >> >> > life.  I can do some measurement.
>> >> >> 
>> >> >> I tested the patch with 1 swap device and 1 process to eat memory
>> >> >> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
>> >> >> worse case because there is no lock contention.  The memory freeing time
>> >> >> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
>> >> >> overhead for some cases.  I change the algorithm to something like
>> >> >> below,
>> >> >> 
>> >> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >> >>  {
>> >> >>  	struct swap_info_struct *p, *prev;
>> >> >>  	int i;
>> >> >> +	swp_entry_t entry;
>> >> >> +	unsigned int prev_swp_type;
>> >> >>  
>> >> >>  	if (n <= 0)
>> >> >>  		return;
>> >> >>  
>> >> >> +	prev_swp_type = swp_type(entries[0]);
>> >> >> +	for (i = n - 1; i > 0; i--) {
>> >> >> +		if (swp_type(entries[i]) != prev_swp_type)
>> >> >> +			break;
>> >> >> +	}
>> >> >
>> >> > That's really what I want to avoid. For many swap usecases,
>> >> > it adds unnecessary overhead.
>> >> >
>> >> >> +
>> >> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> >> +	if (i)
>> >> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >> >>  	prev = NULL;
>> >> >>  	p = NULL;
>> >> >>  	for (i = 0; i < n; ++i) {
>> >> >> -		p = swap_info_get_cont(entries[i], prev);
>> >> >> +		entry = entries[i];
>> >> >> +		p = swap_info_get_cont(entry, prev);
>> >> >>  		if (p)
>> >> >> -			swap_entry_free(p, entries[i]);
>> >> >> +			swap_entry_free(p, entry);
>> >> >>  		prev = p;
>> >> >>  	}
>> >> >>  	if (p)
>> >> >> 
>> >> >> With this patch, the memory freeing time increased from 1.94s to 1.97s.
>> >> >> I think this is good enough.  Do you think so?
>> >> >
>> >> > What I mean is as follows(I didn't test it at all):
>> >> >
>> >> > With this, sort entries if we found multiple entries in current
>> >> > entries. It adds some condition checks for non-multiple swap
>> >> > usecase but it would be more cheaper than the sorting.
>> >> > And it adds a [un]lock overhead for multiple swap usecase but
>> >> > it should be a compromise for single-swap usecase which is more
>> >> > popular.
>> >> >
>> >> 
>> >> How about the following solution?  It can avoid [un]lock overhead and
>> >> double lock issue for multiple swap user case and has good performance
>> >> for one swap user case too.
>> >
>> > How worse with approach I suggested compared to as-is?
>> 
>> The performance difference between your version and my version is small
>> for my testing.
>
> If so, why should we add code to optimize further?
>
>> 
>> > Unless it's too bad, let's not add more complicated thing to just
>> > enhance the minor usecase in such even *slow* path.
>> > It adds code size/maintainance overead.
>> > With your suggestion, it might enhance a bit with speicific benchmark
>> > but not sure it's really worth for real practice.
>> 
>> I don't think the code complexity has much difference between our latest
>> versions.  As for complexity, I think my original version which just
>
> What I suggested is to avoid pointless overhead for *major* usecase
> and the code you are adding now is to optimize further for *minor*
> usecase. And now I dobut the code you are adding is really worth
> unless it makes a meaningful output.
> If it doesn't, it adds just overhead(code size, maintainance, power and
> performance). You might argue it's really *small* so it would be okay
> but think about that you would be not only one in the community so
> kernel bloats day by day with code to handle corner cases.
>
>> uses nr_swapfiles to avoid sort() for single swap device is simple and
>> good enough for this task.  Maybe we can just improve the correctness of
>
> But it hurts *major* usecase.
>
>> swap device counting as Tim suggested.
>
> I don't know what Tim suggested. Anyway, my point is that minor
> usecase doesn't hurt major usecase and justify the benefit
> if you want to put more. So I'm okay with either solution to
> meet it.

Tim suggested to add a mechanism to correctly track how many swap
devices are in use in swapon/swapoff.  So we only sort if the number of
the swap device > 1.  This will not cover multiple swap devices with
different priorities, but will cover the major usecases.  The code
should be simpler.

Best Regards,
Huang, Ying

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-28  8:05                     ` Huang, Ying
@ 2017-04-28  9:00                       ` Minchan Kim
  2017-04-28 11:48                         ` Huang, Ying
  0 siblings, 1 reply; 30+ messages in thread
From: Minchan Kim @ 2017-04-28  9:00 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel

On Fri, Apr 28, 2017 at 04:05:26PM +0800, Huang, Ying wrote:
> Minchan Kim <minchan@kernel.org> writes:
> 
> > On Fri, Apr 28, 2017 at 09:09:53AM +0800, Huang, Ying wrote:
> >> Minchan Kim <minchan@kernel.org> writes:
> >> 
> >> > On Wed, Apr 26, 2017 at 08:42:10PM +0800, Huang, Ying wrote:
> >> >> Minchan Kim <minchan@kernel.org> writes:
> >> >> 
> >> >> > On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
> >> >> >> "Huang, Ying" <ying.huang@intel.com> writes:
> >> >> >> 
> >> >> >> > Minchan Kim <minchan@kernel.org> writes:
> >> >> >> >
> >> >> >> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
> >> >> >> >>> Minchan Kim <minchan@kernel.org> writes:
> >> >> >> >>> 
> >> >> >> >>> > Hi Huang,
> >> >> >> >>> >
> >> >> >> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
> >> >> >> >>> >> From: Huang Ying <ying.huang@intel.com>
> >> >> >> >>> >> 
> >> >> >> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
> >> >> >> >>> >>  {
> >> >> >> >>> >>  	struct swap_info_struct *p, *prev;
> >> >> >> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
> >> >> >> >>> >>  
> >> >> >> >>> >>  	prev = NULL;
> >> >> >> >>> >>  	p = NULL;
> >> >> >> >>> >> +
> >> >> >> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> >> >> >> >>> >> +	if (nr_swapfiles > 1)
> >> >> >> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> >> >> >> >>> >
> >> >> >> >>> > Let's think on other cases.
> >> >> >> >>> >
> >> >> >> >>> > There are two swaps and they are configured by priority so a swap's usage
> >> >> >> >>> > would be zero unless other swap used up. In case of that, this sorting
> >> >> >> >>> > is pointless.
> >> >> >> >>> >
> >> >> >> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
> >> >> >> >>> > swaps and then disable until a swap is remained, this sorting is
> >> >> >> >>> > pointelss, too.
> >> >> >> >>> >
> >> >> >> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
> >> >> >> >>> > then we can sort it.
> >> >> >> >>> 
> >> >> >> >>> Yes.  That should be better.  I just don't know whether the added
> >> >> >> >>> complexity is necessary, given the array is short and sort is fast.
> >> >> >> >>
> >> >> >> >> Huh?
> >> >> >> >>
> >> >> >> >> 1. swapon /dev/XXX1
> >> >> >> >> 2. swapon /dev/XXX2
> >> >> >> >> 3. swapoff /dev/XXX2
> >> >> >> >> 4. use only one swap
> >> >> >> >> 5. then, always pointless sort.
> >> >> >> >
> >> >> >> > Yes.  In this situation we will do unnecessary sorting.  What I don't
> >> >> >> > know is whether the unnecessary sorting will hurt performance in real
> >> >> >> > life.  I can do some measurement.
> >> >> >> 
> >> >> >> I tested the patch with 1 swap device and 1 process to eat memory
> >> >> >> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
> >> >> >> worse case because there is no lock contention.  The memory freeing time
> >> >> >> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
> >> >> >> overhead for some cases.  I change the algorithm to something like
> >> >> >> below,
> >> >> >> 
> >> >> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
> >> >> >>  {
> >> >> >>  	struct swap_info_struct *p, *prev;
> >> >> >>  	int i;
> >> >> >> +	swp_entry_t entry;
> >> >> >> +	unsigned int prev_swp_type;
> >> >> >>  
> >> >> >>  	if (n <= 0)
> >> >> >>  		return;
> >> >> >>  
> >> >> >> +	prev_swp_type = swp_type(entries[0]);
> >> >> >> +	for (i = n - 1; i > 0; i--) {
> >> >> >> +		if (swp_type(entries[i]) != prev_swp_type)
> >> >> >> +			break;
> >> >> >> +	}
> >> >> >
> >> >> > That's really what I want to avoid. For many swap usecases,
> >> >> > it adds unnecessary overhead.
> >> >> >
> >> >> >> +
> >> >> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
> >> >> >> +	if (i)
> >> >> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
> >> >> >>  	prev = NULL;
> >> >> >>  	p = NULL;
> >> >> >>  	for (i = 0; i < n; ++i) {
> >> >> >> -		p = swap_info_get_cont(entries[i], prev);
> >> >> >> +		entry = entries[i];
> >> >> >> +		p = swap_info_get_cont(entry, prev);
> >> >> >>  		if (p)
> >> >> >> -			swap_entry_free(p, entries[i]);
> >> >> >> +			swap_entry_free(p, entry);
> >> >> >>  		prev = p;
> >> >> >>  	}
> >> >> >>  	if (p)
> >> >> >> 
> >> >> >> With this patch, the memory freeing time increased from 1.94s to 1.97s.
> >> >> >> I think this is good enough.  Do you think so?
> >> >> >
> >> >> > What I mean is as follows(I didn't test it at all):
> >> >> >
> >> >> > With this, sort entries if we found multiple entries in current
> >> >> > entries. It adds some condition checks for non-multiple swap
> >> >> > usecase but it would be more cheaper than the sorting.
> >> >> > And it adds a [un]lock overhead for multiple swap usecase but
> >> >> > it should be a compromise for single-swap usecase which is more
> >> >> > popular.
> >> >> >
> >> >> 
> >> >> How about the following solution?  It can avoid [un]lock overhead and
> >> >> double lock issue for multiple swap user case and has good performance
> >> >> for one swap user case too.
> >> >
> >> > How worse with approach I suggested compared to as-is?
> >> 
> >> The performance difference between your version and my version is small
> >> for my testing.
> >
> > If so, why should we add code to optimize further?
> >
> >> 
> >> > Unless it's too bad, let's not add more complicated thing to just
> >> > enhance the minor usecase in such even *slow* path.
> >> > It adds code size/maintainance overead.
> >> > With your suggestion, it might enhance a bit with speicific benchmark
> >> > but not sure it's really worth for real practice.
> >> 
> >> I don't think the code complexity has much difference between our latest
> >> versions.  As for complexity, I think my original version which just
> >
> > What I suggested is to avoid pointless overhead for *major* usecase
> > and the code you are adding now is to optimize further for *minor*
> > usecase. And now I dobut the code you are adding is really worth
> > unless it makes a meaningful output.
> > If it doesn't, it adds just overhead(code size, maintainance, power and
> > performance). You might argue it's really *small* so it would be okay
> > but think about that you would be not only one in the community so
> > kernel bloats day by day with code to handle corner cases.
> >
> >> uses nr_swapfiles to avoid sort() for single swap device is simple and
> >> good enough for this task.  Maybe we can just improve the correctness of
> >
> > But it hurts *major* usecase.
> >
> >> swap device counting as Tim suggested.
> >
> > I don't know what Tim suggested. Anyway, my point is that minor
> > usecase doesn't hurt major usecase and justify the benefit
> > if you want to put more. So I'm okay with either solution to
> > meet it.
> 
> Tim suggested to add a mechanism to correctly track how many swap
> devices are in use in swapon/swapoff.  So we only sort if the number of
> the swap device > 1.  This will not cover multiple swap devices with
> different priorities, but will cover the major usecases.  The code
> should be simpler.

As you know, it doesn't solve multiple swaps by priority.
Even, there are cases full with entries same swap device
although multiple swap devices are used.

So, I think runtime sorting by judging need to be sored is still
better.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-28  9:00                       ` Minchan Kim
@ 2017-04-28 11:48                         ` Huang, Ying
  2017-04-28 13:35                           ` Huang, Ying
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-04-28 11:48 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Minchan Kim <minchan@kernel.org> writes:

> On Fri, Apr 28, 2017 at 04:05:26PM +0800, Huang, Ying wrote:
>> Minchan Kim <minchan@kernel.org> writes:
>> 
>> > On Fri, Apr 28, 2017 at 09:09:53AM +0800, Huang, Ying wrote:
>> >> Minchan Kim <minchan@kernel.org> writes:
>> >> 
>> >> > On Wed, Apr 26, 2017 at 08:42:10PM +0800, Huang, Ying wrote:
>> >> >> Minchan Kim <minchan@kernel.org> writes:
>> >> >> 
>> >> >> > On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
>> >> >> >> "Huang, Ying" <ying.huang@intel.com> writes:
>> >> >> >> 
>> >> >> >> > Minchan Kim <minchan@kernel.org> writes:
>> >> >> >> >
>> >> >> >> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>> >> >> >> >>> Minchan Kim <minchan@kernel.org> writes:
>> >> >> >> >>> 
>> >> >> >> >>> > Hi Huang,
>> >> >> >> >>> >
>> >> >> >> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>> >> >> >> >>> >> From: Huang Ying <ying.huang@intel.com>
>> >> >> >> >>> >> 
>> >> >> >> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >> >> >> >>> >>  {
>> >> >> >> >>> >>  	struct swap_info_struct *p, *prev;
>> >> >> >> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>> >> >> >> >>> >>  
>> >> >> >> >>> >>  	prev = NULL;
>> >> >> >> >>> >>  	p = NULL;
>> >> >> >> >>> >> +
>> >> >> >> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> >> >> >>> >> +	if (nr_swapfiles > 1)
>> >> >> >> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >> >> >> >>> >
>> >> >> >> >>> > Let's think on other cases.
>> >> >> >> >>> >
>> >> >> >> >>> > There are two swaps and they are configured by priority so a swap's usage
>> >> >> >> >>> > would be zero unless other swap used up. In case of that, this sorting
>> >> >> >> >>> > is pointless.
>> >> >> >> >>> >
>> >> >> >> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
>> >> >> >> >>> > swaps and then disable until a swap is remained, this sorting is
>> >> >> >> >>> > pointelss, too.
>> >> >> >> >>> >
>> >> >> >> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
>> >> >> >> >>> > then we can sort it.
>> >> >> >> >>> 
>> >> >> >> >>> Yes.  That should be better.  I just don't know whether the added
>> >> >> >> >>> complexity is necessary, given the array is short and sort is fast.
>> >> >> >> >>
>> >> >> >> >> Huh?
>> >> >> >> >>
>> >> >> >> >> 1. swapon /dev/XXX1
>> >> >> >> >> 2. swapon /dev/XXX2
>> >> >> >> >> 3. swapoff /dev/XXX2
>> >> >> >> >> 4. use only one swap
>> >> >> >> >> 5. then, always pointless sort.
>> >> >> >> >
>> >> >> >> > Yes.  In this situation we will do unnecessary sorting.  What I don't
>> >> >> >> > know is whether the unnecessary sorting will hurt performance in real
>> >> >> >> > life.  I can do some measurement.
>> >> >> >> 
>> >> >> >> I tested the patch with 1 swap device and 1 process to eat memory
>> >> >> >> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
>> >> >> >> worse case because there is no lock contention.  The memory freeing time
>> >> >> >> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
>> >> >> >> overhead for some cases.  I change the algorithm to something like
>> >> >> >> below,
>> >> >> >> 
>> >> >> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>> >> >> >>  {
>> >> >> >>  	struct swap_info_struct *p, *prev;
>> >> >> >>  	int i;
>> >> >> >> +	swp_entry_t entry;
>> >> >> >> +	unsigned int prev_swp_type;
>> >> >> >>  
>> >> >> >>  	if (n <= 0)
>> >> >> >>  		return;
>> >> >> >>  
>> >> >> >> +	prev_swp_type = swp_type(entries[0]);
>> >> >> >> +	for (i = n - 1; i > 0; i--) {
>> >> >> >> +		if (swp_type(entries[i]) != prev_swp_type)
>> >> >> >> +			break;
>> >> >> >> +	}
>> >> >> >
>> >> >> > That's really what I want to avoid. For many swap usecases,
>> >> >> > it adds unnecessary overhead.
>> >> >> >
>> >> >> >> +
>> >> >> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>> >> >> >> +	if (i)
>> >> >> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>> >> >> >>  	prev = NULL;
>> >> >> >>  	p = NULL;
>> >> >> >>  	for (i = 0; i < n; ++i) {
>> >> >> >> -		p = swap_info_get_cont(entries[i], prev);
>> >> >> >> +		entry = entries[i];
>> >> >> >> +		p = swap_info_get_cont(entry, prev);
>> >> >> >>  		if (p)
>> >> >> >> -			swap_entry_free(p, entries[i]);
>> >> >> >> +			swap_entry_free(p, entry);
>> >> >> >>  		prev = p;
>> >> >> >>  	}
>> >> >> >>  	if (p)
>> >> >> >> 
>> >> >> >> With this patch, the memory freeing time increased from 1.94s to 1.97s.
>> >> >> >> I think this is good enough.  Do you think so?
>> >> >> >
>> >> >> > What I mean is as follows(I didn't test it at all):
>> >> >> >
>> >> >> > With this, sort entries if we found multiple entries in current
>> >> >> > entries. It adds some condition checks for non-multiple swap
>> >> >> > usecase but it would be more cheaper than the sorting.
>> >> >> > And it adds a [un]lock overhead for multiple swap usecase but
>> >> >> > it should be a compromise for single-swap usecase which is more
>> >> >> > popular.
>> >> >> >
>> >> >> 
>> >> >> How about the following solution?  It can avoid [un]lock overhead and
>> >> >> double lock issue for multiple swap user case and has good performance
>> >> >> for one swap user case too.
>> >> >
>> >> > How worse with approach I suggested compared to as-is?
>> >> 
>> >> The performance difference between your version and my version is small
>> >> for my testing.
>> >
>> > If so, why should we add code to optimize further?
>> >
>> >> 
>> >> > Unless it's too bad, let's not add more complicated thing to just
>> >> > enhance the minor usecase in such even *slow* path.
>> >> > It adds code size/maintainance overead.
>> >> > With your suggestion, it might enhance a bit with speicific benchmark
>> >> > but not sure it's really worth for real practice.
>> >> 
>> >> I don't think the code complexity has much difference between our latest
>> >> versions.  As for complexity, I think my original version which just
>> >
>> > What I suggested is to avoid pointless overhead for *major* usecase
>> > and the code you are adding now is to optimize further for *minor*
>> > usecase. And now I dobut the code you are adding is really worth
>> > unless it makes a meaningful output.
>> > If it doesn't, it adds just overhead(code size, maintainance, power and
>> > performance). You might argue it's really *small* so it would be okay
>> > but think about that you would be not only one in the community so
>> > kernel bloats day by day with code to handle corner cases.
>> >
>> >> uses nr_swapfiles to avoid sort() for single swap device is simple and
>> >> good enough for this task.  Maybe we can just improve the correctness of
>> >
>> > But it hurts *major* usecase.
>> >
>> >> swap device counting as Tim suggested.
>> >
>> > I don't know what Tim suggested. Anyway, my point is that minor
>> > usecase doesn't hurt major usecase and justify the benefit
>> > if you want to put more. So I'm okay with either solution to
>> > meet it.
>> 
>> Tim suggested to add a mechanism to correctly track how many swap
>> devices are in use in swapon/swapoff.  So we only sort if the number of
>> the swap device > 1.  This will not cover multiple swap devices with
>> different priorities, but will cover the major usecases.  The code
>> should be simpler.
>
> As you know, it doesn't solve multiple swaps by priority.

I don't think this is *major* usecase.

> Even, there are cases full with entries same swap device
> although multiple swap devices are used.

Why, if you have multiple swap device, every time you will allocate from
different swap device.  Although there are swap alloc slots cache, the
possibility of full alignment is low.

Even if there are cases all entries come from one swap device, the
sorting is fast in fact because the array is short and the elements are
sorted (same swap type) already.  So it is not necessary to worry about
that too much.

Best Regards,
Huang, Ying

> So, I think runtime sorting by judging need to be sored is still
> better.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-28 11:48                         ` Huang, Ying
@ 2017-04-28 13:35                           ` Huang, Ying
  2017-05-02  5:02                             ` Minchan Kim
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-04-28 13:35 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Minchan Kim, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

"Huang, Ying" <ying.huang@intel.com> writes:

> Minchan Kim <minchan@kernel.org> writes:
>
>> On Fri, Apr 28, 2017 at 04:05:26PM +0800, Huang, Ying wrote:
>>> Minchan Kim <minchan@kernel.org> writes:
>>> 
>>> > On Fri, Apr 28, 2017 at 09:09:53AM +0800, Huang, Ying wrote:
>>> >> Minchan Kim <minchan@kernel.org> writes:
>>> >> 
>>> >> > On Wed, Apr 26, 2017 at 08:42:10PM +0800, Huang, Ying wrote:
>>> >> >> Minchan Kim <minchan@kernel.org> writes:
>>> >> >> 
>>> >> >> > On Fri, Apr 21, 2017 at 08:29:30PM +0800, Huang, Ying wrote:
>>> >> >> >> "Huang, Ying" <ying.huang@intel.com> writes:
>>> >> >> >> 
>>> >> >> >> > Minchan Kim <minchan@kernel.org> writes:
>>> >> >> >> >
>>> >> >> >> >> On Wed, Apr 19, 2017 at 04:14:43PM +0800, Huang, Ying wrote:
>>> >> >> >> >>> Minchan Kim <minchan@kernel.org> writes:
>>> >> >> >> >>> 
>>> >> >> >> >>> > Hi Huang,
>>> >> >> >> >>> >
>>> >> >> >> >>> > On Fri, Apr 07, 2017 at 02:49:01PM +0800, Huang, Ying wrote:
>>> >> >> >> >>> >> From: Huang Ying <ying.huang@intel.com>
>>> >> >> >> >>> >> 
>>> >> >> >> >>> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>>> >> >> >> >>> >>  {
>>> >> >> >> >>> >>  	struct swap_info_struct *p, *prev;
>>> >> >> >> >>> >> @@ -1075,6 +1083,10 @@ void swapcache_free_entries(swp_entry_t *entries, int n)
>>> >> >> >> >>> >>  
>>> >> >> >> >>> >>  	prev = NULL;
>>> >> >> >> >>> >>  	p = NULL;
>>> >> >> >> >>> >> +
>>> >> >> >> >>> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>>> >> >> >> >>> >> +	if (nr_swapfiles > 1)
>>> >> >> >> >>> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>>> >> >> >> >>> >
>>> >> >> >> >>> > Let's think on other cases.
>>> >> >> >> >>> >
>>> >> >> >> >>> > There are two swaps and they are configured by priority so a swap's usage
>>> >> >> >> >>> > would be zero unless other swap used up. In case of that, this sorting
>>> >> >> >> >>> > is pointless.
>>> >> >> >> >>> >
>>> >> >> >> >>> > As well, nr_swapfiles is never decreased so if we enable multiple
>>> >> >> >> >>> > swaps and then disable until a swap is remained, this sorting is
>>> >> >> >> >>> > pointelss, too.
>>> >> >> >> >>> >
>>> >> >> >> >>> > How about lazy sorting approach? IOW, if we found prev != p and,
>>> >> >> >> >>> > then we can sort it.
>>> >> >> >> >>> 
>>> >> >> >> >>> Yes.  That should be better.  I just don't know whether the added
>>> >> >> >> >>> complexity is necessary, given the array is short and sort is fast.
>>> >> >> >> >>
>>> >> >> >> >> Huh?
>>> >> >> >> >>
>>> >> >> >> >> 1. swapon /dev/XXX1
>>> >> >> >> >> 2. swapon /dev/XXX2
>>> >> >> >> >> 3. swapoff /dev/XXX2
>>> >> >> >> >> 4. use only one swap
>>> >> >> >> >> 5. then, always pointless sort.
>>> >> >> >> >
>>> >> >> >> > Yes.  In this situation we will do unnecessary sorting.  What I don't
>>> >> >> >> > know is whether the unnecessary sorting will hurt performance in real
>>> >> >> >> > life.  I can do some measurement.
>>> >> >> >> 
>>> >> >> >> I tested the patch with 1 swap device and 1 process to eat memory
>>> >> >> >> (remove the "if (nr_swapfiles > 1)" for test).  I think this is the
>>> >> >> >> worse case because there is no lock contention.  The memory freeing time
>>> >> >> >> increased from 1.94s to 2.12s (increase ~9.2%).  So there is some
>>> >> >> >> overhead for some cases.  I change the algorithm to something like
>>> >> >> >> below,
>>> >> >> >> 
>>> >> >> >>  void swapcache_free_entries(swp_entry_t *entries, int n)
>>> >> >> >>  {
>>> >> >> >>  	struct swap_info_struct *p, *prev;
>>> >> >> >>  	int i;
>>> >> >> >> +	swp_entry_t entry;
>>> >> >> >> +	unsigned int prev_swp_type;
>>> >> >> >>  
>>> >> >> >>  	if (n <= 0)
>>> >> >> >>  		return;
>>> >> >> >>  
>>> >> >> >> +	prev_swp_type = swp_type(entries[0]);
>>> >> >> >> +	for (i = n - 1; i > 0; i--) {
>>> >> >> >> +		if (swp_type(entries[i]) != prev_swp_type)
>>> >> >> >> +			break;
>>> >> >> >> +	}
>>> >> >> >
>>> >> >> > That's really what I want to avoid. For many swap usecases,
>>> >> >> > it adds unnecessary overhead.
>>> >> >> >
>>> >> >> >> +
>>> >> >> >> +	/* Sort swap entries by swap device, so each lock is only taken once. */
>>> >> >> >> +	if (i)
>>> >> >> >> +		sort(entries, n, sizeof(entries[0]), swp_entry_cmp, NULL);
>>> >> >> >>  	prev = NULL;
>>> >> >> >>  	p = NULL;
>>> >> >> >>  	for (i = 0; i < n; ++i) {
>>> >> >> >> -		p = swap_info_get_cont(entries[i], prev);
>>> >> >> >> +		entry = entries[i];
>>> >> >> >> +		p = swap_info_get_cont(entry, prev);
>>> >> >> >>  		if (p)
>>> >> >> >> -			swap_entry_free(p, entries[i]);
>>> >> >> >> +			swap_entry_free(p, entry);
>>> >> >> >>  		prev = p;
>>> >> >> >>  	}
>>> >> >> >>  	if (p)
>>> >> >> >> 
>>> >> >> >> With this patch, the memory freeing time increased from 1.94s to 1.97s.
>>> >> >> >> I think this is good enough.  Do you think so?
>>> >> >> >
>>> >> >> > What I mean is as follows(I didn't test it at all):
>>> >> >> >
>>> >> >> > With this, sort entries if we found multiple entries in current
>>> >> >> > entries. It adds some condition checks for non-multiple swap
>>> >> >> > usecase but it would be more cheaper than the sorting.
>>> >> >> > And it adds a [un]lock overhead for multiple swap usecase but
>>> >> >> > it should be a compromise for single-swap usecase which is more
>>> >> >> > popular.
>>> >> >> >
>>> >> >> 
>>> >> >> How about the following solution?  It can avoid [un]lock overhead and
>>> >> >> double lock issue for multiple swap user case and has good performance
>>> >> >> for one swap user case too.
>>> >> >
>>> >> > How worse with approach I suggested compared to as-is?
>>> >> 
>>> >> The performance difference between your version and my version is small
>>> >> for my testing.
>>> >
>>> > If so, why should we add code to optimize further?
>>> >
>>> >> 
>>> >> > Unless it's too bad, let's not add more complicated thing to just
>>> >> > enhance the minor usecase in such even *slow* path.
>>> >> > It adds code size/maintainance overead.
>>> >> > With your suggestion, it might enhance a bit with speicific benchmark
>>> >> > but not sure it's really worth for real practice.
>>> >> 
>>> >> I don't think the code complexity has much difference between our latest
>>> >> versions.  As for complexity, I think my original version which just
>>> >
>>> > What I suggested is to avoid pointless overhead for *major* usecase
>>> > and the code you are adding now is to optimize further for *minor*
>>> > usecase. And now I dobut the code you are adding is really worth
>>> > unless it makes a meaningful output.
>>> > If it doesn't, it adds just overhead(code size, maintainance, power and
>>> > performance). You might argue it's really *small* so it would be okay
>>> > but think about that you would be not only one in the community so
>>> > kernel bloats day by day with code to handle corner cases.
>>> >
>>> >> uses nr_swapfiles to avoid sort() for single swap device is simple and
>>> >> good enough for this task.  Maybe we can just improve the correctness of
>>> >
>>> > But it hurts *major* usecase.
>>> >
>>> >> swap device counting as Tim suggested.
>>> >
>>> > I don't know what Tim suggested. Anyway, my point is that minor
>>> > usecase doesn't hurt major usecase and justify the benefit
>>> > if you want to put more. So I'm okay with either solution to
>>> > meet it.
>>> 
>>> Tim suggested to add a mechanism to correctly track how many swap
>>> devices are in use in swapon/swapoff.  So we only sort if the number of
>>> the swap device > 1.  This will not cover multiple swap devices with
>>> different priorities, but will cover the major usecases.  The code
>>> should be simpler.
>>
>> As you know, it doesn't solve multiple swaps by priority.
>
> I don't think this is *major* usecase.
>
>> Even, there are cases full with entries same swap device
>> although multiple swap devices are used.
>
> Why, if you have multiple swap device, every time you will allocate from
> different swap device.  Although there are swap alloc slots cache, the
> possibility of full alignment is low.
>
> Even if there are cases all entries come from one swap device, the
> sorting is fast in fact because the array is short and the elements are
> sorted (same swap type) already.  So it is not necessary to worry about
> that too much.

In fact, during the test, I found the overhead of sort() is comparable
with the performance difference of adding likely()/unlikely() to the
"if" in the function.

Best Regards,
Huang, Ying

> Best Regards,
> Huang, Ying
>
>> So, I think runtime sorting by judging need to be sored is still
>> better.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-04-28 13:35                           ` Huang, Ying
@ 2017-05-02  5:02                             ` Minchan Kim
  2017-05-02  5:35                               ` Huang, Ying
  0 siblings, 1 reply; 30+ messages in thread
From: Minchan Kim @ 2017-05-02  5:02 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel

On Fri, Apr 28, 2017 at 09:35:37PM +0800, Huang, Ying wrote:
> In fact, during the test, I found the overhead of sort() is comparable
> with the performance difference of adding likely()/unlikely() to the
> "if" in the function.

Huang,

This discussion is started from your optimization code:

        if (nr_swapfiles > 1)
                sort();

I don't have such fast machine so cannot test it. However, you added
such optimization code in there so I guess it's *worth* to review so
with spending my time, I pointed out what you are missing and
suggested a idea to find a compromise.

Now you are saying sort is so fast so no worth to add more logics
to avoid the overhead?
Then, please just drop that if condition part and instead, sort
it unconditionally.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-05-02  5:02                             ` Minchan Kim
@ 2017-05-02  5:35                               ` Huang, Ying
  2017-05-02  5:48                                 ` Minchan Kim
  0 siblings, 1 reply; 30+ messages in thread
From: Huang, Ying @ 2017-05-02  5:35 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Hi, Minchan,

Minchan Kim <minchan@kernel.org> writes:

> On Fri, Apr 28, 2017 at 09:35:37PM +0800, Huang, Ying wrote:
>> In fact, during the test, I found the overhead of sort() is comparable
>> with the performance difference of adding likely()/unlikely() to the
>> "if" in the function.
>
> Huang,
>
> This discussion is started from your optimization code:
>
>         if (nr_swapfiles > 1)
>                 sort();
>
> I don't have such fast machine so cannot test it. However, you added
> such optimization code in there so I guess it's *worth* to review so
> with spending my time, I pointed out what you are missing and
> suggested a idea to find a compromise.

Sorry for wasting your time and Thanks a lot for your review and
suggestion!

When I started talking this with you, I found there is some measurable
overhead of sort().  But later when I done more tests, I found the
measurable overhead is at the same level of likely()/unlikely() compiler
notation.  So you help me to find that, Thanks again!

> Now you are saying sort is so fast so no worth to add more logics
> to avoid the overhead?
> Then, please just drop that if condition part and instead, sort
> it unconditionally.

Now, because we found the overhead of sort() is low, I suggest to put
minimal effort to avoid it.  Like the original implementation,

         if (nr_swapfiles > 1)
                 sort();

Or, we can make nr_swapfiles more correct as Tim suggested (tracking
the number of the swap devices during swap on/off).

Best Regards,
Huang, Ying

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-05-02  5:35                               ` Huang, Ying
@ 2017-05-02  5:48                                 ` Minchan Kim
  2017-05-02  6:08                                   ` Huang, Ying
  0 siblings, 1 reply; 30+ messages in thread
From: Minchan Kim @ 2017-05-02  5:48 UTC (permalink / raw)
  To: Huang, Ying
  Cc: Andrew Morton, linux-mm, linux-kernel, Hugh Dickins, Shaohua Li,
	Rik van Riel

Hi Huang,

On Tue, May 02, 2017 at 01:35:24PM +0800, Huang, Ying wrote:
> Hi, Minchan,
> 
> Minchan Kim <minchan@kernel.org> writes:
> 
> > On Fri, Apr 28, 2017 at 09:35:37PM +0800, Huang, Ying wrote:
> >> In fact, during the test, I found the overhead of sort() is comparable
> >> with the performance difference of adding likely()/unlikely() to the
> >> "if" in the function.
> >
> > Huang,
> >
> > This discussion is started from your optimization code:
> >
> >         if (nr_swapfiles > 1)
> >                 sort();
> >
> > I don't have such fast machine so cannot test it. However, you added
> > such optimization code in there so I guess it's *worth* to review so
> > with spending my time, I pointed out what you are missing and
> > suggested a idea to find a compromise.
> 
> Sorry for wasting your time and Thanks a lot for your review and
> suggestion!
> 
> When I started talking this with you, I found there is some measurable
> overhead of sort().  But later when I done more tests, I found the
> measurable overhead is at the same level of likely()/unlikely() compiler
> notation.  So you help me to find that, Thanks again!
> 
> > Now you are saying sort is so fast so no worth to add more logics
> > to avoid the overhead?
> > Then, please just drop that if condition part and instead, sort
> > it unconditionally.
> 
> Now, because we found the overhead of sort() is low, I suggest to put
> minimal effort to avoid it.  Like the original implementation,
> 
>          if (nr_swapfiles > 1)
>                  sort();

It might confuse someone in future and would make him/her send a patch
to fix like we discussed. If the logic is not clear and doesn't have
measureable overhead, just leave it which is more simple/clear.

> 
> Or, we can make nr_swapfiles more correct as Tim suggested (tracking
> the number of the swap devices during swap on/off).

It might be better option but it's still hard to justify the patch
because you said it's hard to measure. Such optimiztion patch should
be from numbers.

^ permalink raw reply	[flat|nested] 30+ messages in thread

* Re: [PATCH -mm -v3] mm, swap: Sort swap entries before free
  2017-05-02  5:48                                 ` Minchan Kim
@ 2017-05-02  6:08                                   ` Huang, Ying
  0 siblings, 0 replies; 30+ messages in thread
From: Huang, Ying @ 2017-05-02  6:08 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Huang, Ying, Andrew Morton, linux-mm, linux-kernel, Hugh Dickins,
	Shaohua Li, Rik van Riel

Minchan Kim <minchan@kernel.org> writes:

> Hi Huang,
>
> On Tue, May 02, 2017 at 01:35:24PM +0800, Huang, Ying wrote:
>> Hi, Minchan,
>> 
>> Minchan Kim <minchan@kernel.org> writes:
>> 
>> > On Fri, Apr 28, 2017 at 09:35:37PM +0800, Huang, Ying wrote:
>> >> In fact, during the test, I found the overhead of sort() is comparable
>> >> with the performance difference of adding likely()/unlikely() to the
>> >> "if" in the function.
>> >
>> > Huang,
>> >
>> > This discussion is started from your optimization code:
>> >
>> >         if (nr_swapfiles > 1)
>> >                 sort();
>> >
>> > I don't have such fast machine so cannot test it. However, you added
>> > such optimization code in there so I guess it's *worth* to review so
>> > with spending my time, I pointed out what you are missing and
>> > suggested a idea to find a compromise.
>> 
>> Sorry for wasting your time and Thanks a lot for your review and
>> suggestion!
>> 
>> When I started talking this with you, I found there is some measurable
>> overhead of sort().  But later when I done more tests, I found the
>> measurable overhead is at the same level of likely()/unlikely() compiler
>> notation.  So you help me to find that, Thanks again!
>> 
>> > Now you are saying sort is so fast so no worth to add more logics
>> > to avoid the overhead?
>> > Then, please just drop that if condition part and instead, sort
>> > it unconditionally.
>> 
>> Now, because we found the overhead of sort() is low, I suggest to put
>> minimal effort to avoid it.  Like the original implementation,
>> 
>>          if (nr_swapfiles > 1)
>>                  sort();
>
> It might confuse someone in future and would make him/her send a patch
> to fix like we discussed. If the logic is not clear and doesn't have
> measureable overhead, just leave it which is more simple/clear.

Because the added code is minimal and cheap, I tend to keep it and add
some comments to avoid confusion.  For example,

/*
 * Although nr_swapfiles isn't absolute correct, but the overhead of sort()
 * is so low that it isn't necessary to optimize further.
 */

>> 
>> Or, we can make nr_swapfiles more correct as Tim suggested (tracking
>> the number of the swap devices during swap on/off).
>
> It might be better option but it's still hard to justify the patch
> because you said it's hard to measure. Such optimiztion patch should
> be from numbers.

OK.

Best Regards,
Huang, Ying

^ permalink raw reply	[flat|nested] 30+ messages in thread

end of thread, other threads:[~2017-05-02  6:08 UTC | newest]

Thread overview: 30+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2017-04-07  6:49 [PATCH -mm -v3] mm, swap: Sort swap entries before free Huang, Ying
2017-04-07 13:05 ` Rik van Riel
2017-04-07 21:43 ` Andrew Morton
2017-04-11  7:03   ` Huang, Ying
2017-04-14  1:36   ` Huang, Ying
2017-04-18  4:59 ` Minchan Kim
2017-04-19  8:14   ` Huang, Ying
2017-04-20  6:38     ` Minchan Kim
2017-04-20  7:15       ` Huang, Ying
2017-04-21 12:29         ` Huang, Ying
2017-04-21 23:29           ` Tim Chen
2017-04-23 13:16             ` Huang, Ying
2017-04-24 16:03               ` Tim Chen
2017-04-24  4:52           ` Minchan Kim
2017-04-24  6:47             ` Huang, Ying
2017-04-26 12:42             ` Huang, Ying
2017-04-26 20:13               ` Tim Chen
2017-04-27  1:21                 ` Huang, Ying
2017-04-27 16:48                   ` Tim Chen
2017-04-27  4:35               ` Minchan Kim
2017-04-28  1:09                 ` Huang, Ying
2017-04-28  7:42                   ` Minchan Kim
2017-04-28  8:05                     ` Huang, Ying
2017-04-28  9:00                       ` Minchan Kim
2017-04-28 11:48                         ` Huang, Ying
2017-04-28 13:35                           ` Huang, Ying
2017-05-02  5:02                             ` Minchan Kim
2017-05-02  5:35                               ` Huang, Ying
2017-05-02  5:48                                 ` Minchan Kim
2017-05-02  6:08                                   ` Huang, Ying

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).